var/home/core/zuul-output/0000755000175000017500000000000015114477741014540 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114533654015500 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006701345115114533645017713 0ustar rootrootDec 05 06:54:58 crc systemd[1]: Starting Kubernetes Kubelet... Dec 05 06:54:58 crc restorecon[4815]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:58 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 06:54:59 crc restorecon[4815]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 06:54:59 crc restorecon[4815]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 05 06:54:59 crc kubenswrapper[4997]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 06:54:59 crc kubenswrapper[4997]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 05 06:54:59 crc kubenswrapper[4997]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 06:54:59 crc kubenswrapper[4997]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 06:54:59 crc kubenswrapper[4997]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 05 06:54:59 crc kubenswrapper[4997]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.592679 4997 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596397 4997 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596423 4997 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596430 4997 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596436 4997 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596442 4997 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596447 4997 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596453 4997 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596459 4997 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596464 4997 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596473 4997 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596502 4997 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596508 4997 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596514 4997 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596519 4997 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596524 4997 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596529 4997 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596534 4997 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596540 4997 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596545 4997 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596550 4997 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596555 4997 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596561 4997 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596568 4997 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596574 4997 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596579 4997 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596584 4997 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596589 4997 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596595 4997 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596601 4997 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596607 4997 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596643 4997 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596649 4997 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596655 4997 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596660 4997 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596665 4997 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596670 4997 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596675 4997 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596680 4997 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596685 4997 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596690 4997 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596695 4997 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596700 4997 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596705 4997 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596710 4997 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596715 4997 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596720 4997 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596725 4997 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596730 4997 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596734 4997 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596739 4997 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596744 4997 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596749 4997 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596754 4997 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596758 4997 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596764 4997 feature_gate.go:330] unrecognized feature gate: Example Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596768 4997 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596773 4997 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596778 4997 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596783 4997 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596791 4997 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596796 4997 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596801 4997 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596808 4997 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596813 4997 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596820 4997 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596825 4997 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596830 4997 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596836 4997 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596841 4997 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596846 4997 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.596852 4997 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597153 4997 flags.go:64] FLAG: --address="0.0.0.0" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597170 4997 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597179 4997 flags.go:64] FLAG: --anonymous-auth="true" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597187 4997 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597195 4997 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597201 4997 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597209 4997 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597217 4997 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597222 4997 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597229 4997 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597236 4997 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597242 4997 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597248 4997 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597254 4997 flags.go:64] FLAG: --cgroup-root="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597260 4997 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597266 4997 flags.go:64] FLAG: --client-ca-file="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597272 4997 flags.go:64] FLAG: --cloud-config="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597278 4997 flags.go:64] FLAG: --cloud-provider="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597284 4997 flags.go:64] FLAG: --cluster-dns="[]" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597292 4997 flags.go:64] FLAG: --cluster-domain="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597298 4997 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597304 4997 flags.go:64] FLAG: --config-dir="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597309 4997 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597315 4997 flags.go:64] FLAG: --container-log-max-files="5" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597323 4997 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597330 4997 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597336 4997 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597341 4997 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597347 4997 flags.go:64] FLAG: --contention-profiling="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597353 4997 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597358 4997 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597364 4997 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597370 4997 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597377 4997 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597383 4997 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597390 4997 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597395 4997 flags.go:64] FLAG: --enable-load-reader="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597401 4997 flags.go:64] FLAG: --enable-server="true" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597406 4997 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597413 4997 flags.go:64] FLAG: --event-burst="100" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597419 4997 flags.go:64] FLAG: --event-qps="50" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597426 4997 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597432 4997 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597437 4997 flags.go:64] FLAG: --eviction-hard="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597444 4997 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597450 4997 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597456 4997 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597461 4997 flags.go:64] FLAG: --eviction-soft="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597467 4997 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597474 4997 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597480 4997 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597487 4997 flags.go:64] FLAG: --experimental-mounter-path="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597492 4997 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597497 4997 flags.go:64] FLAG: --fail-swap-on="true" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597503 4997 flags.go:64] FLAG: --feature-gates="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597510 4997 flags.go:64] FLAG: --file-check-frequency="20s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597515 4997 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597522 4997 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597528 4997 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597533 4997 flags.go:64] FLAG: --healthz-port="10248" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597539 4997 flags.go:64] FLAG: --help="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597545 4997 flags.go:64] FLAG: --hostname-override="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597550 4997 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597556 4997 flags.go:64] FLAG: --http-check-frequency="20s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597562 4997 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597568 4997 flags.go:64] FLAG: --image-credential-provider-config="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597573 4997 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597580 4997 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597585 4997 flags.go:64] FLAG: --image-service-endpoint="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597591 4997 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597597 4997 flags.go:64] FLAG: --kube-api-burst="100" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597602 4997 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597634 4997 flags.go:64] FLAG: --kube-api-qps="50" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597649 4997 flags.go:64] FLAG: --kube-reserved="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597657 4997 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597665 4997 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597674 4997 flags.go:64] FLAG: --kubelet-cgroups="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597681 4997 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597688 4997 flags.go:64] FLAG: --lock-file="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597695 4997 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597702 4997 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597710 4997 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597722 4997 flags.go:64] FLAG: --log-json-split-stream="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597728 4997 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597735 4997 flags.go:64] FLAG: --log-text-split-stream="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597741 4997 flags.go:64] FLAG: --logging-format="text" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597747 4997 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597753 4997 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597758 4997 flags.go:64] FLAG: --manifest-url="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597764 4997 flags.go:64] FLAG: --manifest-url-header="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597772 4997 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597778 4997 flags.go:64] FLAG: --max-open-files="1000000" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597785 4997 flags.go:64] FLAG: --max-pods="110" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597791 4997 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597797 4997 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597803 4997 flags.go:64] FLAG: --memory-manager-policy="None" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597809 4997 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597817 4997 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597825 4997 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597832 4997 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597852 4997 flags.go:64] FLAG: --node-status-max-images="50" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597860 4997 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597867 4997 flags.go:64] FLAG: --oom-score-adj="-999" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597876 4997 flags.go:64] FLAG: --pod-cidr="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597882 4997 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597894 4997 flags.go:64] FLAG: --pod-manifest-path="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597901 4997 flags.go:64] FLAG: --pod-max-pids="-1" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597909 4997 flags.go:64] FLAG: --pods-per-core="0" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597915 4997 flags.go:64] FLAG: --port="10250" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597921 4997 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597927 4997 flags.go:64] FLAG: --provider-id="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597933 4997 flags.go:64] FLAG: --qos-reserved="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597940 4997 flags.go:64] FLAG: --read-only-port="10255" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597947 4997 flags.go:64] FLAG: --register-node="true" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597953 4997 flags.go:64] FLAG: --register-schedulable="true" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597960 4997 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597971 4997 flags.go:64] FLAG: --registry-burst="10" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597978 4997 flags.go:64] FLAG: --registry-qps="5" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597984 4997 flags.go:64] FLAG: --reserved-cpus="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597991 4997 flags.go:64] FLAG: --reserved-memory="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.597998 4997 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598005 4997 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598011 4997 flags.go:64] FLAG: --rotate-certificates="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598018 4997 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598023 4997 flags.go:64] FLAG: --runonce="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598029 4997 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598035 4997 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598041 4997 flags.go:64] FLAG: --seccomp-default="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598047 4997 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598053 4997 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598059 4997 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598065 4997 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598071 4997 flags.go:64] FLAG: --storage-driver-password="root" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598076 4997 flags.go:64] FLAG: --storage-driver-secure="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598083 4997 flags.go:64] FLAG: --storage-driver-table="stats" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598088 4997 flags.go:64] FLAG: --storage-driver-user="root" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598094 4997 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598100 4997 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598106 4997 flags.go:64] FLAG: --system-cgroups="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598112 4997 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598120 4997 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598126 4997 flags.go:64] FLAG: --tls-cert-file="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598131 4997 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598139 4997 flags.go:64] FLAG: --tls-min-version="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598145 4997 flags.go:64] FLAG: --tls-private-key-file="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598151 4997 flags.go:64] FLAG: --topology-manager-policy="none" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598156 4997 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598162 4997 flags.go:64] FLAG: --topology-manager-scope="container" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598174 4997 flags.go:64] FLAG: --v="2" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598182 4997 flags.go:64] FLAG: --version="false" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598195 4997 flags.go:64] FLAG: --vmodule="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598202 4997 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598208 4997 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598356 4997 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598363 4997 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598368 4997 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598374 4997 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598380 4997 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598386 4997 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598393 4997 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598398 4997 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598403 4997 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598408 4997 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598413 4997 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598417 4997 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598423 4997 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598429 4997 feature_gate.go:330] unrecognized feature gate: Example Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598433 4997 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598438 4997 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598443 4997 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598448 4997 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598453 4997 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598457 4997 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598462 4997 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598467 4997 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598471 4997 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598476 4997 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598481 4997 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598485 4997 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598490 4997 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598495 4997 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598500 4997 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598507 4997 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598512 4997 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598517 4997 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598522 4997 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598527 4997 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598532 4997 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598539 4997 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598544 4997 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598550 4997 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598562 4997 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598567 4997 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598574 4997 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598579 4997 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598585 4997 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598591 4997 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598598 4997 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598604 4997 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598639 4997 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598645 4997 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598651 4997 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598660 4997 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598668 4997 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598681 4997 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598687 4997 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598695 4997 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598703 4997 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598711 4997 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598717 4997 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598724 4997 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598730 4997 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598736 4997 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598743 4997 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598755 4997 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598763 4997 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598770 4997 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598784 4997 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598790 4997 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598796 4997 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598804 4997 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598811 4997 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598818 4997 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.598828 4997 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.598857 4997 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.614015 4997 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.614084 4997 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614221 4997 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614235 4997 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614244 4997 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614256 4997 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614264 4997 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614272 4997 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614280 4997 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614287 4997 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614295 4997 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614303 4997 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614310 4997 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614318 4997 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614325 4997 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614333 4997 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614341 4997 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614349 4997 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614360 4997 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614370 4997 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614379 4997 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614389 4997 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614399 4997 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614408 4997 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614417 4997 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614428 4997 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614437 4997 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614447 4997 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614456 4997 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614464 4997 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614471 4997 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614479 4997 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614489 4997 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614499 4997 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614508 4997 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614518 4997 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614531 4997 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614540 4997 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614550 4997 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614564 4997 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614579 4997 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614589 4997 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614600 4997 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614644 4997 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614658 4997 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614669 4997 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614679 4997 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614689 4997 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614699 4997 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614710 4997 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614720 4997 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614732 4997 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614745 4997 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614757 4997 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614768 4997 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614779 4997 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614789 4997 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614800 4997 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614811 4997 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614821 4997 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614832 4997 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614844 4997 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614857 4997 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614870 4997 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614882 4997 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614892 4997 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614903 4997 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614913 4997 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614921 4997 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614929 4997 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614936 4997 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614944 4997 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.614954 4997 feature_gate.go:330] unrecognized feature gate: Example Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.614969 4997 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615256 4997 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615272 4997 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615281 4997 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615290 4997 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615299 4997 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615307 4997 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615317 4997 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615327 4997 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615337 4997 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615348 4997 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615357 4997 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615364 4997 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615372 4997 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615379 4997 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615387 4997 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615394 4997 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615402 4997 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615409 4997 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615417 4997 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615426 4997 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615434 4997 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615441 4997 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615481 4997 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615490 4997 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615498 4997 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615505 4997 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615513 4997 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615520 4997 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615530 4997 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615539 4997 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615547 4997 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615554 4997 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615562 4997 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615569 4997 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615578 4997 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615586 4997 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615594 4997 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615602 4997 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615609 4997 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615645 4997 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615654 4997 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615662 4997 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615670 4997 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615678 4997 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615686 4997 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615694 4997 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615704 4997 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615714 4997 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615723 4997 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615731 4997 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615740 4997 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615748 4997 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615756 4997 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615763 4997 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615772 4997 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615779 4997 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615787 4997 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615794 4997 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615804 4997 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615814 4997 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615822 4997 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615830 4997 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615837 4997 feature_gate.go:330] unrecognized feature gate: Example Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615845 4997 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615853 4997 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615860 4997 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615867 4997 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615875 4997 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615885 4997 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615893 4997 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.615903 4997 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.615916 4997 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.616268 4997 server.go:940] "Client rotation is on, will bootstrap in background" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.621163 4997 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.621307 4997 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.622523 4997 server.go:997] "Starting client certificate rotation" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.622573 4997 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.623345 4997 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-23 01:16:35.319056688 +0000 UTC Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.623522 4997 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.629332 4997 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 06:54:59 crc kubenswrapper[4997]: E1205 06:54:59.632322 4997 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.155:6443: connect: connection refused" logger="UnhandledError" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.635053 4997 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.643274 4997 log.go:25] "Validated CRI v1 runtime API" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.656658 4997 log.go:25] "Validated CRI v1 image API" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.658221 4997 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.660832 4997 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-05-06-41-22-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.660870 4997 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.679187 4997 manager.go:217] Machine: {Timestamp:2025-12-05 06:54:59.677693052 +0000 UTC m=+0.206600343 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:35ecf2fa-8290-45a9-ad37-8d68d8e2ec69 BootID:e3043b93-5986-433d-b39e-eae93f14d589 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:16:59:93 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:16:59:93 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:63:a4:4b Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:ee:f6:80 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:a8:91:95 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:4f:d6:98 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:19:1e:83 Speed:-1 Mtu:1496} {Name:ens7.44 MacAddress:52:54:00:53:ac:33 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:be:b9:99:fd:8d:1d Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:5a:5b:5f:e2:3b:d9 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.679516 4997 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.679721 4997 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.680387 4997 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.680691 4997 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.680737 4997 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.680998 4997 topology_manager.go:138] "Creating topology manager with none policy" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.681014 4997 container_manager_linux.go:303] "Creating device plugin manager" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.681270 4997 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.681313 4997 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.681745 4997 state_mem.go:36] "Initialized new in-memory state store" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.681848 4997 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.682465 4997 kubelet.go:418] "Attempting to sync node with API server" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.682490 4997 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.682516 4997 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.682533 4997 kubelet.go:324] "Adding apiserver pod source" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.682549 4997 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.684895 4997 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 05 06:54:59 crc kubenswrapper[4997]: E1205 06:54:59.684974 4997 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.155:6443: connect: connection refused" logger="UnhandledError" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.685019 4997 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.685335 4997 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.685344 4997 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 05 06:54:59 crc kubenswrapper[4997]: E1205 06:54:59.685386 4997 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.155:6443: connect: connection refused" logger="UnhandledError" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.688172 4997 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.689205 4997 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.689245 4997 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.689255 4997 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.689267 4997 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.689282 4997 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.689291 4997 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.689299 4997 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.689312 4997 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.689323 4997 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.689333 4997 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.689367 4997 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.689378 4997 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.689559 4997 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.690149 4997 server.go:1280] "Started kubelet" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.690436 4997 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.690481 4997 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.690953 4997 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.691065 4997 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 05 06:54:59 crc systemd[1]: Started Kubernetes Kubelet. Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.692791 4997 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.692835 4997 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.693107 4997 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 05:56:04.086274097 +0000 UTC Dec 05 06:54:59 crc kubenswrapper[4997]: E1205 06:54:59.693369 4997 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.693435 4997 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.693449 4997 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.693521 4997 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 05 06:54:59 crc kubenswrapper[4997]: E1205 06:54:59.693802 4997 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="200ms" Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.694077 4997 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 05 06:54:59 crc kubenswrapper[4997]: E1205 06:54:59.694166 4997 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.155:6443: connect: connection refused" logger="UnhandledError" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.694713 4997 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 191h1m4.393096207s for next certificate rotation Dec 05 06:54:59 crc kubenswrapper[4997]: E1205 06:54:59.693308 4997 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.155:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e3f4c03d3fcdb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-05 06:54:59.690101979 +0000 UTC m=+0.219009240,LastTimestamp:2025-12-05 06:54:59.690101979 +0000 UTC m=+0.219009240,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.695595 4997 factory.go:55] Registering systemd factory Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.695648 4997 factory.go:221] Registration of the systemd container factory successfully Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.695699 4997 server.go:460] "Adding debug handlers to kubelet server" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.696452 4997 factory.go:153] Registering CRI-O factory Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.696579 4997 factory.go:221] Registration of the crio container factory successfully Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.696767 4997 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.696878 4997 factory.go:103] Registering Raw factory Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.696963 4997 manager.go:1196] Started watching for new ooms in manager Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.697751 4997 manager.go:319] Starting recovery of all containers Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.709206 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.709310 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.709331 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.709353 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.709381 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.709445 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.709532 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.709655 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.710891 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.710968 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.710986 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711003 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711061 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711079 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711094 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711110 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711124 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711149 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711164 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711181 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711197 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711213 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711229 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711247 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711262 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711277 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711299 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711319 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711335 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711351 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711368 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711385 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711418 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711547 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711578 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711591 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711703 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711784 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711807 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711827 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711843 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711858 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711872 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711887 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711901 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711914 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711928 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711942 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711956 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711970 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711984 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.711998 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.712020 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.712036 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.712081 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.712099 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.712930 4997 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.712990 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713009 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713025 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713039 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713072 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713088 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713101 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713116 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713151 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713164 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713179 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713200 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713214 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713230 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713245 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713258 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713274 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713288 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713303 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713319 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713334 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713347 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713410 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713428 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713445 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713463 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713482 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713500 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713519 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713539 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713555 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713569 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713583 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713595 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713629 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713644 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713659 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713675 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713691 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713705 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713721 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713738 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713753 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713765 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713778 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713794 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713808 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713822 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713844 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713859 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713875 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713891 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713907 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713921 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713938 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713953 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713969 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.713985 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714002 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714019 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714034 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714050 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714066 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714083 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714100 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714113 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714137 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714152 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714208 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714221 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714236 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714250 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714266 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714280 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714295 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714311 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714325 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714344 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714358 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714374 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714388 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714402 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714416 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714430 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714445 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714461 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714476 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714493 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714507 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714524 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714540 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714553 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714570 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714592 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714631 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714652 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714670 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714686 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714701 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714717 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714732 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714747 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714760 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714776 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714792 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714806 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714821 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714837 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714852 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714868 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714883 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714897 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714914 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714935 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714952 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714968 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714982 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.714999 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715013 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715027 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715044 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715057 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715072 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715086 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715099 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715117 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715131 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715146 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715159 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715172 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715185 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715199 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715215 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715232 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715251 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715271 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715291 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715308 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715327 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715349 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715363 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715378 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715392 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715407 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715421 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715477 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715496 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715510 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715525 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715538 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715554 4997 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715567 4997 reconstruct.go:97] "Volume reconstruction finished" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.715577 4997 reconciler.go:26] "Reconciler: start to sync state" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.727979 4997 manager.go:324] Recovery completed Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.740267 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.742316 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.742373 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.742405 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.743607 4997 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.743650 4997 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.743687 4997 state_mem.go:36] "Initialized new in-memory state store" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.745946 4997 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.747702 4997 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.747769 4997 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.747807 4997 kubelet.go:2335] "Starting kubelet main sync loop" Dec 05 06:54:59 crc kubenswrapper[4997]: E1205 06:54:59.748045 4997 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 05 06:54:59 crc kubenswrapper[4997]: W1205 06:54:59.750518 4997 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 05 06:54:59 crc kubenswrapper[4997]: E1205 06:54:59.750604 4997 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.155:6443: connect: connection refused" logger="UnhandledError" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.753163 4997 policy_none.go:49] "None policy: Start" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.754025 4997 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.754049 4997 state_mem.go:35] "Initializing new in-memory state store" Dec 05 06:54:59 crc kubenswrapper[4997]: E1205 06:54:59.794256 4997 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.808053 4997 manager.go:334] "Starting Device Plugin manager" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.808121 4997 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.808135 4997 server.go:79] "Starting device plugin registration server" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.808762 4997 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.808783 4997 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.809093 4997 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.809380 4997 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.809407 4997 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 05 06:54:59 crc kubenswrapper[4997]: E1205 06:54:59.818267 4997 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.848582 4997 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.848713 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.850300 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.850338 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.850348 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.850479 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.850885 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.850958 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.851814 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.851885 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.851908 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.852225 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.852326 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.852361 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.852374 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.852399 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.852435 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.856539 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.856859 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.856980 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.859079 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.859155 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.859169 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.859338 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.860470 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.860511 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.862561 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.862583 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.862593 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.863119 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.863277 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.863468 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.863906 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.863989 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.864248 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.866370 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.866434 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.866449 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.866409 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.866713 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.866792 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.866732 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.866922 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.867781 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.867881 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.867971 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:54:59 crc kubenswrapper[4997]: E1205 06:54:59.895060 4997 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="400ms" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.909461 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.911376 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.911491 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.911523 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.911579 4997 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 06:54:59 crc kubenswrapper[4997]: E1205 06:54:59.912580 4997 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.155:6443: connect: connection refused" node="crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.917974 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.918032 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.918158 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.918245 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.918286 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.918327 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.918365 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.918406 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.918443 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.918528 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.918589 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.918672 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.918718 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.918754 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 06:54:59 crc kubenswrapper[4997]: I1205 06:54:59.918788 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.019982 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020069 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020113 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020158 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020204 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020245 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020283 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020324 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020369 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020401 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020491 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020583 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020413 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020774 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020791 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020822 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020692 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020737 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020834 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020849 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020882 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.020783 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.021040 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.021079 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.021238 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.021168 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.021310 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.021330 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.021270 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.021549 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.113791 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.116784 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.116859 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.116880 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.116916 4997 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 06:55:00 crc kubenswrapper[4997]: E1205 06:55:00.117514 4997 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.155:6443: connect: connection refused" node="crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.182368 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.185308 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.199879 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.221110 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: W1205 06:55:00.227253 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-3f25d3464cb664649bec8e0b2c9b338fe44c99e8592159e17145840b4da89d5b WatchSource:0}: Error finding container 3f25d3464cb664649bec8e0b2c9b338fe44c99e8592159e17145840b4da89d5b: Status 404 returned error can't find the container with id 3f25d3464cb664649bec8e0b2c9b338fe44c99e8592159e17145840b4da89d5b Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.227353 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 06:55:00 crc kubenswrapper[4997]: W1205 06:55:00.228946 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-b5f31891cc84d1a8b182d4de375e48afded8577fcf295757f3cd2c5936c70642 WatchSource:0}: Error finding container b5f31891cc84d1a8b182d4de375e48afded8577fcf295757f3cd2c5936c70642: Status 404 returned error can't find the container with id b5f31891cc84d1a8b182d4de375e48afded8577fcf295757f3cd2c5936c70642 Dec 05 06:55:00 crc kubenswrapper[4997]: W1205 06:55:00.231435 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-2228c4b99ca971257a1960aeb86b8440c5ad471147f89ca76d30f267ea0ee9b3 WatchSource:0}: Error finding container 2228c4b99ca971257a1960aeb86b8440c5ad471147f89ca76d30f267ea0ee9b3: Status 404 returned error can't find the container with id 2228c4b99ca971257a1960aeb86b8440c5ad471147f89ca76d30f267ea0ee9b3 Dec 05 06:55:00 crc kubenswrapper[4997]: W1205 06:55:00.242124 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-fc8478943d52e86f4b4f4d2afa6bd01139900b7da9d533363e91e7b4570c344b WatchSource:0}: Error finding container fc8478943d52e86f4b4f4d2afa6bd01139900b7da9d533363e91e7b4570c344b: Status 404 returned error can't find the container with id fc8478943d52e86f4b4f4d2afa6bd01139900b7da9d533363e91e7b4570c344b Dec 05 06:55:00 crc kubenswrapper[4997]: W1205 06:55:00.242780 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-63bd45d78c711c56fca843b70037758c9ee650f9e8732857449faf5b8337438a WatchSource:0}: Error finding container 63bd45d78c711c56fca843b70037758c9ee650f9e8732857449faf5b8337438a: Status 404 returned error can't find the container with id 63bd45d78c711c56fca843b70037758c9ee650f9e8732857449faf5b8337438a Dec 05 06:55:00 crc kubenswrapper[4997]: E1205 06:55:00.296693 4997 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="800ms" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.518167 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.520793 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.520836 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.520850 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.520884 4997 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 06:55:00 crc kubenswrapper[4997]: E1205 06:55:00.521560 4997 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.155:6443: connect: connection refused" node="crc" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.691566 4997 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 05 06:55:00 crc kubenswrapper[4997]: W1205 06:55:00.720725 4997 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 05 06:55:00 crc kubenswrapper[4997]: E1205 06:55:00.720855 4997 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.155:6443: connect: connection refused" logger="UnhandledError" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.753832 4997 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ca4b4c5077572acddfe579d6782bb1b8f7d6ca7ca4b1b82c93cf4f8f5dbdbecf" exitCode=0 Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.753925 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ca4b4c5077572acddfe579d6782bb1b8f7d6ca7ca4b1b82c93cf4f8f5dbdbecf"} Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.754033 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b5f31891cc84d1a8b182d4de375e48afded8577fcf295757f3cd2c5936c70642"} Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.754162 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.755601 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.755698 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.755717 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.758574 4997 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="160fdc21bf6853ca46fa12b0313dc8b1745966d507a77808fc496af986911252" exitCode=0 Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.758645 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"160fdc21bf6853ca46fa12b0313dc8b1745966d507a77808fc496af986911252"} Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.758716 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"63bd45d78c711c56fca843b70037758c9ee650f9e8732857449faf5b8337438a"} Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.758821 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.759735 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.759778 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.759794 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.761334 4997 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8" exitCode=0 Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.761416 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8"} Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.761438 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fc8478943d52e86f4b4f4d2afa6bd01139900b7da9d533363e91e7b4570c344b"} Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.761548 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.762342 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.762399 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.762418 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.763252 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451"} Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.763279 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3f25d3464cb664649bec8e0b2c9b338fe44c99e8592159e17145840b4da89d5b"} Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.765218 4997 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef" exitCode=0 Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.765261 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef"} Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.765323 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2228c4b99ca971257a1960aeb86b8440c5ad471147f89ca76d30f267ea0ee9b3"} Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.765452 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.767294 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.767379 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.767501 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.772294 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.773904 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.773945 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:00 crc kubenswrapper[4997]: I1205 06:55:00.773958 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:00 crc kubenswrapper[4997]: W1205 06:55:00.957345 4997 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 05 06:55:00 crc kubenswrapper[4997]: E1205 06:55:00.957449 4997 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.155:6443: connect: connection refused" logger="UnhandledError" Dec 05 06:55:01 crc kubenswrapper[4997]: E1205 06:55:01.098378 4997 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="1.6s" Dec 05 06:55:01 crc kubenswrapper[4997]: W1205 06:55:01.207056 4997 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 05 06:55:01 crc kubenswrapper[4997]: E1205 06:55:01.207193 4997 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.155:6443: connect: connection refused" logger="UnhandledError" Dec 05 06:55:01 crc kubenswrapper[4997]: W1205 06:55:01.209252 4997 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 05 06:55:01 crc kubenswrapper[4997]: E1205 06:55:01.209303 4997 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.155:6443: connect: connection refused" logger="UnhandledError" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.322294 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.324317 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.324353 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.324363 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.324388 4997 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.769315 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ee2199aceb08b750dc55c65e408e96ee6220e4f3a4be77d72ce68debcbccbb33"} Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.769425 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.770899 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.770940 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.770954 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.772064 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"98126b01cc3216f78df3301553e52376d7dc895cb69ebec44c5681e7b9262d86"} Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.772088 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"03a8cf621e0221d9dc37843dd6dff27479bda2bbd5919953b6ec2cdd11f088c7"} Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.772098 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c4ff6e25dbcddad8fb0ae5ead4aaad30ce69d09dfc8b477217a7dd52bb4d3862"} Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.772156 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.772774 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.772790 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.772797 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.774793 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a"} Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.774834 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38"} Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.774849 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d"} Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.774931 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.775732 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.775763 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.775773 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.778296 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5"} Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.778335 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28"} Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.778351 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c"} Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.778372 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5"} Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.781054 4997 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3850a907a089671274bf5e95460244abc7642759485eef596328c4372f2e217e" exitCode=0 Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.781176 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3850a907a089671274bf5e95460244abc7642759485eef596328c4372f2e217e"} Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.781366 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.782275 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.782387 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.782462 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:01 crc kubenswrapper[4997]: I1205 06:55:01.827427 4997 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 05 06:55:02 crc kubenswrapper[4997]: I1205 06:55:02.732689 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:55:02 crc kubenswrapper[4997]: I1205 06:55:02.737481 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:55:02 crc kubenswrapper[4997]: I1205 06:55:02.786411 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696"} Dec 05 06:55:02 crc kubenswrapper[4997]: I1205 06:55:02.786513 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:02 crc kubenswrapper[4997]: I1205 06:55:02.787379 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:02 crc kubenswrapper[4997]: I1205 06:55:02.787412 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:02 crc kubenswrapper[4997]: I1205 06:55:02.787423 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:02 crc kubenswrapper[4997]: I1205 06:55:02.788587 4997 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="6dc25a863fab6a2c364c8737f58e598e3ddb4ccff124d1fe84e9260da083c06f" exitCode=0 Dec 05 06:55:02 crc kubenswrapper[4997]: I1205 06:55:02.788662 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"6dc25a863fab6a2c364c8737f58e598e3ddb4ccff124d1fe84e9260da083c06f"} Dec 05 06:55:02 crc kubenswrapper[4997]: I1205 06:55:02.788742 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:02 crc kubenswrapper[4997]: I1205 06:55:02.788767 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:02 crc kubenswrapper[4997]: I1205 06:55:02.789509 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:02 crc kubenswrapper[4997]: I1205 06:55:02.789532 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:02 crc kubenswrapper[4997]: I1205 06:55:02.789543 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:02 crc kubenswrapper[4997]: I1205 06:55:02.789754 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:02 crc kubenswrapper[4997]: I1205 06:55:02.789817 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:02 crc kubenswrapper[4997]: I1205 06:55:02.789837 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:03 crc kubenswrapper[4997]: I1205 06:55:03.137170 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:55:03 crc kubenswrapper[4997]: I1205 06:55:03.137414 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:55:03 crc kubenswrapper[4997]: I1205 06:55:03.798873 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"410a246c1bd895bb9f69b2eee579ebe89cbcff2e72e5e009aecacd483257a7b3"} Dec 05 06:55:03 crc kubenswrapper[4997]: I1205 06:55:03.798973 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"cf5e819910551a055a2c7e9733e2c78587844c95ad4c0eef266983a0adce1dd0"} Dec 05 06:55:03 crc kubenswrapper[4997]: I1205 06:55:03.799062 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:03 crc kubenswrapper[4997]: I1205 06:55:03.799136 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:03 crc kubenswrapper[4997]: I1205 06:55:03.799061 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6c3d52c77f78baf450d2b3e6e68bd2426339877ddc0e336c4a64be336dec2222"} Dec 05 06:55:03 crc kubenswrapper[4997]: I1205 06:55:03.799426 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:55:03 crc kubenswrapper[4997]: I1205 06:55:03.799463 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"59a30514ff428e57109913523223c2c45bcca3b0ff27077d8e5df5f7d94b8df1"} Dec 05 06:55:03 crc kubenswrapper[4997]: I1205 06:55:03.801547 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:03 crc kubenswrapper[4997]: I1205 06:55:03.801677 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:03 crc kubenswrapper[4997]: I1205 06:55:03.801755 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:03 crc kubenswrapper[4997]: I1205 06:55:03.807050 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:03 crc kubenswrapper[4997]: I1205 06:55:03.807120 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:03 crc kubenswrapper[4997]: I1205 06:55:03.807140 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.241957 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.242357 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.244669 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.244731 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.244756 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.543448 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.810144 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f23ebdccc82cc5e90a6efd59df5cd7ce1a045cb867cbffe679ce0ca635e9f117"} Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.810294 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.810448 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.811507 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.812472 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.812530 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.812555 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.812824 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.812874 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.812897 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.812995 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.813026 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.813043 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:04 crc kubenswrapper[4997]: I1205 06:55:04.923668 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:55:05 crc kubenswrapper[4997]: I1205 06:55:05.190544 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 05 06:55:05 crc kubenswrapper[4997]: I1205 06:55:05.386339 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 05 06:55:05 crc kubenswrapper[4997]: I1205 06:55:05.814491 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:05 crc kubenswrapper[4997]: I1205 06:55:05.814583 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:05 crc kubenswrapper[4997]: I1205 06:55:05.814491 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:05 crc kubenswrapper[4997]: I1205 06:55:05.816998 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:05 crc kubenswrapper[4997]: I1205 06:55:05.817039 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:05 crc kubenswrapper[4997]: I1205 06:55:05.817057 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:05 crc kubenswrapper[4997]: I1205 06:55:05.817122 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:05 crc kubenswrapper[4997]: I1205 06:55:05.817154 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:05 crc kubenswrapper[4997]: I1205 06:55:05.817171 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:05 crc kubenswrapper[4997]: I1205 06:55:05.817668 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:05 crc kubenswrapper[4997]: I1205 06:55:05.817743 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:05 crc kubenswrapper[4997]: I1205 06:55:05.817769 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:06 crc kubenswrapper[4997]: I1205 06:55:06.415065 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:55:06 crc kubenswrapper[4997]: I1205 06:55:06.816434 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:06 crc kubenswrapper[4997]: I1205 06:55:06.816434 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:06 crc kubenswrapper[4997]: I1205 06:55:06.817579 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:06 crc kubenswrapper[4997]: I1205 06:55:06.817642 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:06 crc kubenswrapper[4997]: I1205 06:55:06.817656 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:06 crc kubenswrapper[4997]: I1205 06:55:06.817673 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:06 crc kubenswrapper[4997]: I1205 06:55:06.817699 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:06 crc kubenswrapper[4997]: I1205 06:55:06.817714 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:09 crc kubenswrapper[4997]: I1205 06:55:09.416069 4997 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 06:55:09 crc kubenswrapper[4997]: I1205 06:55:09.416195 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 06:55:09 crc kubenswrapper[4997]: E1205 06:55:09.818552 4997 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 05 06:55:11 crc kubenswrapper[4997]: E1205 06:55:11.326488 4997 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 05 06:55:11 crc kubenswrapper[4997]: I1205 06:55:11.692801 4997 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 05 06:55:11 crc kubenswrapper[4997]: E1205 06:55:11.828740 4997 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 05 06:55:12 crc kubenswrapper[4997]: I1205 06:55:12.611626 4997 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 05 06:55:12 crc kubenswrapper[4997]: I1205 06:55:12.611711 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 05 06:55:12 crc kubenswrapper[4997]: I1205 06:55:12.616699 4997 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 05 06:55:12 crc kubenswrapper[4997]: I1205 06:55:12.616761 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 05 06:55:12 crc kubenswrapper[4997]: I1205 06:55:12.927314 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:12 crc kubenswrapper[4997]: I1205 06:55:12.928544 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:12 crc kubenswrapper[4997]: I1205 06:55:12.928589 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:12 crc kubenswrapper[4997]: I1205 06:55:12.928600 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:12 crc kubenswrapper[4997]: I1205 06:55:12.928663 4997 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 06:55:13 crc kubenswrapper[4997]: I1205 06:55:13.142596 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:55:13 crc kubenswrapper[4997]: I1205 06:55:13.142777 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:13 crc kubenswrapper[4997]: I1205 06:55:13.144006 4997 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]log ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]etcd ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/generic-apiserver-start-informers ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/priority-and-fairness-filter ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/start-apiextensions-informers ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/start-apiextensions-controllers ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/crd-informer-synced ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/start-system-namespaces-controller ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 05 06:55:13 crc kubenswrapper[4997]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 05 06:55:13 crc kubenswrapper[4997]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/bootstrap-controller ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/start-kube-aggregator-informers ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/apiservice-registration-controller ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/apiservice-discovery-controller ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]autoregister-completion ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/apiservice-openapi-controller ok Dec 05 06:55:13 crc kubenswrapper[4997]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 05 06:55:13 crc kubenswrapper[4997]: livez check failed Dec 05 06:55:13 crc kubenswrapper[4997]: I1205 06:55:13.145365 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 06:55:13 crc kubenswrapper[4997]: I1205 06:55:13.145505 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:13 crc kubenswrapper[4997]: I1205 06:55:13.145543 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:13 crc kubenswrapper[4997]: I1205 06:55:13.145561 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:15 crc kubenswrapper[4997]: I1205 06:55:15.408194 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 05 06:55:15 crc kubenswrapper[4997]: I1205 06:55:15.408410 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:15 crc kubenswrapper[4997]: I1205 06:55:15.409928 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:15 crc kubenswrapper[4997]: I1205 06:55:15.409966 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:15 crc kubenswrapper[4997]: I1205 06:55:15.409974 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:15 crc kubenswrapper[4997]: I1205 06:55:15.422895 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 05 06:55:15 crc kubenswrapper[4997]: I1205 06:55:15.838164 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:15 crc kubenswrapper[4997]: I1205 06:55:15.839250 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:15 crc kubenswrapper[4997]: I1205 06:55:15.839280 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:15 crc kubenswrapper[4997]: I1205 06:55:15.839292 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:16 crc kubenswrapper[4997]: I1205 06:55:16.095735 4997 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 05 06:55:16 crc kubenswrapper[4997]: I1205 06:55:16.110238 4997 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 05 06:55:16 crc kubenswrapper[4997]: I1205 06:55:16.517949 4997 csr.go:261] certificate signing request csr-gvwxq is approved, waiting to be issued Dec 05 06:55:16 crc kubenswrapper[4997]: I1205 06:55:16.532716 4997 csr.go:257] certificate signing request csr-gvwxq is issued Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.534314 4997 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-05 06:50:16 +0000 UTC, rotation deadline is 2026-09-05 12:27:42.335321989 +0000 UTC Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.534356 4997 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6581h32m24.800968443s for next certificate rotation Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.610894 4997 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.613905 4997 trace.go:236] Trace[310317783]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 06:55:03.031) (total time: 14582ms): Dec 05 06:55:17 crc kubenswrapper[4997]: Trace[310317783]: ---"Objects listed" error: 14582ms (06:55:17.613) Dec 05 06:55:17 crc kubenswrapper[4997]: Trace[310317783]: [14.582184093s] [14.582184093s] END Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.613943 4997 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.615211 4997 trace.go:236] Trace[316079389]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 06:55:03.567) (total time: 14048ms): Dec 05 06:55:17 crc kubenswrapper[4997]: Trace[316079389]: ---"Objects listed" error: 14047ms (06:55:17.615) Dec 05 06:55:17 crc kubenswrapper[4997]: Trace[316079389]: [14.048081912s] [14.048081912s] END Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.615233 4997 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.616860 4997 trace.go:236] Trace[580496700]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 06:55:02.905) (total time: 14711ms): Dec 05 06:55:17 crc kubenswrapper[4997]: Trace[580496700]: ---"Objects listed" error: 14711ms (06:55:17.616) Dec 05 06:55:17 crc kubenswrapper[4997]: Trace[580496700]: [14.711540525s] [14.711540525s] END Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.616883 4997 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.617595 4997 trace.go:236] Trace[41811248]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 06:55:03.125) (total time: 14491ms): Dec 05 06:55:17 crc kubenswrapper[4997]: Trace[41811248]: ---"Objects listed" error: 14491ms (06:55:17.617) Dec 05 06:55:17 crc kubenswrapper[4997]: Trace[41811248]: [14.49165243s] [14.49165243s] END Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.617629 4997 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.623139 4997 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.658487 4997 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51270->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.658571 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51270->192.168.126.11:17697: read: connection reset by peer" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.683374 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.686726 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.692824 4997 apiserver.go:52] "Watching apiserver" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.695900 4997 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.696145 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.696485 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.696576 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.696641 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.696650 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.696682 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.696751 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.696749 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.696789 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.697010 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.699467 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.699578 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.699655 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.699755 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.700342 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.700423 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.700596 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.701759 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.705077 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.727996 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.743969 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.758524 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.774825 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.786473 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.794820 4997 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.796557 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.809565 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.823976 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824027 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824051 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824074 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824094 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824118 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824147 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824170 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824192 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824212 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824237 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824263 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824286 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824309 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824330 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824353 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824375 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824400 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824475 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824503 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824521 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824540 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824558 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824575 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824595 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824631 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824648 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824671 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824689 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824708 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824727 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824744 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824761 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824855 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824872 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824924 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824943 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824958 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824974 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824994 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825009 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825026 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825041 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825058 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825081 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825098 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825115 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825130 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825145 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825159 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825176 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825195 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825210 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825227 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825271 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825289 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825304 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825322 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825338 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825382 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825397 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825412 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825428 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825443 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825459 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825475 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825489 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825508 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825523 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825540 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825558 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825579 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825597 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825628 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825646 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825664 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825679 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825694 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825710 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825727 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825742 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825757 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825777 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825794 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825810 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825825 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825842 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825858 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825874 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825889 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825904 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825921 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825939 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825959 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825978 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825997 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826022 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826044 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826062 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826088 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826107 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826127 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826147 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826168 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826195 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826220 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826240 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826261 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826285 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826305 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826326 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826350 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826386 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826432 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826458 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826483 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826505 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826530 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826554 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826575 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826597 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826642 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826669 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826695 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826720 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826743 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826764 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826787 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826811 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826834 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826857 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826882 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826905 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826928 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826954 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826976 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827001 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827032 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827059 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827081 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827104 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827127 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827151 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827168 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827184 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827206 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827230 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827261 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827277 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824598 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827302 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827329 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824714 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.824900 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825074 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825565 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825630 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825818 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827389 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827422 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.825996 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826165 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826156 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826309 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826489 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826591 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826678 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826766 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826902 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.826947 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827137 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827289 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827561 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827597 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827773 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827943 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.828123 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.828177 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.828183 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.828314 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.828450 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.828574 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.829698 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.831153 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.831437 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.831986 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.832174 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.832437 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.832544 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.832720 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:55:18.332691116 +0000 UTC m=+18.861598617 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.832947 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.833294 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.833755 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.833021 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.834148 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.834359 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.834572 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.834839 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.834927 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.835102 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.835141 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.835138 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.835406 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.835422 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.835587 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.835658 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.835789 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.835796 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.836007 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.836158 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.836293 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.836330 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.836494 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.836742 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.836764 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.837048 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.837501 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.837660 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.837826 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.838104 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.838238 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.838497 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.838561 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.838964 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.839118 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.839302 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.839304 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.839352 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.839491 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.839759 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.839771 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.839787 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.839917 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.840060 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.840055 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.840404 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.840409 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.840603 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.840670 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.840990 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.841007 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842315 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842343 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842358 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.827347 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842445 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842464 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842471 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842503 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842524 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842545 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842565 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842585 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842603 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842637 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842655 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842671 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842688 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842705 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842723 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842749 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842772 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842794 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842820 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842855 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842882 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842901 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842922 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842945 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842972 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842995 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843019 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843052 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843077 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843099 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843125 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843149 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843352 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843375 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843430 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843455 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843525 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843553 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843578 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843596 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843642 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843666 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843689 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843713 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843739 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843762 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844420 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844499 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844523 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844550 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844576 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844602 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844640 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844660 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844682 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844706 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844725 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844742 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844761 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844782 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844870 4997 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844883 4997 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844893 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844905 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844915 4997 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844925 4997 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844935 4997 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844945 4997 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844956 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844967 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844976 4997 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844986 4997 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844996 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845005 4997 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845015 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845024 4997 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845033 4997 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845043 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845054 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845064 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845073 4997 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845088 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845097 4997 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845107 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845116 4997 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845127 4997 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845142 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845152 4997 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845161 4997 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845170 4997 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845180 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845190 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845202 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845211 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845222 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845233 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845242 4997 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845252 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845261 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845270 4997 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845279 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845288 4997 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845296 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845305 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845315 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845324 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845334 4997 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845439 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845448 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845457 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845466 4997 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845477 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845485 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845495 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845504 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845513 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845522 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845531 4997 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845541 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845550 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845559 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845569 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845579 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845589 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845599 4997 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845622 4997 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845632 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845641 4997 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845650 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845665 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845674 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845683 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845692 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845701 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845709 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845718 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845727 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845737 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845746 4997 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845761 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845770 4997 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845778 4997 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845787 4997 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845795 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845804 4997 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845813 4997 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845823 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845832 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845840 4997 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845851 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845860 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845869 4997 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845878 4997 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.845886 4997 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842502 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.842888 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843057 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843100 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843425 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843515 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843799 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843852 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.843962 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844125 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844153 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844037 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844200 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844228 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.844508 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.845941 4997 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.851169 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:18.351148232 +0000 UTC m=+18.880055493 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.851487 4997 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.852755 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.853205 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.853497 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.853570 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.853569 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.853651 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.853763 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.854257 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.854115 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.854491 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.854681 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.854970 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.849132 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.849265 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.849476 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.849509 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.849520 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.849684 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.849821 4997 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.855529 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:18.355506679 +0000 UTC m=+18.884413930 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.849827 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.850167 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.850235 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.850502 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.855567 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.850922 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.855070 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.854983 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.855750 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.855770 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.856037 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.856074 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.856413 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.857540 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.857778 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.848690 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.857934 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.858077 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.858096 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.858102 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.858261 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.858296 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.858377 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.858553 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.858922 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.859174 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.859216 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.859560 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.859875 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.859892 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.860055 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.860065 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.860099 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.860123 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.858519 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.862968 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.863057 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.866257 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.866296 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.866321 4997 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.866393 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:18.366367841 +0000 UTC m=+18.895275102 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.867337 4997 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696" exitCode=255 Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.867867 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696"} Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.868533 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.868555 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.868565 4997 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.868604 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:18.368590151 +0000 UTC m=+18.897497412 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.869567 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.870770 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.870794 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.872720 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.873730 4997 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.873793 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.874349 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.874830 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.875572 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.875831 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.877473 4997 scope.go:117] "RemoveContainer" containerID="4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.878724 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.878983 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.879415 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.881715 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.888958 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.888996 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.889805 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.890293 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.890387 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.890729 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.890776 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.891034 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.891167 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.891332 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.891811 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.891852 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.892459 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.893792 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.894187 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.894259 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.897165 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.897573 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.897738 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.897769 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.898948 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.899495 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.899633 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.901310 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.901688 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.901917 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.903674 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.907569 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.912295 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.922724 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.931553 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.933275 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:55:17 crc kubenswrapper[4997]: E1205 06:55:17.934565 4997 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.934710 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.947812 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948175 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948217 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948252 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948330 4997 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948346 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948361 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948375 4997 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948388 4997 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948400 4997 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948413 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948427 4997 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948441 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948457 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948470 4997 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948481 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948491 4997 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948502 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948512 4997 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948522 4997 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948532 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948541 4997 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948551 4997 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948561 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948571 4997 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948581 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948589 4997 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948600 4997 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948623 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948634 4997 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948643 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948653 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948663 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948673 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948684 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948694 4997 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948704 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948714 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948724 4997 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948733 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948742 4997 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948753 4997 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948763 4997 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948947 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948962 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948972 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948982 4997 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948992 4997 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949002 4997 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949012 4997 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949021 4997 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949029 4997 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949039 4997 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949049 4997 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949057 4997 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949067 4997 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949077 4997 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949087 4997 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949095 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949105 4997 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949114 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949122 4997 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949132 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949141 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949150 4997 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949159 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949168 4997 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949179 4997 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949189 4997 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949198 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949206 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949215 4997 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949224 4997 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949233 4997 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949243 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949252 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949261 4997 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949270 4997 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949280 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949289 4997 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949299 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949308 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949318 4997 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949329 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949370 4997 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949380 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949389 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949398 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949407 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949416 4997 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949424 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949434 4997 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949452 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949461 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949483 4997 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949493 4997 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949502 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949512 4997 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949521 4997 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949530 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949539 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949549 4997 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.949558 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.948001 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.950175 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.960636 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.970096 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:17 crc kubenswrapper[4997]: I1205 06:55:17.989860 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.004185 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.012894 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.016680 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.020264 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.025546 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-l5p2z"] Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.025843 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.025865 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-l5p2z" Dec 05 06:55:18 crc kubenswrapper[4997]: W1205 06:55:18.033692 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-67f2350c2154d3ad4e64d95a3c23dc9736e366607e1445c7780fd13163391a8f WatchSource:0}: Error finding container 67f2350c2154d3ad4e64d95a3c23dc9736e366607e1445c7780fd13163391a8f: Status 404 returned error can't find the container with id 67f2350c2154d3ad4e64d95a3c23dc9736e366607e1445c7780fd13163391a8f Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.033839 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.033910 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.033840 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.064601 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.093385 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.142284 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.148752 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.152275 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frk5k\" (UniqueName: \"kubernetes.io/projected/44799328-a533-4c95-9712-683b649c8ebf-kube-api-access-frk5k\") pod \"node-resolver-l5p2z\" (UID: \"44799328-a533-4c95-9712-683b649c8ebf\") " pod="openshift-dns/node-resolver-l5p2z" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.152319 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/44799328-a533-4c95-9712-683b649c8ebf-hosts-file\") pod \"node-resolver-l5p2z\" (UID: \"44799328-a533-4c95-9712-683b649c8ebf\") " pod="openshift-dns/node-resolver-l5p2z" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.187323 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.212311 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.241602 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.252914 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frk5k\" (UniqueName: \"kubernetes.io/projected/44799328-a533-4c95-9712-683b649c8ebf-kube-api-access-frk5k\") pod \"node-resolver-l5p2z\" (UID: \"44799328-a533-4c95-9712-683b649c8ebf\") " pod="openshift-dns/node-resolver-l5p2z" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.252966 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/44799328-a533-4c95-9712-683b649c8ebf-hosts-file\") pod \"node-resolver-l5p2z\" (UID: \"44799328-a533-4c95-9712-683b649c8ebf\") " pod="openshift-dns/node-resolver-l5p2z" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.253241 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/44799328-a533-4c95-9712-683b649c8ebf-hosts-file\") pod \"node-resolver-l5p2z\" (UID: \"44799328-a533-4c95-9712-683b649c8ebf\") " pod="openshift-dns/node-resolver-l5p2z" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.260849 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.285690 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.290319 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frk5k\" (UniqueName: \"kubernetes.io/projected/44799328-a533-4c95-9712-683b649c8ebf-kube-api-access-frk5k\") pod \"node-resolver-l5p2z\" (UID: \"44799328-a533-4c95-9712-683b649c8ebf\") " pod="openshift-dns/node-resolver-l5p2z" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.296694 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.309276 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.329185 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.343880 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.343933 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-l5p2z" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.353884 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.353978 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:18 crc kubenswrapper[4997]: E1205 06:55:18.354078 4997 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 06:55:18 crc kubenswrapper[4997]: E1205 06:55:18.354078 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:55:19.354050971 +0000 UTC m=+19.882958232 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:55:18 crc kubenswrapper[4997]: E1205 06:55:18.354242 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:19.354205356 +0000 UTC m=+19.883112807 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 06:55:18 crc kubenswrapper[4997]: W1205 06:55:18.354335 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44799328_a533_4c95_9712_683b649c8ebf.slice/crio-8941ccf200a75c969d908d096c73da711ea9e272c3c67cf809f5b9354f2dfca4 WatchSource:0}: Error finding container 8941ccf200a75c969d908d096c73da711ea9e272c3c67cf809f5b9354f2dfca4: Status 404 returned error can't find the container with id 8941ccf200a75c969d908d096c73da711ea9e272c3c67cf809f5b9354f2dfca4 Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.354948 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.365647 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.382080 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.390831 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.399952 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.410157 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.429376 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-th9dq"] Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.429679 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-l7hsk"] Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.429907 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.430165 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.433028 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.433261 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.433700 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.433707 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.433774 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.433820 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.433883 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.433999 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.434068 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.436325 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.444135 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.454894 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.455247 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:18 crc kubenswrapper[4997]: E1205 06:55:18.455399 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 06:55:18 crc kubenswrapper[4997]: E1205 06:55:18.455421 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 06:55:18 crc kubenswrapper[4997]: E1205 06:55:18.455434 4997 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:18 crc kubenswrapper[4997]: E1205 06:55:18.455482 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:19.455466028 +0000 UTC m=+19.984373289 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.455275 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.455526 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:18 crc kubenswrapper[4997]: E1205 06:55:18.455606 4997 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 06:55:18 crc kubenswrapper[4997]: E1205 06:55:18.455644 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:19.455638283 +0000 UTC m=+19.984545544 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 06:55:18 crc kubenswrapper[4997]: E1205 06:55:18.455727 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 06:55:18 crc kubenswrapper[4997]: E1205 06:55:18.455767 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 06:55:18 crc kubenswrapper[4997]: E1205 06:55:18.455783 4997 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:18 crc kubenswrapper[4997]: E1205 06:55:18.455862 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:19.455834518 +0000 UTC m=+19.984741779 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.466537 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.480318 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.495948 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.507607 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.517894 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.529071 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.541842 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.549706 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.556662 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/28cde6da-1de5-455a-93c8-dd250f9ea4d0-rootfs\") pod \"machine-config-daemon-l7hsk\" (UID: \"28cde6da-1de5-455a-93c8-dd250f9ea4d0\") " pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.556701 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/28cde6da-1de5-455a-93c8-dd250f9ea4d0-mcd-auth-proxy-config\") pod \"machine-config-daemon-l7hsk\" (UID: \"28cde6da-1de5-455a-93c8-dd250f9ea4d0\") " pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.556719 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v4j8\" (UniqueName: \"kubernetes.io/projected/28cde6da-1de5-455a-93c8-dd250f9ea4d0-kube-api-access-4v4j8\") pod \"machine-config-daemon-l7hsk\" (UID: \"28cde6da-1de5-455a-93c8-dd250f9ea4d0\") " pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.556750 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-system-cni-dir\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.556766 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-os-release\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.556782 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-multus-socket-dir-parent\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.556798 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-run-k8s-cni-cncf-io\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.556814 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/28cde6da-1de5-455a-93c8-dd250f9ea4d0-proxy-tls\") pod \"machine-config-daemon-l7hsk\" (UID: \"28cde6da-1de5-455a-93c8-dd250f9ea4d0\") " pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.556838 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-run-netns\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.556911 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-var-lib-cni-multus\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.556963 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-run-multus-certs\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.557005 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/85f05eb3-df7f-4c46-a0e2-451766bb3297-cni-binary-copy\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.557050 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-multus-cni-dir\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.557072 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-cnibin\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.557091 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-var-lib-kubelet\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.557127 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-etc-kubernetes\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.557175 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-var-lib-cni-bin\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.557199 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-hostroot\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.557221 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-multus-conf-dir\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.557243 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rhm5\" (UniqueName: \"kubernetes.io/projected/85f05eb3-df7f-4c46-a0e2-451766bb3297-kube-api-access-5rhm5\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.557287 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/85f05eb3-df7f-4c46-a0e2-451766bb3297-multus-daemon-config\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.558986 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.567598 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.581566 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.591677 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.601643 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:18Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.612653 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:18Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.630515 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:18Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.657760 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/28cde6da-1de5-455a-93c8-dd250f9ea4d0-mcd-auth-proxy-config\") pod \"machine-config-daemon-l7hsk\" (UID: \"28cde6da-1de5-455a-93c8-dd250f9ea4d0\") " pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658016 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v4j8\" (UniqueName: \"kubernetes.io/projected/28cde6da-1de5-455a-93c8-dd250f9ea4d0-kube-api-access-4v4j8\") pod \"machine-config-daemon-l7hsk\" (UID: \"28cde6da-1de5-455a-93c8-dd250f9ea4d0\") " pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658033 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-system-cni-dir\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658054 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-multus-socket-dir-parent\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658078 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-run-k8s-cni-cncf-io\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658094 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/28cde6da-1de5-455a-93c8-dd250f9ea4d0-proxy-tls\") pod \"machine-config-daemon-l7hsk\" (UID: \"28cde6da-1de5-455a-93c8-dd250f9ea4d0\") " pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658109 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-os-release\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658132 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-var-lib-cni-multus\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658147 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-run-multus-certs\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658161 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/85f05eb3-df7f-4c46-a0e2-451766bb3297-cni-binary-copy\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658176 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-run-netns\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658191 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-multus-cni-dir\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658208 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-cnibin\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658222 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-var-lib-kubelet\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658235 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-etc-kubernetes\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658271 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-var-lib-cni-bin\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658288 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-hostroot\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658305 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-multus-conf-dir\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658320 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rhm5\" (UniqueName: \"kubernetes.io/projected/85f05eb3-df7f-4c46-a0e2-451766bb3297-kube-api-access-5rhm5\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658335 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/85f05eb3-df7f-4c46-a0e2-451766bb3297-multus-daemon-config\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658355 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/28cde6da-1de5-455a-93c8-dd250f9ea4d0-rootfs\") pod \"machine-config-daemon-l7hsk\" (UID: \"28cde6da-1de5-455a-93c8-dd250f9ea4d0\") " pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658410 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/28cde6da-1de5-455a-93c8-dd250f9ea4d0-rootfs\") pod \"machine-config-daemon-l7hsk\" (UID: \"28cde6da-1de5-455a-93c8-dd250f9ea4d0\") " pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658570 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-system-cni-dir\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658599 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-cnibin\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658639 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-var-lib-kubelet\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658653 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-multus-cni-dir\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658660 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-etc-kubernetes\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658678 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-var-lib-cni-bin\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658695 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-hostroot\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658716 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-multus-conf-dir\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658725 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-multus-socket-dir-parent\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658764 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-run-k8s-cni-cncf-io\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.658904 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/28cde6da-1de5-455a-93c8-dd250f9ea4d0-mcd-auth-proxy-config\") pod \"machine-config-daemon-l7hsk\" (UID: \"28cde6da-1de5-455a-93c8-dd250f9ea4d0\") " pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.659022 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-run-multus-certs\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.659315 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/85f05eb3-df7f-4c46-a0e2-451766bb3297-multus-daemon-config\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.659364 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-var-lib-cni-multus\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.659456 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-os-release\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.659526 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/85f05eb3-df7f-4c46-a0e2-451766bb3297-host-run-netns\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.659750 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/85f05eb3-df7f-4c46-a0e2-451766bb3297-cni-binary-copy\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.663153 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/28cde6da-1de5-455a-93c8-dd250f9ea4d0-proxy-tls\") pod \"machine-config-daemon-l7hsk\" (UID: \"28cde6da-1de5-455a-93c8-dd250f9ea4d0\") " pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.664204 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:18Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.686502 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v4j8\" (UniqueName: \"kubernetes.io/projected/28cde6da-1de5-455a-93c8-dd250f9ea4d0-kube-api-access-4v4j8\") pod \"machine-config-daemon-l7hsk\" (UID: \"28cde6da-1de5-455a-93c8-dd250f9ea4d0\") " pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.708523 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rhm5\" (UniqueName: \"kubernetes.io/projected/85f05eb3-df7f-4c46-a0e2-451766bb3297-kube-api-access-5rhm5\") pod \"multus-th9dq\" (UID: \"85f05eb3-df7f-4c46-a0e2-451766bb3297\") " pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.744771 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.745059 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:18Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.753386 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-th9dq" Dec 05 06:55:18 crc kubenswrapper[4997]: W1205 06:55:18.755519 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28cde6da_1de5_455a_93c8_dd250f9ea4d0.slice/crio-2cacf33595db02eccfcc938e9f375deff57d6da54afa9b97676e8fe0cb2f5cd4 WatchSource:0}: Error finding container 2cacf33595db02eccfcc938e9f375deff57d6da54afa9b97676e8fe0cb2f5cd4: Status 404 returned error can't find the container with id 2cacf33595db02eccfcc938e9f375deff57d6da54afa9b97676e8fe0cb2f5cd4 Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.787554 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:18Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.797474 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bchpk"] Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.798507 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-5m9fs"] Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.799209 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.799675 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.827586 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:18Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.834150 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.852910 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.859271 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/76d5e7c0-c9d4-47c8-8a9e-d96565189197-os-release\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.859337 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-cni-netd\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.859379 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovnkube-script-lib\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.859419 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/76d5e7c0-c9d4-47c8-8a9e-d96565189197-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.859456 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-var-lib-openvswitch\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.859492 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovnkube-config\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.859584 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-cni-bin\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.859660 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.859775 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/76d5e7c0-c9d4-47c8-8a9e-d96565189197-cnibin\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.859807 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-kubelet\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.859826 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-log-socket\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.859848 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovn-node-metrics-cert\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.859882 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-slash\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.859900 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-openvswitch\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.859948 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-node-log\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.859985 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-systemd-units\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.860001 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-ovn\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.860023 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-etc-openvswitch\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.860047 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/76d5e7c0-c9d4-47c8-8a9e-d96565189197-cni-binary-copy\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.860064 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qmw5\" (UniqueName: \"kubernetes.io/projected/76d5e7c0-c9d4-47c8-8a9e-d96565189197-kube-api-access-9qmw5\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.860088 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-env-overrides\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.860103 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/76d5e7c0-c9d4-47c8-8a9e-d96565189197-system-cni-dir\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.860122 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-run-netns\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.860165 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-run-ovn-kubernetes\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.860267 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-systemd\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.860371 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/76d5e7c0-c9d4-47c8-8a9e-d96565189197-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.860401 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vmkd\" (UniqueName: \"kubernetes.io/projected/28f65c70-f59b-4393-b9c9-8163bb596f6c-kube-api-access-2vmkd\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.878796 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.878979 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c"} Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.879012 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"88c80b1562189024b3380b094ccc8b8819367a8c3f69b0db884142ba14cf5ce3"} Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.885604 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.892725 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.894199 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b"} Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.895004 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.896016 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-l5p2z" event={"ID":"44799328-a533-4c95-9712-683b649c8ebf","Type":"ContainerStarted","Data":"e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114"} Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.896079 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-l5p2z" event={"ID":"44799328-a533-4c95-9712-683b649c8ebf","Type":"ContainerStarted","Data":"8941ccf200a75c969d908d096c73da711ea9e272c3c67cf809f5b9354f2dfca4"} Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.897438 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-th9dq" event={"ID":"85f05eb3-df7f-4c46-a0e2-451766bb3297","Type":"ContainerStarted","Data":"7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6"} Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.897489 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-th9dq" event={"ID":"85f05eb3-df7f-4c46-a0e2-451766bb3297","Type":"ContainerStarted","Data":"a241a85d8b470c2b70e277137f1ec8bcbe2f391c479b113c54cb75c11e3b3548"} Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.898980 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a"} Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.899006 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"2cacf33595db02eccfcc938e9f375deff57d6da54afa9b97676e8fe0cb2f5cd4"} Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.899637 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"40f274cdb83d2cb22dfa13b2c4fe386af8d4cf3fe19f13b476645585bfa418f0"} Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.900377 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.901107 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918"} Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.901163 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc"} Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.901190 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"67f2350c2154d3ad4e64d95a3c23dc9736e366607e1445c7780fd13163391a8f"} Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.912628 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.932968 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.953133 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.961195 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovnkube-config\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.961262 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/76d5e7c0-c9d4-47c8-8a9e-d96565189197-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.961283 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-var-lib-openvswitch\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.961300 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-cni-bin\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.961401 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/76d5e7c0-c9d4-47c8-8a9e-d96565189197-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.961398 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-var-lib-openvswitch\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.961434 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.961420 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-cni-bin\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.961475 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.961526 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/76d5e7c0-c9d4-47c8-8a9e-d96565189197-cnibin\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.961648 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/76d5e7c0-c9d4-47c8-8a9e-d96565189197-cnibin\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962106 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovnkube-config\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962306 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-kubelet\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962329 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-log-socket\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962386 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-log-socket\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962401 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-kubelet\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962345 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovn-node-metrics-cert\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962447 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-slash\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962463 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-openvswitch\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962482 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-slash\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962494 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-node-log\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962520 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-openvswitch\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962535 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-systemd-units\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962551 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-ovn\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962574 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-etc-openvswitch\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962595 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/76d5e7c0-c9d4-47c8-8a9e-d96565189197-cni-binary-copy\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962649 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-node-log\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.963073 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-ovn\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962686 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qmw5\" (UniqueName: \"kubernetes.io/projected/76d5e7c0-c9d4-47c8-8a9e-d96565189197-kube-api-access-9qmw5\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.963143 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-env-overrides\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.963162 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/76d5e7c0-c9d4-47c8-8a9e-d96565189197-system-cni-dir\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.963195 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-run-netns\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.963215 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-run-ovn-kubernetes\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.963248 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/76d5e7c0-c9d4-47c8-8a9e-d96565189197-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.963270 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-systemd\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.963300 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vmkd\" (UniqueName: \"kubernetes.io/projected/28f65c70-f59b-4393-b9c9-8163bb596f6c-kube-api-access-2vmkd\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.963320 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovnkube-script-lib\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.962679 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-etc-openvswitch\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.963939 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-systemd-units\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.964020 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/76d5e7c0-c9d4-47c8-8a9e-d96565189197-system-cni-dir\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.964089 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-run-ovn-kubernetes\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.964433 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/76d5e7c0-c9d4-47c8-8a9e-d96565189197-cni-binary-copy\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.964792 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-run-netns\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.964840 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/76d5e7c0-c9d4-47c8-8a9e-d96565189197-os-release\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.964875 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-cni-netd\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.964938 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-env-overrides\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.965069 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/76d5e7c0-c9d4-47c8-8a9e-d96565189197-os-release\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.965119 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-cni-netd\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.965254 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-systemd\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.965586 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/76d5e7c0-c9d4-47c8-8a9e-d96565189197-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.966037 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovnkube-script-lib\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.970501 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovn-node-metrics-cert\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.972066 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 05 06:55:18 crc kubenswrapper[4997]: I1205 06:55:18.992814 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.050922 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qmw5\" (UniqueName: \"kubernetes.io/projected/76d5e7c0-c9d4-47c8-8a9e-d96565189197-kube-api-access-9qmw5\") pod \"multus-additional-cni-plugins-5m9fs\" (UID: \"76d5e7c0-c9d4-47c8-8a9e-d96565189197\") " pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.066192 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vmkd\" (UniqueName: \"kubernetes.io/projected/28f65c70-f59b-4393-b9c9-8163bb596f6c-kube-api-access-2vmkd\") pod \"ovnkube-node-bchpk\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.080683 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.117140 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" Dec 05 06:55:19 crc kubenswrapper[4997]: W1205 06:55:19.128414 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76d5e7c0_c9d4_47c8_8a9e_d96565189197.slice/crio-f20e7f38816308e2987c335a9670f16108e6075fe35097a75c8c50fac10c61e8 WatchSource:0}: Error finding container f20e7f38816308e2987c335a9670f16108e6075fe35097a75c8c50fac10c61e8: Status 404 returned error can't find the container with id f20e7f38816308e2987c335a9670f16108e6075fe35097a75c8c50fac10c61e8 Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.129184 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.134957 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.187407 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: W1205 06:55:19.193652 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28f65c70_f59b_4393_b9c9_8163bb596f6c.slice/crio-897a7454e3968995ef11164249ce88aa5e11662142735a63b34b5d3056e89dbd WatchSource:0}: Error finding container 897a7454e3968995ef11164249ce88aa5e11662142735a63b34b5d3056e89dbd: Status 404 returned error can't find the container with id 897a7454e3968995ef11164249ce88aa5e11662142735a63b34b5d3056e89dbd Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.240019 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.265150 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.290241 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.326402 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.363972 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.368367 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:55:19 crc kubenswrapper[4997]: E1205 06:55:19.368507 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:55:21.368483802 +0000 UTC m=+21.897391063 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.368598 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:19 crc kubenswrapper[4997]: E1205 06:55:19.368707 4997 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 06:55:19 crc kubenswrapper[4997]: E1205 06:55:19.368747 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:21.368740549 +0000 UTC m=+21.897647810 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.405382 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.450447 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.469733 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.469773 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.469805 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:19 crc kubenswrapper[4997]: E1205 06:55:19.469917 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 06:55:19 crc kubenswrapper[4997]: E1205 06:55:19.469931 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 06:55:19 crc kubenswrapper[4997]: E1205 06:55:19.469941 4997 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:19 crc kubenswrapper[4997]: E1205 06:55:19.469936 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 06:55:19 crc kubenswrapper[4997]: E1205 06:55:19.469980 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 06:55:19 crc kubenswrapper[4997]: E1205 06:55:19.469991 4997 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:19 crc kubenswrapper[4997]: E1205 06:55:19.470007 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:21.469994371 +0000 UTC m=+21.998901632 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:19 crc kubenswrapper[4997]: E1205 06:55:19.469941 4997 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 06:55:19 crc kubenswrapper[4997]: E1205 06:55:19.470038 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:21.470023562 +0000 UTC m=+21.998930823 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:19 crc kubenswrapper[4997]: E1205 06:55:19.470154 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:21.470122765 +0000 UTC m=+21.999030026 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.482797 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.521343 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.565909 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.618770 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.624263 4997 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 05 06:55:19 crc kubenswrapper[4997]: W1205 06:55:19.624436 4997 reflector.go:484] object-"openshift-ovn-kubernetes"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Dec 05 06:55:19 crc kubenswrapper[4997]: W1205 06:55:19.624489 4997 reflector.go:484] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": Unexpected watch close - watch lasted less than a second and no items received Dec 05 06:55:19 crc kubenswrapper[4997]: W1205 06:55:19.624519 4997 reflector.go:484] object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 05 06:55:19 crc kubenswrapper[4997]: W1205 06:55:19.624572 4997 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovnkube-script-lib": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovnkube-script-lib": Unexpected watch close - watch lasted less than a second and no items received Dec 05 06:55:19 crc kubenswrapper[4997]: W1205 06:55:19.624645 4997 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": watch of *v1.Secret ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": Unexpected watch close - watch lasted less than a second and no items received Dec 05 06:55:19 crc kubenswrapper[4997]: W1205 06:55:19.624756 4997 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovnkube-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovnkube-config": Unexpected watch close - watch lasted less than a second and no items received Dec 05 06:55:19 crc kubenswrapper[4997]: W1205 06:55:19.624997 4997 reflector.go:484] object-"openshift-multus"/"default-cni-sysctl-allowlist": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"default-cni-sysctl-allowlist": Unexpected watch close - watch lasted less than a second and no items received Dec 05 06:55:19 crc kubenswrapper[4997]: W1205 06:55:19.625029 4997 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": Unexpected watch close - watch lasted less than a second and no items received Dec 05 06:55:19 crc kubenswrapper[4997]: W1205 06:55:19.625031 4997 reflector.go:484] object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.644655 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.683294 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.745938 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.748001 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.748045 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.748014 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:19 crc kubenswrapper[4997]: E1205 06:55:19.748146 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:19 crc kubenswrapper[4997]: E1205 06:55:19.748221 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:19 crc kubenswrapper[4997]: E1205 06:55:19.748344 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.752149 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.752826 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.753600 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.754227 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.754803 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.755299 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.755914 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.756449 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.757078 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.757583 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.758095 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.758762 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.759258 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.762031 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.762552 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.763407 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.764034 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.764435 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.767562 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.768228 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.768231 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.768724 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.769686 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.770076 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.771105 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.771561 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.772651 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.773260 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.773804 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.774801 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.775329 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.776161 4997 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.776260 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.778091 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.779131 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.779628 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.781115 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.782199 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.782795 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.786501 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.787308 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.787866 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.788898 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.789988 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.790709 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.791546 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.792280 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.793347 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.794126 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.795080 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.795658 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.796138 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.797184 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.798003 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.798995 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.824957 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.848191 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: E1205 06:55:19.855826 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28f65c70_f59b_4393_b9c9_8163bb596f6c.slice/crio-a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2.scope\": RecentStats: unable to find data in memory cache]" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.885683 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.906877 4997 generic.go:334] "Generic (PLEG): container finished" podID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerID="a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2" exitCode=0 Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.906972 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerDied","Data":"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2"} Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.907196 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerStarted","Data":"897a7454e3968995ef11164249ce88aa5e11662142735a63b34b5d3056e89dbd"} Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.909563 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31"} Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.910872 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" event={"ID":"76d5e7c0-c9d4-47c8-8a9e-d96565189197","Type":"ContainerStarted","Data":"a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55"} Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.910919 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" event={"ID":"76d5e7c0-c9d4-47c8-8a9e-d96565189197","Type":"ContainerStarted","Data":"f20e7f38816308e2987c335a9670f16108e6075fe35097a75c8c50fac10c61e8"} Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.923069 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:19 crc kubenswrapper[4997]: I1205 06:55:19.964121 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.002202 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.042996 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.081280 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.122852 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.162425 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.203458 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.241827 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.279034 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.324482 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.360778 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.402268 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.443273 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.452687 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.506221 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.512801 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.563631 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.601558 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.642126 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.680273 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.712598 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.742915 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.752889 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.772789 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.813337 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.847845 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.852557 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.908247 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.919696 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerStarted","Data":"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b"} Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.919732 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerStarted","Data":"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34"} Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.919743 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerStarted","Data":"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb"} Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.919752 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerStarted","Data":"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf"} Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.919760 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerStarted","Data":"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c"} Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.919767 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerStarted","Data":"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798"} Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.920887 4997 generic.go:334] "Generic (PLEG): container finished" podID="76d5e7c0-c9d4-47c8-8a9e-d96565189197" containerID="a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55" exitCode=0 Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.920959 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" event={"ID":"76d5e7c0-c9d4-47c8-8a9e-d96565189197","Type":"ContainerDied","Data":"a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55"} Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.922253 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942"} Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.943567 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.953351 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 05 06:55:20 crc kubenswrapper[4997]: I1205 06:55:20.997927 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.012076 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.061859 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.101913 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.135345 4997 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.137543 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.137579 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.137591 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.137721 4997 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.145191 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.194551 4997 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.194856 4997 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.196802 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.196843 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.196853 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.196870 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.196882 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:21Z","lastTransitionTime":"2025-12-05T06:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.214173 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.218113 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.218144 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.218154 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.218170 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.218181 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:21Z","lastTransitionTime":"2025-12-05T06:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.231155 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.231845 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.234705 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.234743 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.234762 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.234779 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.234791 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:21Z","lastTransitionTime":"2025-12-05T06:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.249715 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.253253 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.253291 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.253299 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.253314 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.253325 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:21Z","lastTransitionTime":"2025-12-05T06:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.263308 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.266306 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.269686 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.269709 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.269717 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.269730 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.269743 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:21Z","lastTransitionTime":"2025-12-05T06:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.284939 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.285083 4997 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.286708 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.286737 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.286745 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.286760 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.286771 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:21Z","lastTransitionTime":"2025-12-05T06:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.302601 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.340871 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.380599 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.389124 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.389176 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.389185 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.389201 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.389210 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:21Z","lastTransitionTime":"2025-12-05T06:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.390160 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.390245 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.390504 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:55:25.390296384 +0000 UTC m=+25.919203645 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.390538 4997 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.390588 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:25.390574481 +0000 UTC m=+25.919481732 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.420515 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.464601 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.491026 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.491073 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.491094 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.491191 4997 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.491422 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:25.491409652 +0000 UTC m=+26.020316913 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.491448 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.491474 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.491485 4997 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.491505 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.491534 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.491541 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:25.491525135 +0000 UTC m=+26.020432396 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.491547 4997 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.491601 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:25.491586777 +0000 UTC m=+26.020494038 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.492266 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.492295 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.492303 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.492323 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.492334 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:21Z","lastTransitionTime":"2025-12-05T06:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.508894 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.542502 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.582930 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.595294 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.595324 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.595333 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.595348 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.595357 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:21Z","lastTransitionTime":"2025-12-05T06:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.623172 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.662049 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.697381 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.697420 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.697430 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.697445 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.697455 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:21Z","lastTransitionTime":"2025-12-05T06:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.701546 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.748561 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.748596 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.748568 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.748669 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.748762 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:21 crc kubenswrapper[4997]: E1205 06:55:21.748856 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.757659 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.780996 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.799199 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.799227 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.799235 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.799248 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.799258 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:21Z","lastTransitionTime":"2025-12-05T06:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.818861 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.860093 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.901757 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.901791 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.901799 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.901822 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.901835 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:21Z","lastTransitionTime":"2025-12-05T06:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.931431 4997 generic.go:334] "Generic (PLEG): container finished" podID="76d5e7c0-c9d4-47c8-8a9e-d96565189197" containerID="880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5" exitCode=0 Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.931520 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" event={"ID":"76d5e7c0-c9d4-47c8-8a9e-d96565189197","Type":"ContainerDied","Data":"880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5"} Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.948125 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.962122 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:21 crc kubenswrapper[4997]: I1205 06:55:21.981924 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.004366 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.004414 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.004424 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.004439 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.004450 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:22Z","lastTransitionTime":"2025-12-05T06:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.022406 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.065764 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.106151 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.106551 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.106579 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.106587 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.106603 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.106628 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:22Z","lastTransitionTime":"2025-12-05T06:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.140655 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.182642 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.209025 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.209068 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.209080 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.209098 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.209109 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:22Z","lastTransitionTime":"2025-12-05T06:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.222468 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.262292 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.302694 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.311396 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.311424 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.311433 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.311449 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.311458 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:22Z","lastTransitionTime":"2025-12-05T06:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.342138 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.381690 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.414553 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.414580 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.414588 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.414605 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.414649 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:22Z","lastTransitionTime":"2025-12-05T06:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.523266 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.523345 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.523360 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.523379 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.523393 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:22Z","lastTransitionTime":"2025-12-05T06:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.628356 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.628399 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.628412 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.628494 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.628508 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:22Z","lastTransitionTime":"2025-12-05T06:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.731388 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.731433 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.731445 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.731464 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.731477 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:22Z","lastTransitionTime":"2025-12-05T06:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.834949 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.834985 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.834993 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.835007 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.835016 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:22Z","lastTransitionTime":"2025-12-05T06:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.938229 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.938298 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.938325 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.938357 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.938381 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:22Z","lastTransitionTime":"2025-12-05T06:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.941798 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerStarted","Data":"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c"} Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.945340 4997 generic.go:334] "Generic (PLEG): container finished" podID="76d5e7c0-c9d4-47c8-8a9e-d96565189197" containerID="1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550" exitCode=0 Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.945519 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" event={"ID":"76d5e7c0-c9d4-47c8-8a9e-d96565189197","Type":"ContainerDied","Data":"1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550"} Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.966423 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.979232 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:22 crc kubenswrapper[4997]: I1205 06:55:22.992066 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.006878 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:23Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.022228 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:23Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.034421 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:23Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.040067 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.040095 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.040105 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.040118 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.040128 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:23Z","lastTransitionTime":"2025-12-05T06:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.048329 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:23Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.061312 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:23Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.075429 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:23Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.109954 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:23Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.130745 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:23Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.142674 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.142975 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.142984 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.142996 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.143005 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:23Z","lastTransitionTime":"2025-12-05T06:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.160341 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:23Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.181725 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:23Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.245119 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.245165 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.245174 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.245188 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.245198 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:23Z","lastTransitionTime":"2025-12-05T06:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.347118 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.347154 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.347172 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.347187 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.347197 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:23Z","lastTransitionTime":"2025-12-05T06:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.449713 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.449771 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.449788 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.449811 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.449828 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:23Z","lastTransitionTime":"2025-12-05T06:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.552603 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.552661 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.552674 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.552714 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.552725 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:23Z","lastTransitionTime":"2025-12-05T06:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.655511 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.655550 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.655559 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.655576 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.655585 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:23Z","lastTransitionTime":"2025-12-05T06:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.748437 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.748539 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.748444 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:23 crc kubenswrapper[4997]: E1205 06:55:23.748609 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:23 crc kubenswrapper[4997]: E1205 06:55:23.748773 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:23 crc kubenswrapper[4997]: E1205 06:55:23.748878 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.758739 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.758802 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.758813 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.758866 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.758881 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:23Z","lastTransitionTime":"2025-12-05T06:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.861189 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.861227 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.861238 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.861255 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.861265 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:23Z","lastTransitionTime":"2025-12-05T06:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.950125 4997 generic.go:334] "Generic (PLEG): container finished" podID="76d5e7c0-c9d4-47c8-8a9e-d96565189197" containerID="90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98" exitCode=0 Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.950166 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" event={"ID":"76d5e7c0-c9d4-47c8-8a9e-d96565189197","Type":"ContainerDied","Data":"90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98"} Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.963897 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.963928 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.963936 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.963949 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.963959 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:23Z","lastTransitionTime":"2025-12-05T06:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.974027 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:23Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:23 crc kubenswrapper[4997]: I1205 06:55:23.988534 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:23Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.008525 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.026337 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.042154 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.055131 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-2jj4c"] Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.055693 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2jj4c" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.058764 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.059276 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.059574 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.060858 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.067137 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.067131 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.067204 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.067253 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.067275 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.067311 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:24Z","lastTransitionTime":"2025-12-05T06:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.084545 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.096597 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.107570 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.118822 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.120144 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2df2c8f1-3f95-4b85-825f-d79671f90630-host\") pod \"node-ca-2jj4c\" (UID: \"2df2c8f1-3f95-4b85-825f-d79671f90630\") " pod="openshift-image-registry/node-ca-2jj4c" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.120182 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2df2c8f1-3f95-4b85-825f-d79671f90630-serviceca\") pod \"node-ca-2jj4c\" (UID: \"2df2c8f1-3f95-4b85-825f-d79671f90630\") " pod="openshift-image-registry/node-ca-2jj4c" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.120288 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2m8r\" (UniqueName: \"kubernetes.io/projected/2df2c8f1-3f95-4b85-825f-d79671f90630-kube-api-access-c2m8r\") pod \"node-ca-2jj4c\" (UID: \"2df2c8f1-3f95-4b85-825f-d79671f90630\") " pod="openshift-image-registry/node-ca-2jj4c" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.135489 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.150609 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.169922 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.170369 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.170378 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.170390 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.170400 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:24Z","lastTransitionTime":"2025-12-05T06:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.173358 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.188053 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.202767 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.220989 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2m8r\" (UniqueName: \"kubernetes.io/projected/2df2c8f1-3f95-4b85-825f-d79671f90630-kube-api-access-c2m8r\") pod \"node-ca-2jj4c\" (UID: \"2df2c8f1-3f95-4b85-825f-d79671f90630\") " pod="openshift-image-registry/node-ca-2jj4c" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.221089 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2df2c8f1-3f95-4b85-825f-d79671f90630-host\") pod \"node-ca-2jj4c\" (UID: \"2df2c8f1-3f95-4b85-825f-d79671f90630\") " pod="openshift-image-registry/node-ca-2jj4c" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.221110 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2df2c8f1-3f95-4b85-825f-d79671f90630-serviceca\") pod \"node-ca-2jj4c\" (UID: \"2df2c8f1-3f95-4b85-825f-d79671f90630\") " pod="openshift-image-registry/node-ca-2jj4c" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.221258 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2df2c8f1-3f95-4b85-825f-d79671f90630-host\") pod \"node-ca-2jj4c\" (UID: \"2df2c8f1-3f95-4b85-825f-d79671f90630\") " pod="openshift-image-registry/node-ca-2jj4c" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.221789 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.222108 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2df2c8f1-3f95-4b85-825f-d79671f90630-serviceca\") pod \"node-ca-2jj4c\" (UID: \"2df2c8f1-3f95-4b85-825f-d79671f90630\") " pod="openshift-image-registry/node-ca-2jj4c" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.233101 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.244481 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2m8r\" (UniqueName: \"kubernetes.io/projected/2df2c8f1-3f95-4b85-825f-d79671f90630-kube-api-access-c2m8r\") pod \"node-ca-2jj4c\" (UID: \"2df2c8f1-3f95-4b85-825f-d79671f90630\") " pod="openshift-image-registry/node-ca-2jj4c" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.254995 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.272186 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.272912 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.272975 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.272986 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.273008 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.273030 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:24Z","lastTransitionTime":"2025-12-05T06:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.291011 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.305222 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.317337 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.332379 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.348002 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.366060 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.373355 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2jj4c" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.376267 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.376301 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.376310 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.376327 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.376337 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:24Z","lastTransitionTime":"2025-12-05T06:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.378828 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.392548 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.480396 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.480449 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.480461 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.480481 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.480495 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:24Z","lastTransitionTime":"2025-12-05T06:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.584184 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.584245 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.584259 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.584286 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.584305 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:24Z","lastTransitionTime":"2025-12-05T06:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.688226 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.688297 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.688319 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.688353 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.688377 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:24Z","lastTransitionTime":"2025-12-05T06:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.791670 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.791701 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.791709 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.791724 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.791734 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:24Z","lastTransitionTime":"2025-12-05T06:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.894893 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.894980 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.894998 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.895314 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.895825 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:24Z","lastTransitionTime":"2025-12-05T06:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.959696 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2jj4c" event={"ID":"2df2c8f1-3f95-4b85-825f-d79671f90630","Type":"ContainerStarted","Data":"3ca270a30472a65b1739d46c8f2db09aa3602fd696a5d4fe9259cc2a32c02f18"} Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.964704 4997 generic.go:334] "Generic (PLEG): container finished" podID="76d5e7c0-c9d4-47c8-8a9e-d96565189197" containerID="871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e" exitCode=0 Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.964774 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" event={"ID":"76d5e7c0-c9d4-47c8-8a9e-d96565189197","Type":"ContainerDied","Data":"871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e"} Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.985015 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:24Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.999380 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.999429 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.999439 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.999460 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:24 crc kubenswrapper[4997]: I1205 06:55:24.999472 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:24Z","lastTransitionTime":"2025-12-05T06:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.002840 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:25Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.021285 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:25Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.040023 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:25Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.055910 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:25Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.071712 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:25Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.089824 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:25Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.102196 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.102239 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.102248 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.102262 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.102274 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:25Z","lastTransitionTime":"2025-12-05T06:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.106922 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:25Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.121124 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:25Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.133494 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:25Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.146399 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:25Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.165511 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:25Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.184286 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:25Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.200369 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:25Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.206383 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.206626 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.206644 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.206665 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.206679 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:25Z","lastTransitionTime":"2025-12-05T06:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.309462 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.309507 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.309518 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.309537 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.309549 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:25Z","lastTransitionTime":"2025-12-05T06:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.411522 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.411550 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.411558 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.411573 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.411581 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:25Z","lastTransitionTime":"2025-12-05T06:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.433432 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.433519 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:25 crc kubenswrapper[4997]: E1205 06:55:25.433640 4997 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 06:55:25 crc kubenswrapper[4997]: E1205 06:55:25.433686 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:33.433669991 +0000 UTC m=+33.962577252 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 06:55:25 crc kubenswrapper[4997]: E1205 06:55:25.433729 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:55:33.433723743 +0000 UTC m=+33.962631004 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.514223 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.514267 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.514279 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.514294 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.514305 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:25Z","lastTransitionTime":"2025-12-05T06:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.534942 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.534987 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.535006 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:25 crc kubenswrapper[4997]: E1205 06:55:25.535114 4997 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 06:55:25 crc kubenswrapper[4997]: E1205 06:55:25.535167 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:33.535147879 +0000 UTC m=+34.064055150 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 06:55:25 crc kubenswrapper[4997]: E1205 06:55:25.535249 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 06:55:25 crc kubenswrapper[4997]: E1205 06:55:25.535299 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 06:55:25 crc kubenswrapper[4997]: E1205 06:55:25.535319 4997 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:25 crc kubenswrapper[4997]: E1205 06:55:25.535311 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 06:55:25 crc kubenswrapper[4997]: E1205 06:55:25.535385 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 06:55:25 crc kubenswrapper[4997]: E1205 06:55:25.535408 4997 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:25 crc kubenswrapper[4997]: E1205 06:55:25.535413 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:33.535384225 +0000 UTC m=+34.064291666 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:25 crc kubenswrapper[4997]: E1205 06:55:25.535514 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:33.535478128 +0000 UTC m=+34.064385549 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.618103 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.618153 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.618348 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.618368 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.618383 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:25Z","lastTransitionTime":"2025-12-05T06:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.721825 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.721888 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.721906 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.721931 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.721945 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:25Z","lastTransitionTime":"2025-12-05T06:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.748860 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.749035 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:25 crc kubenswrapper[4997]: E1205 06:55:25.749089 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.749169 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:25 crc kubenswrapper[4997]: E1205 06:55:25.749457 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:25 crc kubenswrapper[4997]: E1205 06:55:25.749316 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.826813 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.827817 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.827888 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.827944 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.827969 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:25Z","lastTransitionTime":"2025-12-05T06:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.937464 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.937744 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.937780 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.937809 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.937836 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:25Z","lastTransitionTime":"2025-12-05T06:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.971587 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2jj4c" event={"ID":"2df2c8f1-3f95-4b85-825f-d79671f90630","Type":"ContainerStarted","Data":"44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6"} Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.974841 4997 generic.go:334] "Generic (PLEG): container finished" podID="76d5e7c0-c9d4-47c8-8a9e-d96565189197" containerID="72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c" exitCode=0 Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.975022 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" event={"ID":"76d5e7c0-c9d4-47c8-8a9e-d96565189197","Type":"ContainerDied","Data":"72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c"} Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.984091 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerStarted","Data":"7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185"} Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.984379 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.984537 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:25 crc kubenswrapper[4997]: I1205 06:55:25.985825 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:25Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.003253 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.021016 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.043807 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.043866 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.043880 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.043900 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.043917 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:26Z","lastTransitionTime":"2025-12-05T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.045826 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.062360 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.082280 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.102445 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.120938 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.133816 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.134421 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.136501 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.148522 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.148744 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.148769 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.148799 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.148818 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:26Z","lastTransitionTime":"2025-12-05T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.150317 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.166205 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.184147 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.206025 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.221704 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.235659 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.247390 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.251683 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.251740 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.251753 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.251781 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.251796 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:26Z","lastTransitionTime":"2025-12-05T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.263152 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.284310 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.299491 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.312876 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.331745 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.349327 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.354193 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.354247 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.354263 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.354292 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.354309 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:26Z","lastTransitionTime":"2025-12-05T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.365019 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.381943 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.399769 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.419774 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.431679 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.451973 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:26Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.457172 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.457350 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.457421 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.457484 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.457549 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:26Z","lastTransitionTime":"2025-12-05T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.561015 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.561070 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.561086 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.561107 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.561122 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:26Z","lastTransitionTime":"2025-12-05T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.665840 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.665917 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.665936 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.665968 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.665988 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:26Z","lastTransitionTime":"2025-12-05T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.768554 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.768581 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.768589 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.768602 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.768627 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:26Z","lastTransitionTime":"2025-12-05T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.870940 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.870981 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.870993 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.871007 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.871017 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:26Z","lastTransitionTime":"2025-12-05T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.974758 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.974809 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.974823 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.974843 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:26 crc kubenswrapper[4997]: I1205 06:55:26.974856 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:26Z","lastTransitionTime":"2025-12-05T06:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:26.999951 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" event={"ID":"76d5e7c0-c9d4-47c8-8a9e-d96565189197","Type":"ContainerStarted","Data":"d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf"} Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.000081 4997 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.020816 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:27Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.036590 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:27Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.058396 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:27Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.078547 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.078655 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.078684 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.078713 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.078733 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:27Z","lastTransitionTime":"2025-12-05T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.090209 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:27Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.113114 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:27Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.131375 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:27Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.145080 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:27Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.169399 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:27Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.181602 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.181716 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.181736 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.181762 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.181779 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:27Z","lastTransitionTime":"2025-12-05T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.183521 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:27Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.195932 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:27Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.217253 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:27Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.239300 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:27Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.263117 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:27Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.278419 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:27Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.284067 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.284109 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.284121 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.284139 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.284151 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:27Z","lastTransitionTime":"2025-12-05T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.387086 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.387132 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.387141 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.387157 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.387166 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:27Z","lastTransitionTime":"2025-12-05T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.489494 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.489531 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.489545 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.489564 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.489574 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:27Z","lastTransitionTime":"2025-12-05T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.591839 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.591902 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.591914 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.591940 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.591954 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:27Z","lastTransitionTime":"2025-12-05T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.694593 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.694679 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.694693 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.694713 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.694724 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:27Z","lastTransitionTime":"2025-12-05T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.749831 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.749935 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.749934 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:27 crc kubenswrapper[4997]: E1205 06:55:27.750082 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:27 crc kubenswrapper[4997]: E1205 06:55:27.750991 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:27 crc kubenswrapper[4997]: E1205 06:55:27.751202 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.798297 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.798333 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.798343 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.798357 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.798367 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:27Z","lastTransitionTime":"2025-12-05T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.901746 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.901788 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.901800 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.901817 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:27 crc kubenswrapper[4997]: I1205 06:55:27.901829 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:27Z","lastTransitionTime":"2025-12-05T06:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.002485 4997 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.003473 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.003537 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.003550 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.003570 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.003580 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:28Z","lastTransitionTime":"2025-12-05T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.106378 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.106419 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.106428 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.106442 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.106452 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:28Z","lastTransitionTime":"2025-12-05T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.209166 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.209242 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.209263 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.209296 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.209316 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:28Z","lastTransitionTime":"2025-12-05T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.312453 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.312495 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.312508 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.312528 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.312542 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:28Z","lastTransitionTime":"2025-12-05T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.414937 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.414994 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.415010 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.415040 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.415061 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:28Z","lastTransitionTime":"2025-12-05T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.518606 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.518696 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.518713 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.518736 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.518752 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:28Z","lastTransitionTime":"2025-12-05T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.622375 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.622432 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.622452 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.622475 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.622486 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:28Z","lastTransitionTime":"2025-12-05T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.725530 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.725661 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.725684 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.725717 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.725739 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:28Z","lastTransitionTime":"2025-12-05T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.829232 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.829301 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.829323 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.829358 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.829378 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:28Z","lastTransitionTime":"2025-12-05T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.932381 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.932762 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.932897 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.933021 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:28 crc kubenswrapper[4997]: I1205 06:55:28.933138 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:28Z","lastTransitionTime":"2025-12-05T06:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.010993 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovnkube-controller/0.log" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.014988 4997 generic.go:334] "Generic (PLEG): container finished" podID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerID="7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185" exitCode=1 Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.015215 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerDied","Data":"7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185"} Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.016580 4997 scope.go:117] "RemoveContainer" containerID="7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.038164 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.038198 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.038209 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.038225 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.038236 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:29Z","lastTransitionTime":"2025-12-05T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.041182 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.062959 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.079759 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.102137 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:28Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 06:55:27.880124 6256 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 06:55:27.880257 6256 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 06:55:27.880321 6256 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 06:55:27.880335 6256 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 06:55:27.880362 6256 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 06:55:27.880378 6256 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 06:55:27.880389 6256 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 06:55:27.880399 6256 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 06:55:27.880414 6256 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 06:55:27.880420 6256 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 06:55:27.880423 6256 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 06:55:27.880434 6256 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 06:55:27.880445 6256 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 06:55:27.880516 6256 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 06:55:27.880555 6256 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.124594 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.140920 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.141017 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.141042 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.141081 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.141111 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:29Z","lastTransitionTime":"2025-12-05T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.141390 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.158161 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.174921 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.190694 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.207687 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.224972 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.245249 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.245304 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.245315 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.245336 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.245349 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:29Z","lastTransitionTime":"2025-12-05T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.250846 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.264181 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.279080 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.348422 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.348488 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.348511 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.348542 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.348563 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:29Z","lastTransitionTime":"2025-12-05T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.451621 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.451670 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.451680 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.451698 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.451715 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:29Z","lastTransitionTime":"2025-12-05T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.554030 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.554345 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.554438 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.554505 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.554565 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:29Z","lastTransitionTime":"2025-12-05T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.657856 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.657890 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.657897 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.657912 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.657923 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:29Z","lastTransitionTime":"2025-12-05T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.748409 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.748502 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.748569 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:29 crc kubenswrapper[4997]: E1205 06:55:29.749003 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:29 crc kubenswrapper[4997]: E1205 06:55:29.749104 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:29 crc kubenswrapper[4997]: E1205 06:55:29.749245 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.760554 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.761004 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.761184 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.761439 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.761639 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:29Z","lastTransitionTime":"2025-12-05T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.763181 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.779193 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.805015 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.827276 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.850075 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.864506 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.864571 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.864585 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.864630 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.864646 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:29Z","lastTransitionTime":"2025-12-05T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.872391 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.885071 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.903721 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.918133 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.934683 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.953386 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:28Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 06:55:27.880124 6256 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 06:55:27.880257 6256 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 06:55:27.880321 6256 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 06:55:27.880335 6256 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 06:55:27.880362 6256 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 06:55:27.880378 6256 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 06:55:27.880389 6256 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 06:55:27.880399 6256 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 06:55:27.880414 6256 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 06:55:27.880420 6256 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 06:55:27.880423 6256 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 06:55:27.880434 6256 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 06:55:27.880445 6256 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 06:55:27.880516 6256 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 06:55:27.880555 6256 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.967410 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.968427 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.968551 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.968655 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.968761 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.968845 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:29Z","lastTransitionTime":"2025-12-05T06:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.982493 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:29 crc kubenswrapper[4997]: I1205 06:55:29.995702 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.032288 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovnkube-controller/0.log" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.035493 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerStarted","Data":"cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7"} Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.035669 4997 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.052756 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.064370 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.071820 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.071990 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.072057 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.072150 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.072224 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:30Z","lastTransitionTime":"2025-12-05T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.078575 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.090781 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.106217 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.116254 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.127872 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.141003 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.156458 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.174551 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.174725 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.174907 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.175161 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.175228 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:30Z","lastTransitionTime":"2025-12-05T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.181810 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:28Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 06:55:27.880124 6256 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 06:55:27.880257 6256 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 06:55:27.880321 6256 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 06:55:27.880335 6256 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 06:55:27.880362 6256 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 06:55:27.880378 6256 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 06:55:27.880389 6256 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 06:55:27.880399 6256 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 06:55:27.880414 6256 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 06:55:27.880420 6256 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 06:55:27.880423 6256 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 06:55:27.880434 6256 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 06:55:27.880445 6256 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 06:55:27.880516 6256 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 06:55:27.880555 6256 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.195505 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.210552 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.223877 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.243093 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.277682 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.277711 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.277722 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.277735 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.277744 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:30Z","lastTransitionTime":"2025-12-05T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.380649 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.380682 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.380693 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.380708 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.380718 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:30Z","lastTransitionTime":"2025-12-05T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.483659 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.483739 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.483763 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.483800 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.483828 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:30Z","lastTransitionTime":"2025-12-05T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.586386 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.586469 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.586502 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.586539 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.586559 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:30Z","lastTransitionTime":"2025-12-05T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.673337 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.690520 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.690559 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.690568 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.690588 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.690598 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:30Z","lastTransitionTime":"2025-12-05T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.692303 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.708945 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.723386 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.736290 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.747566 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.760839 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.775164 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.788108 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.793867 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.793917 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.793932 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.793954 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.793970 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:30Z","lastTransitionTime":"2025-12-05T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.801033 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.812290 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.827562 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.845421 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.867123 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:28Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 06:55:27.880124 6256 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 06:55:27.880257 6256 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 06:55:27.880321 6256 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 06:55:27.880335 6256 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 06:55:27.880362 6256 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 06:55:27.880378 6256 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 06:55:27.880389 6256 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 06:55:27.880399 6256 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 06:55:27.880414 6256 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 06:55:27.880420 6256 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 06:55:27.880423 6256 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 06:55:27.880434 6256 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 06:55:27.880445 6256 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 06:55:27.880516 6256 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 06:55:27.880555 6256 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.881345 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:30Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.896190 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.896216 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.896225 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.896240 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.896250 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:30Z","lastTransitionTime":"2025-12-05T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.999586 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.999645 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.999655 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.999670 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:30 crc kubenswrapper[4997]: I1205 06:55:30.999683 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:30Z","lastTransitionTime":"2025-12-05T06:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.041677 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovnkube-controller/1.log" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.042311 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovnkube-controller/0.log" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.045455 4997 generic.go:334] "Generic (PLEG): container finished" podID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerID="cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7" exitCode=1 Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.045520 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerDied","Data":"cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7"} Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.045805 4997 scope.go:117] "RemoveContainer" containerID="7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.046559 4997 scope.go:117] "RemoveContainer" containerID="cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7" Dec 05 06:55:31 crc kubenswrapper[4997]: E1205 06:55:31.046761 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.060312 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.074405 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.092581 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.102454 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.102502 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.102512 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.102533 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.102542 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:31Z","lastTransitionTime":"2025-12-05T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.118909 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:28Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 06:55:27.880124 6256 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 06:55:27.880257 6256 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 06:55:27.880321 6256 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 06:55:27.880335 6256 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 06:55:27.880362 6256 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 06:55:27.880378 6256 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 06:55:27.880389 6256 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 06:55:27.880399 6256 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 06:55:27.880414 6256 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 06:55:27.880420 6256 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 06:55:27.880423 6256 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 06:55:27.880434 6256 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 06:55:27.880445 6256 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 06:55:27.880516 6256 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 06:55:27.880555 6256 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"message\\\":\\\"] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-5m9fs\\\\nI1205 06:55:29.888179 6428 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-bchpk after 0 failed attempt(s)\\\\nI1205 06:55:29.888181 6428 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-5m9fs in node crc\\\\nI1205 06:55:29.888185 6428 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-bchpk\\\\nI1205 06:55:29.888188 6428 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-5m9fs after 0 failed attempt(s)\\\\nF1205 06:55:29.887701 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.137050 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.152429 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.163736 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.175790 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.186388 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.201431 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.205021 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.205176 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.205310 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.205396 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.205469 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:31Z","lastTransitionTime":"2025-12-05T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.217264 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.231589 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.244565 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.256318 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.307964 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.308168 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.308256 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.308328 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.308409 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:31Z","lastTransitionTime":"2025-12-05T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.412243 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.412312 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.412331 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.412360 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.412386 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:31Z","lastTransitionTime":"2025-12-05T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.515189 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.515221 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.515231 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.515245 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.515256 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:31Z","lastTransitionTime":"2025-12-05T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.532555 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5"] Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.533002 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.535775 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.535911 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.555426 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.577580 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:28Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 06:55:27.880124 6256 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 06:55:27.880257 6256 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 06:55:27.880321 6256 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 06:55:27.880335 6256 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 06:55:27.880362 6256 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 06:55:27.880378 6256 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 06:55:27.880389 6256 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 06:55:27.880399 6256 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 06:55:27.880414 6256 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 06:55:27.880420 6256 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 06:55:27.880423 6256 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 06:55:27.880434 6256 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 06:55:27.880445 6256 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 06:55:27.880516 6256 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 06:55:27.880555 6256 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"message\\\":\\\"] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-5m9fs\\\\nI1205 06:55:29.888179 6428 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-bchpk after 0 failed attempt(s)\\\\nI1205 06:55:29.888181 6428 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-5m9fs in node crc\\\\nI1205 06:55:29.888185 6428 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-bchpk\\\\nI1205 06:55:29.888188 6428 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-5m9fs after 0 failed attempt(s)\\\\nF1205 06:55:29.887701 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.592419 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.604290 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.604330 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.604338 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.604353 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.604363 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:31Z","lastTransitionTime":"2025-12-05T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.608670 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: E1205 06:55:31.628123 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.633089 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.633688 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.633777 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.633861 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.633940 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:31Z","lastTransitionTime":"2025-12-05T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.636855 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: E1205 06:55:31.650054 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.653606 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.655047 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.655084 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.655101 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.655121 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.655133 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:31Z","lastTransitionTime":"2025-12-05T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:31 crc kubenswrapper[4997]: E1205 06:55:31.668710 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.674883 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.674928 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.674939 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.674959 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.674972 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:31Z","lastTransitionTime":"2025-12-05T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.679381 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: E1205 06:55:31.686781 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.690644 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.690667 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.690677 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.690692 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.690702 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:31Z","lastTransitionTime":"2025-12-05T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.697274 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.700107 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1d2f43a-f27f-4739-9f27-095ef2ff7ce0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-k85r5\" (UID: \"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.700283 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1d2f43a-f27f-4739-9f27-095ef2ff7ce0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-k85r5\" (UID: \"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.700448 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2kb4\" (UniqueName: \"kubernetes.io/projected/d1d2f43a-f27f-4739-9f27-095ef2ff7ce0-kube-api-access-g2kb4\") pod \"ovnkube-control-plane-749d76644c-k85r5\" (UID: \"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.700516 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1d2f43a-f27f-4739-9f27-095ef2ff7ce0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-k85r5\" (UID: \"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" Dec 05 06:55:31 crc kubenswrapper[4997]: E1205 06:55:31.702689 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: E1205 06:55:31.702795 4997 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.704843 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.704894 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.704905 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.704921 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.704932 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:31Z","lastTransitionTime":"2025-12-05T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.711143 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.725206 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.740509 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.748956 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.749075 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:31 crc kubenswrapper[4997]: E1205 06:55:31.749190 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.749213 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:31 crc kubenswrapper[4997]: E1205 06:55:31.749298 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:31 crc kubenswrapper[4997]: E1205 06:55:31.749348 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.759406 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.771461 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.783288 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.799753 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:31Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.801017 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1d2f43a-f27f-4739-9f27-095ef2ff7ce0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-k85r5\" (UID: \"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.801079 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1d2f43a-f27f-4739-9f27-095ef2ff7ce0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-k85r5\" (UID: \"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.801115 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2kb4\" (UniqueName: \"kubernetes.io/projected/d1d2f43a-f27f-4739-9f27-095ef2ff7ce0-kube-api-access-g2kb4\") pod \"ovnkube-control-plane-749d76644c-k85r5\" (UID: \"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.801133 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1d2f43a-f27f-4739-9f27-095ef2ff7ce0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-k85r5\" (UID: \"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.801780 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1d2f43a-f27f-4739-9f27-095ef2ff7ce0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-k85r5\" (UID: \"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.801948 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1d2f43a-f27f-4739-9f27-095ef2ff7ce0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-k85r5\" (UID: \"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.807083 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.807108 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.807117 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.807131 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.807142 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:31Z","lastTransitionTime":"2025-12-05T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.811228 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1d2f43a-f27f-4739-9f27-095ef2ff7ce0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-k85r5\" (UID: \"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.817800 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2kb4\" (UniqueName: \"kubernetes.io/projected/d1d2f43a-f27f-4739-9f27-095ef2ff7ce0-kube-api-access-g2kb4\") pod \"ovnkube-control-plane-749d76644c-k85r5\" (UID: \"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.846368 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" Dec 05 06:55:31 crc kubenswrapper[4997]: W1205 06:55:31.861293 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1d2f43a_f27f_4739_9f27_095ef2ff7ce0.slice/crio-f32328926190a28cafdfa58b1fc869263fc64b35a6c3183ada24980e10f4fd82 WatchSource:0}: Error finding container f32328926190a28cafdfa58b1fc869263fc64b35a6c3183ada24980e10f4fd82: Status 404 returned error can't find the container with id f32328926190a28cafdfa58b1fc869263fc64b35a6c3183ada24980e10f4fd82 Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.908737 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.908773 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.908783 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.908798 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:31 crc kubenswrapper[4997]: I1205 06:55:31.908807 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:31Z","lastTransitionTime":"2025-12-05T06:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.011711 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.011749 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.011759 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.011776 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.011793 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:32Z","lastTransitionTime":"2025-12-05T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.050946 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovnkube-controller/1.log" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.055596 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" event={"ID":"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0","Type":"ContainerStarted","Data":"f32328926190a28cafdfa58b1fc869263fc64b35a6c3183ada24980e10f4fd82"} Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.114164 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.114207 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.114214 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.114229 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.114238 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:32Z","lastTransitionTime":"2025-12-05T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.216554 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.216928 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.216939 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.216953 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.216963 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:32Z","lastTransitionTime":"2025-12-05T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.291865 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-b2vlg"] Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.292278 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:32 crc kubenswrapper[4997]: E1205 06:55:32.292328 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.305149 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:32Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.318430 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:32Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.327632 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.327706 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.327718 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.327745 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.327754 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:32Z","lastTransitionTime":"2025-12-05T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.338702 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:32Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.355816 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:32Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.374481 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:32Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.386409 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:32Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.399550 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:32Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.406116 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs\") pod \"network-metrics-daemon-b2vlg\" (UID: \"81f6790f-312b-42ce-9bd4-58bac09ce615\") " pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.406175 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bk5c\" (UniqueName: \"kubernetes.io/projected/81f6790f-312b-42ce-9bd4-58bac09ce615-kube-api-access-6bk5c\") pod \"network-metrics-daemon-b2vlg\" (UID: \"81f6790f-312b-42ce-9bd4-58bac09ce615\") " pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.422758 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:32Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.430076 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.430311 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.430381 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.430447 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.430517 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:32Z","lastTransitionTime":"2025-12-05T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.443041 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:28Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 06:55:27.880124 6256 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 06:55:27.880257 6256 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 06:55:27.880321 6256 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 06:55:27.880335 6256 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 06:55:27.880362 6256 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 06:55:27.880378 6256 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 06:55:27.880389 6256 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 06:55:27.880399 6256 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 06:55:27.880414 6256 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 06:55:27.880420 6256 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 06:55:27.880423 6256 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 06:55:27.880434 6256 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 06:55:27.880445 6256 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 06:55:27.880516 6256 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 06:55:27.880555 6256 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"message\\\":\\\"] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-5m9fs\\\\nI1205 06:55:29.888179 6428 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-bchpk after 0 failed attempt(s)\\\\nI1205 06:55:29.888181 6428 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-5m9fs in node crc\\\\nI1205 06:55:29.888185 6428 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-bchpk\\\\nI1205 06:55:29.888188 6428 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-5m9fs after 0 failed attempt(s)\\\\nF1205 06:55:29.887701 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:32Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.462392 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:32Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.476784 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:32Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.489150 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:32Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.500084 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:32Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.507753 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs\") pod \"network-metrics-daemon-b2vlg\" (UID: \"81f6790f-312b-42ce-9bd4-58bac09ce615\") " pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.507805 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bk5c\" (UniqueName: \"kubernetes.io/projected/81f6790f-312b-42ce-9bd4-58bac09ce615-kube-api-access-6bk5c\") pod \"network-metrics-daemon-b2vlg\" (UID: \"81f6790f-312b-42ce-9bd4-58bac09ce615\") " pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:32 crc kubenswrapper[4997]: E1205 06:55:32.507910 4997 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 06:55:32 crc kubenswrapper[4997]: E1205 06:55:32.507992 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs podName:81f6790f-312b-42ce-9bd4-58bac09ce615 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:33.007973218 +0000 UTC m=+33.536880479 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs") pod "network-metrics-daemon-b2vlg" (UID: "81f6790f-312b-42ce-9bd4-58bac09ce615") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.514583 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:32Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.525997 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bk5c\" (UniqueName: \"kubernetes.io/projected/81f6790f-312b-42ce-9bd4-58bac09ce615-kube-api-access-6bk5c\") pod \"network-metrics-daemon-b2vlg\" (UID: \"81f6790f-312b-42ce-9bd4-58bac09ce615\") " pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.531802 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:32Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.535297 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.535342 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.535353 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.535373 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.535385 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:32Z","lastTransitionTime":"2025-12-05T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.545279 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:32Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.638549 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.638648 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.638659 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.638680 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.638691 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:32Z","lastTransitionTime":"2025-12-05T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.741927 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.741982 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.741993 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.742016 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.742031 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:32Z","lastTransitionTime":"2025-12-05T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.844841 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.844888 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.844899 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.844922 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.844938 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:32Z","lastTransitionTime":"2025-12-05T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.948104 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.948150 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.948162 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.948184 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:32 crc kubenswrapper[4997]: I1205 06:55:32.948201 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:32Z","lastTransitionTime":"2025-12-05T06:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.014455 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs\") pod \"network-metrics-daemon-b2vlg\" (UID: \"81f6790f-312b-42ce-9bd4-58bac09ce615\") " pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.014732 4997 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.014872 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs podName:81f6790f-312b-42ce-9bd4-58bac09ce615 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:34.014840474 +0000 UTC m=+34.543747725 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs") pod "network-metrics-daemon-b2vlg" (UID: "81f6790f-312b-42ce-9bd4-58bac09ce615") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.052479 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.052579 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.052593 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.052642 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.052673 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:33Z","lastTransitionTime":"2025-12-05T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.061939 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" event={"ID":"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0","Type":"ContainerStarted","Data":"5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066"} Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.061997 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" event={"ID":"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0","Type":"ContainerStarted","Data":"53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40"} Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.079323 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:33Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.095322 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:33Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.113388 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:33Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.130418 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:33Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.149595 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:33Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.156332 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.156385 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.156396 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.156424 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.156437 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:33Z","lastTransitionTime":"2025-12-05T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.165274 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:33Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.183363 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:33Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.202360 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:33Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.220887 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:33Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.238981 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:33Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.255565 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:33Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.260417 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.260468 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.260481 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.260502 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.260518 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:33Z","lastTransitionTime":"2025-12-05T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.276676 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:33Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.301829 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:28Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 06:55:27.880124 6256 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 06:55:27.880257 6256 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 06:55:27.880321 6256 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 06:55:27.880335 6256 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 06:55:27.880362 6256 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 06:55:27.880378 6256 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 06:55:27.880389 6256 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 06:55:27.880399 6256 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 06:55:27.880414 6256 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 06:55:27.880420 6256 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 06:55:27.880423 6256 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 06:55:27.880434 6256 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 06:55:27.880445 6256 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 06:55:27.880516 6256 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 06:55:27.880555 6256 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"message\\\":\\\"] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-5m9fs\\\\nI1205 06:55:29.888179 6428 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-bchpk after 0 failed attempt(s)\\\\nI1205 06:55:29.888181 6428 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-5m9fs in node crc\\\\nI1205 06:55:29.888185 6428 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-bchpk\\\\nI1205 06:55:29.888188 6428 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-5m9fs after 0 failed attempt(s)\\\\nF1205 06:55:29.887701 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:33Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.317500 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:33Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.333480 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:33Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.350246 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:33Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.364096 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.364150 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.364160 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.364204 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.364216 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:33Z","lastTransitionTime":"2025-12-05T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.466523 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.466596 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.466608 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.466642 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.466652 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:33Z","lastTransitionTime":"2025-12-05T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.522516 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.522800 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:55:49.522763418 +0000 UTC m=+50.051670709 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.522959 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.523119 4997 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.523231 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:49.52320713 +0000 UTC m=+50.052114391 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.569699 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.569748 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.569763 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.569794 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.569808 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:33Z","lastTransitionTime":"2025-12-05T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.624522 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.624659 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.624707 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.624807 4997 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.624902 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.624934 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.624944 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.624992 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.625008 4997 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.624953 4997 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.624914 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:49.624894674 +0000 UTC m=+50.153801935 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.625170 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:49.625143911 +0000 UTC m=+50.154051182 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.625188 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:49.625180482 +0000 UTC m=+50.154087743 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.673107 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.673163 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.673174 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.673188 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.673199 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:33Z","lastTransitionTime":"2025-12-05T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.748482 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.748812 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.748919 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.749007 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.749087 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.749417 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.749581 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:33 crc kubenswrapper[4997]: E1205 06:55:33.749945 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.775632 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.775693 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.775705 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.775726 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.775744 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:33Z","lastTransitionTime":"2025-12-05T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.878658 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.878714 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.878727 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.878744 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.878757 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:33Z","lastTransitionTime":"2025-12-05T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.982399 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.982493 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.982529 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.982557 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:33 crc kubenswrapper[4997]: I1205 06:55:33.982573 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:33Z","lastTransitionTime":"2025-12-05T06:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.030122 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs\") pod \"network-metrics-daemon-b2vlg\" (UID: \"81f6790f-312b-42ce-9bd4-58bac09ce615\") " pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:34 crc kubenswrapper[4997]: E1205 06:55:34.030488 4997 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 06:55:34 crc kubenswrapper[4997]: E1205 06:55:34.030709 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs podName:81f6790f-312b-42ce-9bd4-58bac09ce615 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:36.030664323 +0000 UTC m=+36.559571624 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs") pod "network-metrics-daemon-b2vlg" (UID: "81f6790f-312b-42ce-9bd4-58bac09ce615") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.085778 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.085831 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.085842 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.085899 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.085911 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:34Z","lastTransitionTime":"2025-12-05T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.189918 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.189967 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.189980 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.189996 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.190007 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:34Z","lastTransitionTime":"2025-12-05T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.293679 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.293740 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.293760 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.293790 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.293816 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:34Z","lastTransitionTime":"2025-12-05T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.397489 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.397582 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.397609 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.397693 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.397767 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:34Z","lastTransitionTime":"2025-12-05T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.503500 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.503571 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.503582 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.503608 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.503639 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:34Z","lastTransitionTime":"2025-12-05T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.606420 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.606515 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.606586 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.606683 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.606709 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:34Z","lastTransitionTime":"2025-12-05T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.710572 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.710646 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.710655 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.710674 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.710683 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:34Z","lastTransitionTime":"2025-12-05T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.814106 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.814210 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.814231 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.814300 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.814319 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:34Z","lastTransitionTime":"2025-12-05T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.917292 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.917363 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.917388 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.917416 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:34 crc kubenswrapper[4997]: I1205 06:55:34.917435 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:34Z","lastTransitionTime":"2025-12-05T06:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.022001 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.022072 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.022089 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.022116 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.022138 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:35Z","lastTransitionTime":"2025-12-05T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.126041 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.126111 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.126127 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.126151 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.126169 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:35Z","lastTransitionTime":"2025-12-05T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.233268 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.233373 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.233384 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.233405 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.233418 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:35Z","lastTransitionTime":"2025-12-05T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.336968 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.337043 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.337069 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.337103 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.337127 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:35Z","lastTransitionTime":"2025-12-05T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.440852 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.440925 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.440944 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.440974 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.440996 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:35Z","lastTransitionTime":"2025-12-05T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.544024 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.544067 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.544075 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.544090 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.544101 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:35Z","lastTransitionTime":"2025-12-05T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.648343 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.648582 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.648658 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.648698 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.648726 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:35Z","lastTransitionTime":"2025-12-05T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.748481 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.748481 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:35 crc kubenswrapper[4997]: E1205 06:55:35.748725 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:55:35 crc kubenswrapper[4997]: E1205 06:55:35.748909 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.748513 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:35 crc kubenswrapper[4997]: E1205 06:55:35.749097 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.749385 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:35 crc kubenswrapper[4997]: E1205 06:55:35.749511 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.751865 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.751921 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.751933 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.751954 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.751966 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:35Z","lastTransitionTime":"2025-12-05T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.854867 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.854909 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.854920 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.854935 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.854947 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:35Z","lastTransitionTime":"2025-12-05T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.957517 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.957588 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.957607 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.957661 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:35 crc kubenswrapper[4997]: I1205 06:55:35.957687 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:35Z","lastTransitionTime":"2025-12-05T06:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.053206 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs\") pod \"network-metrics-daemon-b2vlg\" (UID: \"81f6790f-312b-42ce-9bd4-58bac09ce615\") " pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:36 crc kubenswrapper[4997]: E1205 06:55:36.053488 4997 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 06:55:36 crc kubenswrapper[4997]: E1205 06:55:36.053667 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs podName:81f6790f-312b-42ce-9bd4-58bac09ce615 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:40.053606615 +0000 UTC m=+40.582514106 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs") pod "network-metrics-daemon-b2vlg" (UID: "81f6790f-312b-42ce-9bd4-58bac09ce615") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.062174 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.062219 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.062232 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.062254 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.062267 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:36Z","lastTransitionTime":"2025-12-05T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.165399 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.165475 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.165496 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.165529 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.165552 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:36Z","lastTransitionTime":"2025-12-05T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.268785 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.268882 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.268909 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.268948 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.268976 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:36Z","lastTransitionTime":"2025-12-05T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.372428 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.372488 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.372497 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.372516 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.372526 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:36Z","lastTransitionTime":"2025-12-05T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.476763 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.476825 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.476840 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.476867 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.476883 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:36Z","lastTransitionTime":"2025-12-05T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.579993 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.580071 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.580089 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.580120 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.580138 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:36Z","lastTransitionTime":"2025-12-05T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.683019 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.683066 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.683080 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.683131 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.683141 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:36Z","lastTransitionTime":"2025-12-05T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.786217 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.786266 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.786278 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.786300 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.786315 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:36Z","lastTransitionTime":"2025-12-05T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.889358 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.889398 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.889410 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.889426 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.889437 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:36Z","lastTransitionTime":"2025-12-05T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.993353 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.993419 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.993433 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.993457 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:36 crc kubenswrapper[4997]: I1205 06:55:36.993474 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:36Z","lastTransitionTime":"2025-12-05T06:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.097752 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.097845 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.097866 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.097914 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.097952 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:37Z","lastTransitionTime":"2025-12-05T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.202102 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.202148 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.202159 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.202183 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.202196 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:37Z","lastTransitionTime":"2025-12-05T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.310994 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.311093 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.311117 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.311167 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.311201 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:37Z","lastTransitionTime":"2025-12-05T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.414602 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.414720 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.414790 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.414825 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.414846 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:37Z","lastTransitionTime":"2025-12-05T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.519423 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.519488 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.519509 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.519537 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.519556 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:37Z","lastTransitionTime":"2025-12-05T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.623739 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.623822 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.623842 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.623873 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.623895 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:37Z","lastTransitionTime":"2025-12-05T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.728055 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.728120 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.728139 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.728175 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.728211 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:37Z","lastTransitionTime":"2025-12-05T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.748532 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.748569 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.748670 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:37 crc kubenswrapper[4997]: E1205 06:55:37.748803 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.748969 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:37 crc kubenswrapper[4997]: E1205 06:55:37.749009 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:37 crc kubenswrapper[4997]: E1205 06:55:37.749279 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:37 crc kubenswrapper[4997]: E1205 06:55:37.749380 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.830402 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.830828 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.830917 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.831000 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.831089 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:37Z","lastTransitionTime":"2025-12-05T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.934086 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.934169 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.934203 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.934235 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:37 crc kubenswrapper[4997]: I1205 06:55:37.934258 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:37Z","lastTransitionTime":"2025-12-05T06:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.036686 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.036714 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.036725 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.036742 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.036753 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:38Z","lastTransitionTime":"2025-12-05T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.139393 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.139438 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.139449 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.139463 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.139471 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:38Z","lastTransitionTime":"2025-12-05T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.241747 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.241776 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.241785 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.241799 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.241808 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:38Z","lastTransitionTime":"2025-12-05T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.343727 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.343760 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.343768 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.343780 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.343789 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:38Z","lastTransitionTime":"2025-12-05T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.446439 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.446469 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.446479 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.446495 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.446505 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:38Z","lastTransitionTime":"2025-12-05T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.548909 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.548952 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.548960 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.548977 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.548985 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:38Z","lastTransitionTime":"2025-12-05T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.651737 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.651780 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.651789 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.651804 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.651813 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:38Z","lastTransitionTime":"2025-12-05T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.755387 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.755420 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.755430 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.755446 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.755459 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:38Z","lastTransitionTime":"2025-12-05T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.859558 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.859594 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.859603 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.859642 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.859653 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:38Z","lastTransitionTime":"2025-12-05T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.961854 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.961893 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.961905 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.961921 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:38 crc kubenswrapper[4997]: I1205 06:55:38.961933 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:38Z","lastTransitionTime":"2025-12-05T06:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.064051 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.064099 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.064112 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.064130 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.064185 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:39Z","lastTransitionTime":"2025-12-05T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.166496 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.166545 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.166558 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.166571 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.166582 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:39Z","lastTransitionTime":"2025-12-05T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.269920 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.269972 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.269989 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.270012 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.270029 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:39Z","lastTransitionTime":"2025-12-05T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.372839 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.372937 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.372950 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.372970 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.372984 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:39Z","lastTransitionTime":"2025-12-05T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.476863 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.477205 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.477268 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.477408 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.477509 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:39Z","lastTransitionTime":"2025-12-05T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.581811 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.581887 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.581906 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.581936 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.581958 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:39Z","lastTransitionTime":"2025-12-05T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.684370 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.684425 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.684438 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.684454 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.684465 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:39Z","lastTransitionTime":"2025-12-05T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.748499 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.748574 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.748508 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.748797 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:39 crc kubenswrapper[4997]: E1205 06:55:39.748806 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:39 crc kubenswrapper[4997]: E1205 06:55:39.748906 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:39 crc kubenswrapper[4997]: E1205 06:55:39.748938 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:39 crc kubenswrapper[4997]: E1205 06:55:39.748999 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.764282 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:39Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.779062 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:39Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.787754 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.787827 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.787847 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.787878 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.787904 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:39Z","lastTransitionTime":"2025-12-05T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.802661 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:39Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.817419 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:39Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.839650 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:39Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.852063 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:39Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.868149 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:39Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.885573 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:39Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.891028 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.891078 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.891090 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.891111 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.891125 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:39Z","lastTransitionTime":"2025-12-05T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.898785 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:39Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.929227 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b3e32aa7df2bbdd1266282f4086971ad94d9cb727bfa2da9f987ced55aaa185\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:28Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 06:55:27.880124 6256 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 06:55:27.880257 6256 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 06:55:27.880321 6256 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 06:55:27.880335 6256 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 06:55:27.880362 6256 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 06:55:27.880378 6256 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 06:55:27.880389 6256 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 06:55:27.880399 6256 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 06:55:27.880414 6256 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 06:55:27.880420 6256 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 06:55:27.880423 6256 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 06:55:27.880434 6256 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 06:55:27.880445 6256 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 06:55:27.880516 6256 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 06:55:27.880555 6256 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"message\\\":\\\"] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-5m9fs\\\\nI1205 06:55:29.888179 6428 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-bchpk after 0 failed attempt(s)\\\\nI1205 06:55:29.888181 6428 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-5m9fs in node crc\\\\nI1205 06:55:29.888185 6428 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-bchpk\\\\nI1205 06:55:29.888188 6428 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-5m9fs after 0 failed attempt(s)\\\\nF1205 06:55:29.887701 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:39Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.947550 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:39Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.963998 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:39Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.978579 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:39Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.994124 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.994174 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.994183 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.994202 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.994213 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:39Z","lastTransitionTime":"2025-12-05T06:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:39 crc kubenswrapper[4997]: I1205 06:55:39.996151 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:39Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.012243 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:40Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.031482 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:40Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.096473 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.096529 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.096537 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.096554 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.096563 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:40Z","lastTransitionTime":"2025-12-05T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.116776 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs\") pod \"network-metrics-daemon-b2vlg\" (UID: \"81f6790f-312b-42ce-9bd4-58bac09ce615\") " pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:40 crc kubenswrapper[4997]: E1205 06:55:40.117036 4997 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 06:55:40 crc kubenswrapper[4997]: E1205 06:55:40.117142 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs podName:81f6790f-312b-42ce-9bd4-58bac09ce615 nodeName:}" failed. No retries permitted until 2025-12-05 06:55:48.117113094 +0000 UTC m=+48.646020365 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs") pod "network-metrics-daemon-b2vlg" (UID: "81f6790f-312b-42ce-9bd4-58bac09ce615") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.200306 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.200693 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.200710 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.200730 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.200743 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:40Z","lastTransitionTime":"2025-12-05T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.303688 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.303787 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.303809 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.303842 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.303861 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:40Z","lastTransitionTime":"2025-12-05T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.408430 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.408495 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.408516 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.408543 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.408563 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:40Z","lastTransitionTime":"2025-12-05T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.510982 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.511019 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.511029 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.511046 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.511056 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:40Z","lastTransitionTime":"2025-12-05T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.614299 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.614358 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.614375 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.614397 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.614413 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:40Z","lastTransitionTime":"2025-12-05T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.717350 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.717402 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.717416 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.717436 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.717452 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:40Z","lastTransitionTime":"2025-12-05T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.819826 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.819870 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.819886 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.819908 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.819924 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:40Z","lastTransitionTime":"2025-12-05T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.922170 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.922203 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.922279 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.922298 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:40 crc kubenswrapper[4997]: I1205 06:55:40.922329 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:40Z","lastTransitionTime":"2025-12-05T06:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.026248 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.026287 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.026300 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.026318 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.026329 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:41Z","lastTransitionTime":"2025-12-05T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.129414 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.129473 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.129487 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.129509 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.129527 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:41Z","lastTransitionTime":"2025-12-05T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.233882 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.234148 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.234168 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.234198 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.234219 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:41Z","lastTransitionTime":"2025-12-05T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.337783 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.337834 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.337844 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.337862 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.337875 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:41Z","lastTransitionTime":"2025-12-05T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.440946 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.441013 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.441033 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.441063 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.441083 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:41Z","lastTransitionTime":"2025-12-05T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.544842 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.544883 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.544891 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.544907 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.544916 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:41Z","lastTransitionTime":"2025-12-05T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.647595 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.647676 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.647688 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.647707 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.647719 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:41Z","lastTransitionTime":"2025-12-05T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.748740 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.748771 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.748752 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.748751 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:41 crc kubenswrapper[4997]: E1205 06:55:41.748918 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:41 crc kubenswrapper[4997]: E1205 06:55:41.749229 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:55:41 crc kubenswrapper[4997]: E1205 06:55:41.749371 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:41 crc kubenswrapper[4997]: E1205 06:55:41.749533 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.750315 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.750351 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.750364 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.750385 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.750397 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:41Z","lastTransitionTime":"2025-12-05T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.853180 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.853229 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.853243 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.853261 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.853273 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:41Z","lastTransitionTime":"2025-12-05T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.860313 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.860353 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.860363 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.860380 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.860393 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:41Z","lastTransitionTime":"2025-12-05T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:41 crc kubenswrapper[4997]: E1205 06:55:41.874782 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:41Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.878699 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.878762 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.878775 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.878795 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.878809 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:41Z","lastTransitionTime":"2025-12-05T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:41 crc kubenswrapper[4997]: E1205 06:55:41.894825 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:41Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.900451 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.900504 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.900515 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.900535 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.900548 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:41Z","lastTransitionTime":"2025-12-05T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:41 crc kubenswrapper[4997]: E1205 06:55:41.919004 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:41Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.970237 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.970275 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.970288 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.970371 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.970398 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:41Z","lastTransitionTime":"2025-12-05T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:41 crc kubenswrapper[4997]: E1205 06:55:41.985372 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:41Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.989739 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.989785 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.989798 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.989819 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:41 crc kubenswrapper[4997]: I1205 06:55:41.989833 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:41Z","lastTransitionTime":"2025-12-05T06:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:42 crc kubenswrapper[4997]: E1205 06:55:42.005639 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:42Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:42 crc kubenswrapper[4997]: E1205 06:55:42.005830 4997 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.008257 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.008307 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.008326 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.008350 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.008366 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:42Z","lastTransitionTime":"2025-12-05T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.111416 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.111483 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.111502 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.111533 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.111554 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:42Z","lastTransitionTime":"2025-12-05T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.215255 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.215341 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.215364 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.215405 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.215432 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:42Z","lastTransitionTime":"2025-12-05T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.319521 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.319587 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.319606 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.319681 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.319721 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:42Z","lastTransitionTime":"2025-12-05T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.423368 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.423443 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.423463 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.423500 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.423521 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:42Z","lastTransitionTime":"2025-12-05T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.526702 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.526744 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.526757 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.526775 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.526785 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:42Z","lastTransitionTime":"2025-12-05T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.630501 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.630566 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.630584 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.630697 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.630712 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:42Z","lastTransitionTime":"2025-12-05T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.733401 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.733456 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.733466 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.733484 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.733493 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:42Z","lastTransitionTime":"2025-12-05T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.835708 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.835759 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.835772 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.835791 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.835813 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:42Z","lastTransitionTime":"2025-12-05T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.937913 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.937961 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.937970 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.937988 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:42 crc kubenswrapper[4997]: I1205 06:55:42.938002 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:42Z","lastTransitionTime":"2025-12-05T06:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.041580 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.041681 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.041702 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.041732 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.041752 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:43Z","lastTransitionTime":"2025-12-05T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.144479 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.144550 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.144567 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.144600 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.144687 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:43Z","lastTransitionTime":"2025-12-05T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.249585 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.249700 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.249719 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.249749 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.249770 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:43Z","lastTransitionTime":"2025-12-05T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.354825 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.354928 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.354955 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.354992 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.355017 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:43Z","lastTransitionTime":"2025-12-05T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.458682 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.458758 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.458778 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.458812 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.458835 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:43Z","lastTransitionTime":"2025-12-05T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.562999 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.563086 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.563114 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.563150 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.563174 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:43Z","lastTransitionTime":"2025-12-05T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.666705 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.666778 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.666793 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.666819 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.666835 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:43Z","lastTransitionTime":"2025-12-05T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.748771 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.748774 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.748963 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.749597 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:43 crc kubenswrapper[4997]: E1205 06:55:43.750174 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:43 crc kubenswrapper[4997]: E1205 06:55:43.750416 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:43 crc kubenswrapper[4997]: E1205 06:55:43.750337 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:55:43 crc kubenswrapper[4997]: E1205 06:55:43.750846 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.751177 4997 scope.go:117] "RemoveContainer" containerID="cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.765887 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:43Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.772031 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.772089 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.772100 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.772120 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.772132 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:43Z","lastTransitionTime":"2025-12-05T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.785661 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:43Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.800821 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:43Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.812849 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:43Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.824158 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:43Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.841109 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:43Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.854859 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:43Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.868511 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:43Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.875717 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.875782 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.875807 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.875845 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.875870 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:43Z","lastTransitionTime":"2025-12-05T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.884967 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:43Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.902734 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:43Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.917535 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:43Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.934815 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:43Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.958319 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"message\\\":\\\"] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-5m9fs\\\\nI1205 06:55:29.888179 6428 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-bchpk after 0 failed attempt(s)\\\\nI1205 06:55:29.888181 6428 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-5m9fs in node crc\\\\nI1205 06:55:29.888185 6428 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-bchpk\\\\nI1205 06:55:29.888188 6428 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-5m9fs after 0 failed attempt(s)\\\\nF1205 06:55:29.887701 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:43Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.972597 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:43Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.978987 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.979032 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.979046 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.979066 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.979080 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:43Z","lastTransitionTime":"2025-12-05T06:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:43 crc kubenswrapper[4997]: I1205 06:55:43.987964 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:43Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.001879 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:43Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.081715 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.081749 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.081758 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.081770 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.081782 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:44Z","lastTransitionTime":"2025-12-05T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.107021 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovnkube-controller/1.log" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.109500 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerStarted","Data":"5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f"} Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.109636 4997 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.125475 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:44Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.144603 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:44Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.171284 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:44Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.184043 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.184077 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.184086 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.184108 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.184118 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:44Z","lastTransitionTime":"2025-12-05T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.195459 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"message\\\":\\\"] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-5m9fs\\\\nI1205 06:55:29.888179 6428 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-bchpk after 0 failed attempt(s)\\\\nI1205 06:55:29.888181 6428 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-5m9fs in node crc\\\\nI1205 06:55:29.888185 6428 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-bchpk\\\\nI1205 06:55:29.888188 6428 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-5m9fs after 0 failed attempt(s)\\\\nF1205 06:55:29.887701 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:44Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.214220 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:44Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.226888 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:44Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.240328 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:44Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.250120 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:44Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.262183 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:44Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.273195 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:44Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.282565 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:44Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.285981 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.286015 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.286025 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.286039 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.286050 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:44Z","lastTransitionTime":"2025-12-05T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.293148 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:44Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.304745 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:44Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.322527 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:44Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.338633 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:44Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.352170 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:44Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.388545 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.388580 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.388592 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.388607 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.388635 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:44Z","lastTransitionTime":"2025-12-05T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.497433 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.497472 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.497483 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.497499 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.497509 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:44Z","lastTransitionTime":"2025-12-05T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.600626 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.600673 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.600683 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.600701 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.600712 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:44Z","lastTransitionTime":"2025-12-05T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.702981 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.703046 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.703064 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.703088 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.703106 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:44Z","lastTransitionTime":"2025-12-05T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.805492 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.805559 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.805576 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.806066 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.806121 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:44Z","lastTransitionTime":"2025-12-05T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.908701 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.908737 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.908751 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.908768 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:44 crc kubenswrapper[4997]: I1205 06:55:44.908781 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:44Z","lastTransitionTime":"2025-12-05T06:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.011610 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.011679 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.011687 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.011703 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.011712 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:45Z","lastTransitionTime":"2025-12-05T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.113493 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.113567 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.113591 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.113656 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.113695 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:45Z","lastTransitionTime":"2025-12-05T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.115173 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovnkube-controller/2.log" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.116504 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovnkube-controller/1.log" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.120064 4997 generic.go:334] "Generic (PLEG): container finished" podID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerID="5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f" exitCode=1 Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.120104 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerDied","Data":"5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f"} Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.120173 4997 scope.go:117] "RemoveContainer" containerID="cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.121950 4997 scope.go:117] "RemoveContainer" containerID="5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f" Dec 05 06:55:45 crc kubenswrapper[4997]: E1205 06:55:45.122701 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.142265 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:45Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.162509 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:45Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.182486 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd29c273522f55dfadd8cd5c9b549f169c9447efd75ea5c9c69e94d5623fc9e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"message\\\":\\\"] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-5m9fs\\\\nI1205 06:55:29.888179 6428 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-bchpk after 0 failed attempt(s)\\\\nI1205 06:55:29.888181 6428 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-5m9fs in node crc\\\\nI1205 06:55:29.888185 6428 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-bchpk\\\\nI1205 06:55:29.888188 6428 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-5m9fs after 0 failed attempt(s)\\\\nF1205 06:55:29.887701 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:44Z\\\",\\\"message\\\":\\\"nil\\\\u003e UUID: UUIDName:}]\\\\nI1205 06:55:44.568703 6645 ovnkube.go:599] Stopped ovnkube\\\\nI1205 06:55:44.568739 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1205 06:55:44.568170 6645 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-l5p2z in node crc\\\\nI1205 06:55:44.568775 6645 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-l5p2z after 0 failed attempt(s)\\\\nI1205 06:55:44.568784 6645 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-l5p2z\\\\nI1205 06:55:44.568688 6645 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/api]} name:Service_openshift-apiserver/api_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.37:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {88e20c31-5b8d-4d44-bbd8-dba87b7dbaf0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1205 06:55:44.568837 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:45Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.195158 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:45Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.217142 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.217179 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.217188 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.217203 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.217213 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:45Z","lastTransitionTime":"2025-12-05T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.217592 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:45Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.235930 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:45Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.254234 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:45Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.269412 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:45Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.282214 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:45Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.294112 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:45Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.304559 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:45Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.317202 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:45Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.319841 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.319870 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.319878 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.319896 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.319908 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:45Z","lastTransitionTime":"2025-12-05T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.331570 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:45Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.343725 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:45Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.354302 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:45Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.366215 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:45Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.423027 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.423070 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.423081 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.423098 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.423110 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:45Z","lastTransitionTime":"2025-12-05T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.526039 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.526078 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.526088 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.526101 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.526113 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:45Z","lastTransitionTime":"2025-12-05T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.628432 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.628543 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.628571 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.628603 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.628657 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:45Z","lastTransitionTime":"2025-12-05T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.735097 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.735188 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.735229 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.735268 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.735294 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:45Z","lastTransitionTime":"2025-12-05T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.748361 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.748495 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.748446 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.748554 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:45 crc kubenswrapper[4997]: E1205 06:55:45.748642 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:55:45 crc kubenswrapper[4997]: E1205 06:55:45.748736 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:45 crc kubenswrapper[4997]: E1205 06:55:45.749017 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:45 crc kubenswrapper[4997]: E1205 06:55:45.749156 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.838531 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.838586 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.838599 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.838642 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.838658 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:45Z","lastTransitionTime":"2025-12-05T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.940687 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.940764 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.940787 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.940817 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:45 crc kubenswrapper[4997]: I1205 06:55:45.940839 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:45Z","lastTransitionTime":"2025-12-05T06:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.043448 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.043482 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.043492 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.043506 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.043515 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:46Z","lastTransitionTime":"2025-12-05T06:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.124870 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovnkube-controller/2.log" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.145228 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.145279 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.145289 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.145309 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.145320 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:46Z","lastTransitionTime":"2025-12-05T06:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.247811 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.247872 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.247887 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.247907 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.247920 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:46Z","lastTransitionTime":"2025-12-05T06:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.350844 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.350915 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.350941 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.350985 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.351009 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:46Z","lastTransitionTime":"2025-12-05T06:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.453492 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.453523 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.453532 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.453547 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.453555 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:46Z","lastTransitionTime":"2025-12-05T06:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.555555 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.555600 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.555649 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.555669 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.555683 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:46Z","lastTransitionTime":"2025-12-05T06:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.658788 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.658831 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.658852 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.658881 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.658897 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:46Z","lastTransitionTime":"2025-12-05T06:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.761123 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.761169 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.761193 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.761212 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.761225 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:46Z","lastTransitionTime":"2025-12-05T06:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.864431 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.864493 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.864529 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.864568 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.864591 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:46Z","lastTransitionTime":"2025-12-05T06:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.968130 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.968168 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.968177 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.968193 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:46 crc kubenswrapper[4997]: I1205 06:55:46.968201 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:46Z","lastTransitionTime":"2025-12-05T06:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.071045 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.071115 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.071136 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.071168 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.071187 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:47Z","lastTransitionTime":"2025-12-05T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.174347 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.174411 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.174434 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.174462 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.174485 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:47Z","lastTransitionTime":"2025-12-05T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.277397 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.277478 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.277497 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.277528 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.277549 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:47Z","lastTransitionTime":"2025-12-05T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.380767 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.380840 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.380859 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.380892 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.380913 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:47Z","lastTransitionTime":"2025-12-05T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.482697 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.482736 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.482744 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.482756 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.482764 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:47Z","lastTransitionTime":"2025-12-05T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.586005 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.586105 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.586140 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.586162 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.586184 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:47Z","lastTransitionTime":"2025-12-05T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.689901 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.689977 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.689997 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.690029 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.690049 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:47Z","lastTransitionTime":"2025-12-05T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.749099 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.749099 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.749121 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:47 crc kubenswrapper[4997]: E1205 06:55:47.749423 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:47 crc kubenswrapper[4997]: E1205 06:55:47.749244 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.749115 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:47 crc kubenswrapper[4997]: E1205 06:55:47.749499 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:47 crc kubenswrapper[4997]: E1205 06:55:47.749866 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.792165 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.792392 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.792404 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.792424 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.792436 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:47Z","lastTransitionTime":"2025-12-05T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.896305 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.896348 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.896357 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.896374 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.896384 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:47Z","lastTransitionTime":"2025-12-05T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.999421 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.999462 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.999470 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.999484 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:47 crc kubenswrapper[4997]: I1205 06:55:47.999493 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:47Z","lastTransitionTime":"2025-12-05T06:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.102242 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.102277 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.102305 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.102320 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.102329 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:48Z","lastTransitionTime":"2025-12-05T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.137515 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs\") pod \"network-metrics-daemon-b2vlg\" (UID: \"81f6790f-312b-42ce-9bd4-58bac09ce615\") " pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:48 crc kubenswrapper[4997]: E1205 06:55:48.137763 4997 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 06:55:48 crc kubenswrapper[4997]: E1205 06:55:48.137841 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs podName:81f6790f-312b-42ce-9bd4-58bac09ce615 nodeName:}" failed. No retries permitted until 2025-12-05 06:56:04.137820382 +0000 UTC m=+64.666727663 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs") pod "network-metrics-daemon-b2vlg" (UID: "81f6790f-312b-42ce-9bd4-58bac09ce615") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.205132 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.205182 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.205197 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.205236 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.205250 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:48Z","lastTransitionTime":"2025-12-05T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.307737 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.307790 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.307798 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.307813 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.307821 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:48Z","lastTransitionTime":"2025-12-05T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.410203 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.410257 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.410265 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.410279 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.410287 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:48Z","lastTransitionTime":"2025-12-05T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.512739 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.512784 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.512810 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.512830 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.512841 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:48Z","lastTransitionTime":"2025-12-05T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.615538 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.615585 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.615596 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.615644 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.615658 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:48Z","lastTransitionTime":"2025-12-05T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.718748 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.718958 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.718966 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.718979 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.718988 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:48Z","lastTransitionTime":"2025-12-05T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.740150 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.741155 4997 scope.go:117] "RemoveContainer" containerID="5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f" Dec 05 06:55:48 crc kubenswrapper[4997]: E1205 06:55:48.741324 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.752763 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:48Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.766042 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:48Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.782809 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:48Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.802231 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:44Z\\\",\\\"message\\\":\\\"nil\\\\u003e UUID: UUIDName:}]\\\\nI1205 06:55:44.568703 6645 ovnkube.go:599] Stopped ovnkube\\\\nI1205 06:55:44.568739 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1205 06:55:44.568170 6645 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-l5p2z in node crc\\\\nI1205 06:55:44.568775 6645 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-l5p2z after 0 failed attempt(s)\\\\nI1205 06:55:44.568784 6645 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-l5p2z\\\\nI1205 06:55:44.568688 6645 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/api]} name:Service_openshift-apiserver/api_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.37:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {88e20c31-5b8d-4d44-bbd8-dba87b7dbaf0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1205 06:55:44.568837 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:48Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.815751 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:48Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.821245 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.821289 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.821307 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.821331 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.821348 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:48Z","lastTransitionTime":"2025-12-05T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.829654 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:48Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.842909 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:48Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.855383 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:48Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.866695 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:48Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.878868 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:48Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.887993 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:48Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.898603 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:48Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.910338 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:48Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.923140 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:48Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.923922 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.924037 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.924113 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.924203 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.924277 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:48Z","lastTransitionTime":"2025-12-05T06:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.934798 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:48Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:48 crc kubenswrapper[4997]: I1205 06:55:48.948939 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:48Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.027061 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.027366 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.027474 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.027596 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.027743 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:49Z","lastTransitionTime":"2025-12-05T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.133277 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.133528 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.133784 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.133859 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.133875 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:49Z","lastTransitionTime":"2025-12-05T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.236422 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.236462 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.236470 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.236484 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.236493 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:49Z","lastTransitionTime":"2025-12-05T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.339199 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.339247 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.339261 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.339280 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.339296 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:49Z","lastTransitionTime":"2025-12-05T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.442245 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.442282 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.442293 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.442310 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.442758 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:49Z","lastTransitionTime":"2025-12-05T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.545053 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.545120 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.545144 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.545176 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.545199 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:49Z","lastTransitionTime":"2025-12-05T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.549424 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.549592 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:49 crc kubenswrapper[4997]: E1205 06:55:49.549701 4997 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 06:55:49 crc kubenswrapper[4997]: E1205 06:55:49.549709 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:56:21.549677077 +0000 UTC m=+82.078584378 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:55:49 crc kubenswrapper[4997]: E1205 06:55:49.549804 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 06:56:21.54978635 +0000 UTC m=+82.078693651 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.647253 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.647309 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.647327 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.647351 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.647369 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:49Z","lastTransitionTime":"2025-12-05T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.651041 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.651100 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.651191 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:49 crc kubenswrapper[4997]: E1205 06:55:49.651252 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 06:55:49 crc kubenswrapper[4997]: E1205 06:55:49.651277 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 06:55:49 crc kubenswrapper[4997]: E1205 06:55:49.651288 4997 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:49 crc kubenswrapper[4997]: E1205 06:55:49.651319 4997 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 06:55:49 crc kubenswrapper[4997]: E1205 06:55:49.651341 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 06:56:21.651323696 +0000 UTC m=+82.180230957 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:49 crc kubenswrapper[4997]: E1205 06:55:49.651346 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 06:55:49 crc kubenswrapper[4997]: E1205 06:55:49.651371 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 06:55:49 crc kubenswrapper[4997]: E1205 06:55:49.651391 4997 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:49 crc kubenswrapper[4997]: E1205 06:55:49.651374 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 06:56:21.651358937 +0000 UTC m=+82.180266208 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 06:55:49 crc kubenswrapper[4997]: E1205 06:55:49.651467 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 06:56:21.651448769 +0000 UTC m=+82.180356070 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.748501 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.748588 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:49 crc kubenswrapper[4997]: E1205 06:55:49.748762 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.748814 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:49 crc kubenswrapper[4997]: E1205 06:55:49.748925 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:55:49 crc kubenswrapper[4997]: E1205 06:55:49.749071 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.749129 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:49 crc kubenswrapper[4997]: E1205 06:55:49.749355 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.750064 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.750154 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.750178 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.750209 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.750234 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:49Z","lastTransitionTime":"2025-12-05T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.774003 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:49Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.792008 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:49Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.806349 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:49Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.819339 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:49Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.839865 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:49Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.852287 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.852318 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.852326 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.852338 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.852348 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:49Z","lastTransitionTime":"2025-12-05T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.858370 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:49Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.881258 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:49Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.898107 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:49Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.918491 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:44Z\\\",\\\"message\\\":\\\"nil\\\\u003e UUID: UUIDName:}]\\\\nI1205 06:55:44.568703 6645 ovnkube.go:599] Stopped ovnkube\\\\nI1205 06:55:44.568739 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1205 06:55:44.568170 6645 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-l5p2z in node crc\\\\nI1205 06:55:44.568775 6645 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-l5p2z after 0 failed attempt(s)\\\\nI1205 06:55:44.568784 6645 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-l5p2z\\\\nI1205 06:55:44.568688 6645 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/api]} name:Service_openshift-apiserver/api_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.37:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {88e20c31-5b8d-4d44-bbd8-dba87b7dbaf0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1205 06:55:44.568837 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:49Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.936401 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:49Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.949857 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:49Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.954378 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.954438 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.954456 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.954483 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.954501 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:49Z","lastTransitionTime":"2025-12-05T06:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.962634 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:49Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.975073 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:49Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:49 crc kubenswrapper[4997]: I1205 06:55:49.985830 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:49Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.000510 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:49Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.011889 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:50Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.056019 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.056076 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.056091 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.056111 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.056126 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:50Z","lastTransitionTime":"2025-12-05T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.158007 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.158050 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.158061 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.158079 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.158095 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:50Z","lastTransitionTime":"2025-12-05T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.260452 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.260482 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.260490 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.260502 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.260512 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:50Z","lastTransitionTime":"2025-12-05T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.363069 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.363102 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.363109 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.363124 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.363142 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:50Z","lastTransitionTime":"2025-12-05T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.465557 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.465599 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.465639 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.465656 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.465668 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:50Z","lastTransitionTime":"2025-12-05T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.567609 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.567657 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.567669 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.567687 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.567698 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:50Z","lastTransitionTime":"2025-12-05T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.669861 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.669908 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.669937 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.669959 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.669975 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:50Z","lastTransitionTime":"2025-12-05T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.772513 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.772547 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.772558 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.772572 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.772585 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:50Z","lastTransitionTime":"2025-12-05T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.875854 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.875897 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.875955 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.875991 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.876019 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:50Z","lastTransitionTime":"2025-12-05T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.978348 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.978407 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.978416 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.978432 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:50 crc kubenswrapper[4997]: I1205 06:55:50.978441 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:50Z","lastTransitionTime":"2025-12-05T06:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.080782 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.080818 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.080827 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.080842 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.080852 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:51Z","lastTransitionTime":"2025-12-05T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.183481 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.183521 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.183530 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.183546 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.183556 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:51Z","lastTransitionTime":"2025-12-05T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.286132 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.286181 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.286192 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.286212 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.286224 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:51Z","lastTransitionTime":"2025-12-05T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.388114 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.388186 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.388202 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.388244 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.388253 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:51Z","lastTransitionTime":"2025-12-05T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.490415 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.490442 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.490450 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.490478 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.490488 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:51Z","lastTransitionTime":"2025-12-05T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.592604 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.592698 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.592717 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.592741 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.592759 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:51Z","lastTransitionTime":"2025-12-05T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.695089 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.695122 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.695131 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.695145 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.695153 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:51Z","lastTransitionTime":"2025-12-05T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.748553 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.748603 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.748650 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:51 crc kubenswrapper[4997]: E1205 06:55:51.748714 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.748726 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:51 crc kubenswrapper[4997]: E1205 06:55:51.748770 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:51 crc kubenswrapper[4997]: E1205 06:55:51.748812 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:51 crc kubenswrapper[4997]: E1205 06:55:51.749022 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.796914 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.796944 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.796953 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.796966 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.796974 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:51Z","lastTransitionTime":"2025-12-05T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.900008 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.900062 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.900073 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.900091 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:51 crc kubenswrapper[4997]: I1205 06:55:51.900104 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:51Z","lastTransitionTime":"2025-12-05T06:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.003435 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.004008 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.004046 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.004075 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.004096 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:52Z","lastTransitionTime":"2025-12-05T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.107129 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.107168 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.107178 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.107194 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.107206 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:52Z","lastTransitionTime":"2025-12-05T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.210458 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.210503 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.210514 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.210530 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.210542 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:52Z","lastTransitionTime":"2025-12-05T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.313596 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.313904 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.313950 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.313977 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.314015 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:52Z","lastTransitionTime":"2025-12-05T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.350891 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.350944 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.350955 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.350974 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.350986 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:52Z","lastTransitionTime":"2025-12-05T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:52 crc kubenswrapper[4997]: E1205 06:55:52.369314 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:52Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.372529 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.372799 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.372941 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.373079 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.373205 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:52Z","lastTransitionTime":"2025-12-05T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:52 crc kubenswrapper[4997]: E1205 06:55:52.392846 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:52Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.396239 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.396269 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.396280 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.396295 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.396305 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:52Z","lastTransitionTime":"2025-12-05T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:52 crc kubenswrapper[4997]: E1205 06:55:52.406565 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:52Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.410971 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.411011 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.411020 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.411036 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.411045 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:52Z","lastTransitionTime":"2025-12-05T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:52 crc kubenswrapper[4997]: E1205 06:55:52.427561 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:52Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.431965 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.432021 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.432045 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.432072 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.432088 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:52Z","lastTransitionTime":"2025-12-05T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:52 crc kubenswrapper[4997]: E1205 06:55:52.444481 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:52Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:52 crc kubenswrapper[4997]: E1205 06:55:52.444664 4997 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.446310 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.446354 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.446370 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.446392 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.446408 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:52Z","lastTransitionTime":"2025-12-05T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.548846 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.549415 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.549431 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.549450 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.549460 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:52Z","lastTransitionTime":"2025-12-05T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.651515 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.651706 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.651719 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.651736 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.651747 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:52Z","lastTransitionTime":"2025-12-05T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.753867 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.753913 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.753922 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.753938 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.753948 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:52Z","lastTransitionTime":"2025-12-05T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.856395 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.856443 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.856455 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.856469 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.856478 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:52Z","lastTransitionTime":"2025-12-05T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.958279 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.958312 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.958321 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.958334 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:52 crc kubenswrapper[4997]: I1205 06:55:52.958343 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:52Z","lastTransitionTime":"2025-12-05T06:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.060316 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.060363 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.060376 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.060393 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.060405 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:53Z","lastTransitionTime":"2025-12-05T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.162006 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.162037 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.162046 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.162059 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.162069 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:53Z","lastTransitionTime":"2025-12-05T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.264065 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.264097 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.264105 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.264118 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.264127 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:53Z","lastTransitionTime":"2025-12-05T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.366009 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.366047 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.366064 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.366084 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.366096 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:53Z","lastTransitionTime":"2025-12-05T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.468279 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.468520 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.468595 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.468716 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.468799 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:53Z","lastTransitionTime":"2025-12-05T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.571550 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.571598 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.571607 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.571649 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.571658 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:53Z","lastTransitionTime":"2025-12-05T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.676173 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.676292 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.676305 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.676319 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.676328 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:53Z","lastTransitionTime":"2025-12-05T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.748828 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.748877 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:53 crc kubenswrapper[4997]: E1205 06:55:53.748971 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.748832 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:53 crc kubenswrapper[4997]: E1205 06:55:53.749085 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.749125 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:53 crc kubenswrapper[4997]: E1205 06:55:53.749168 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:55:53 crc kubenswrapper[4997]: E1205 06:55:53.749210 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.777978 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.778011 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.778023 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.778040 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.778051 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:53Z","lastTransitionTime":"2025-12-05T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.880660 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.880703 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.880715 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.880732 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.880744 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:53Z","lastTransitionTime":"2025-12-05T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.983751 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.983803 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.983814 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.983834 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:53 crc kubenswrapper[4997]: I1205 06:55:53.983848 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:53Z","lastTransitionTime":"2025-12-05T06:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.092045 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.092105 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.092131 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.092161 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.092184 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:54Z","lastTransitionTime":"2025-12-05T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.194417 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.194453 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.194464 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.194479 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.194491 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:54Z","lastTransitionTime":"2025-12-05T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.249011 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.260445 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.272585 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:44Z\\\",\\\"message\\\":\\\"nil\\\\u003e UUID: UUIDName:}]\\\\nI1205 06:55:44.568703 6645 ovnkube.go:599] Stopped ovnkube\\\\nI1205 06:55:44.568739 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1205 06:55:44.568170 6645 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-l5p2z in node crc\\\\nI1205 06:55:44.568775 6645 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-l5p2z after 0 failed attempt(s)\\\\nI1205 06:55:44.568784 6645 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-l5p2z\\\\nI1205 06:55:44.568688 6645 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/api]} name:Service_openshift-apiserver/api_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.37:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {88e20c31-5b8d-4d44-bbd8-dba87b7dbaf0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1205 06:55:44.568837 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:54Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.288392 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:54Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.296711 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.296760 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.296779 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.296802 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.296815 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:54Z","lastTransitionTime":"2025-12-05T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.302470 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:54Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.319529 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:54Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.333632 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:54Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.345882 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:54Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.358519 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:54Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.370264 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:54Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.379676 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:54Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.392883 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:54Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.398914 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.398946 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.398954 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.398969 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.398981 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:54Z","lastTransitionTime":"2025-12-05T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.402560 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:54Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.415599 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:54Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.429278 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:54Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.456946 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:54Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.477054 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:54Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.493550 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:54Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.501415 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.501472 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.501486 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.501504 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.501539 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:54Z","lastTransitionTime":"2025-12-05T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.604471 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.604513 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.604522 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.604539 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.604550 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:54Z","lastTransitionTime":"2025-12-05T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.706861 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.706909 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.706922 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.706940 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.706952 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:54Z","lastTransitionTime":"2025-12-05T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.809967 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.810002 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.810011 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.810024 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.810034 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:54Z","lastTransitionTime":"2025-12-05T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.912388 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.912423 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.912434 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.912449 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:54 crc kubenswrapper[4997]: I1205 06:55:54.912460 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:54Z","lastTransitionTime":"2025-12-05T06:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.015232 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.015293 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.015312 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.015335 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.015352 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:55Z","lastTransitionTime":"2025-12-05T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.117934 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.117958 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.117966 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.117979 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.117987 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:55Z","lastTransitionTime":"2025-12-05T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.220688 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.220726 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.220736 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.220750 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.220760 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:55Z","lastTransitionTime":"2025-12-05T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.323417 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.323477 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.323497 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.323526 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.323547 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:55Z","lastTransitionTime":"2025-12-05T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.425985 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.426016 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.426025 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.426039 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.426050 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:55Z","lastTransitionTime":"2025-12-05T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.528579 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.528632 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.528642 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.528654 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.528663 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:55Z","lastTransitionTime":"2025-12-05T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.630188 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.630228 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.630240 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.630254 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.630264 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:55Z","lastTransitionTime":"2025-12-05T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.731790 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.731831 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.731841 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.731857 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.731870 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:55Z","lastTransitionTime":"2025-12-05T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.748077 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.748124 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.748137 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:55 crc kubenswrapper[4997]: E1205 06:55:55.748170 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.748179 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:55 crc kubenswrapper[4997]: E1205 06:55:55.748263 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:55 crc kubenswrapper[4997]: E1205 06:55:55.748339 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:55:55 crc kubenswrapper[4997]: E1205 06:55:55.748402 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.834341 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.834452 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.834470 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.834499 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.834515 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:55Z","lastTransitionTime":"2025-12-05T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.936963 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.937001 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.937013 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.937030 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:55 crc kubenswrapper[4997]: I1205 06:55:55.937043 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:55Z","lastTransitionTime":"2025-12-05T06:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.038647 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.038689 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.038697 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.038712 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.038722 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:56Z","lastTransitionTime":"2025-12-05T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.141079 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.141124 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.141132 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.141152 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.141163 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:56Z","lastTransitionTime":"2025-12-05T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.243020 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.243064 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.243101 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.243116 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.243125 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:56Z","lastTransitionTime":"2025-12-05T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.345770 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.345811 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.345823 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.345839 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.345850 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:56Z","lastTransitionTime":"2025-12-05T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.448371 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.448409 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.448421 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.448438 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.448449 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:56Z","lastTransitionTime":"2025-12-05T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.550460 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.550505 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.550515 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.550530 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.550540 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:56Z","lastTransitionTime":"2025-12-05T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.653270 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.653324 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.653338 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.653358 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.653370 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:56Z","lastTransitionTime":"2025-12-05T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.755983 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.756035 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.756048 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.756068 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.756082 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:56Z","lastTransitionTime":"2025-12-05T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.858462 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.858502 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.858510 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.858525 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.858534 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:56Z","lastTransitionTime":"2025-12-05T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.960295 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.960331 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.960339 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.960351 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:56 crc kubenswrapper[4997]: I1205 06:55:56.960360 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:56Z","lastTransitionTime":"2025-12-05T06:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.062025 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.062300 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.062379 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.062465 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.062531 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:57Z","lastTransitionTime":"2025-12-05T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.166109 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.166139 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.166148 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.166162 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.166171 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:57Z","lastTransitionTime":"2025-12-05T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.268226 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.268272 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.268283 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.268300 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.268313 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:57Z","lastTransitionTime":"2025-12-05T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.370535 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.370572 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.370580 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.370593 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.370601 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:57Z","lastTransitionTime":"2025-12-05T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.472707 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.472945 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.473014 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.473084 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.473148 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:57Z","lastTransitionTime":"2025-12-05T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.575981 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.576055 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.576076 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.576102 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.576122 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:57Z","lastTransitionTime":"2025-12-05T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.679137 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.679203 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.679226 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.679255 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.679277 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:57Z","lastTransitionTime":"2025-12-05T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.748679 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.748739 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.748775 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.748676 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:57 crc kubenswrapper[4997]: E1205 06:55:57.748841 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:57 crc kubenswrapper[4997]: E1205 06:55:57.748955 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:55:57 crc kubenswrapper[4997]: E1205 06:55:57.749018 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:57 crc kubenswrapper[4997]: E1205 06:55:57.749112 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.782044 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.782091 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.782102 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.782121 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.782132 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:57Z","lastTransitionTime":"2025-12-05T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.885416 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.885447 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.885455 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.885468 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.885482 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:57Z","lastTransitionTime":"2025-12-05T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.987958 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.988011 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.988020 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.988036 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:57 crc kubenswrapper[4997]: I1205 06:55:57.988045 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:57Z","lastTransitionTime":"2025-12-05T06:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.090524 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.090581 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.090592 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.090644 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.090656 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:58Z","lastTransitionTime":"2025-12-05T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.193279 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.193357 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.193381 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.193412 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.193434 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:58Z","lastTransitionTime":"2025-12-05T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.297858 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.297917 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.297935 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.297963 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.297979 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:58Z","lastTransitionTime":"2025-12-05T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.400833 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.400890 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.400907 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.400928 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.400943 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:58Z","lastTransitionTime":"2025-12-05T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.504173 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.504244 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.504257 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.504287 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.504317 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:58Z","lastTransitionTime":"2025-12-05T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.608136 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.608218 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.608237 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.608261 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.608279 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:58Z","lastTransitionTime":"2025-12-05T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.710364 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.710400 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.710408 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.710422 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.710431 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:58Z","lastTransitionTime":"2025-12-05T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.813116 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.813182 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.813197 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.813250 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.813267 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:58Z","lastTransitionTime":"2025-12-05T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.916839 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.916921 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.916943 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.916972 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:58 crc kubenswrapper[4997]: I1205 06:55:58.916994 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:58Z","lastTransitionTime":"2025-12-05T06:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.020385 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.020437 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.020450 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.020471 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.020484 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:59Z","lastTransitionTime":"2025-12-05T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.124276 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.124379 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.124401 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.124428 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.124448 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:59Z","lastTransitionTime":"2025-12-05T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.227187 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.227256 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.227268 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.227286 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.227299 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:59Z","lastTransitionTime":"2025-12-05T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.332953 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.333006 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.333018 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.333039 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.333061 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:59Z","lastTransitionTime":"2025-12-05T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.435989 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.436047 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.436060 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.436083 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.436097 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:59Z","lastTransitionTime":"2025-12-05T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.538809 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.538890 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.539069 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.539109 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.539144 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:59Z","lastTransitionTime":"2025-12-05T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.642858 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.642953 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.642980 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.643014 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.643037 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:59Z","lastTransitionTime":"2025-12-05T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.746947 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.747037 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.747153 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.747187 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.747211 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:59Z","lastTransitionTime":"2025-12-05T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.748157 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.748199 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.748199 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.748308 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:55:59 crc kubenswrapper[4997]: E1205 06:55:59.748507 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:55:59 crc kubenswrapper[4997]: E1205 06:55:59.748984 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:55:59 crc kubenswrapper[4997]: E1205 06:55:59.749547 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:55:59 crc kubenswrapper[4997]: E1205 06:55:59.749727 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.768278 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec78c08f-8482-4c7f-b321-903b61c52319\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4ff6e25dbcddad8fb0ae5ead4aaad30ce69d09dfc8b477217a7dd52bb4d3862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03a8cf621e0221d9dc37843dd6dff27479bda2bbd5919953b6ec2cdd11f088c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98126b01cc3216f78df3301553e52376d7dc895cb69ebec44c5681e7b9262d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:59Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.791561 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:59Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.812944 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:59Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.838221 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:59Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.849427 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.849483 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.849500 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.849525 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.849541 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:59Z","lastTransitionTime":"2025-12-05T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.856869 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:59Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.873747 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:59Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.888519 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:59Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.903439 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:59Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.918494 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:59Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.939672 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:59Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.952087 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.952136 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.952154 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.952181 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.952200 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:55:59Z","lastTransitionTime":"2025-12-05T06:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.962476 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:59Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.977535 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:59Z is after 2025-08-24T17:21:41Z" Dec 05 06:55:59 crc kubenswrapper[4997]: I1205 06:55:59.993903 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:55:59Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.007773 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:00Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.027485 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:00Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.050667 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:44Z\\\",\\\"message\\\":\\\"nil\\\\u003e UUID: UUIDName:}]\\\\nI1205 06:55:44.568703 6645 ovnkube.go:599] Stopped ovnkube\\\\nI1205 06:55:44.568739 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1205 06:55:44.568170 6645 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-l5p2z in node crc\\\\nI1205 06:55:44.568775 6645 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-l5p2z after 0 failed attempt(s)\\\\nI1205 06:55:44.568784 6645 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-l5p2z\\\\nI1205 06:55:44.568688 6645 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/api]} name:Service_openshift-apiserver/api_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.37:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {88e20c31-5b8d-4d44-bbd8-dba87b7dbaf0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1205 06:55:44.568837 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:00Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.054938 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.054993 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.055011 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.055040 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.055062 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:00Z","lastTransitionTime":"2025-12-05T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.069089 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:00Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.159226 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.159817 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.159960 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.160111 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.160263 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:00Z","lastTransitionTime":"2025-12-05T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.264352 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.264413 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.264429 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.264454 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.264471 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:00Z","lastTransitionTime":"2025-12-05T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.368716 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.368805 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.368822 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.368847 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.368865 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:00Z","lastTransitionTime":"2025-12-05T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.472822 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.473382 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.473551 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.473762 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.473942 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:00Z","lastTransitionTime":"2025-12-05T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.578061 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.578545 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.578565 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.578594 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.578641 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:00Z","lastTransitionTime":"2025-12-05T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.682963 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.683025 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.683042 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.683070 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.683089 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:00Z","lastTransitionTime":"2025-12-05T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.786877 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.786935 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.786948 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.786969 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.786982 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:00Z","lastTransitionTime":"2025-12-05T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.890527 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.890602 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.890653 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.890686 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.890707 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:00Z","lastTransitionTime":"2025-12-05T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.993471 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.993560 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.993580 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.994087 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:00 crc kubenswrapper[4997]: I1205 06:56:00.994188 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:00Z","lastTransitionTime":"2025-12-05T06:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.098823 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.098869 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.098882 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.098907 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.098926 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:01Z","lastTransitionTime":"2025-12-05T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.201872 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.201923 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.201934 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.201958 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.201971 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:01Z","lastTransitionTime":"2025-12-05T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.305595 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.305669 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.305680 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.305700 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.305713 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:01Z","lastTransitionTime":"2025-12-05T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.408810 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.408895 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.408913 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.408944 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.408964 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:01Z","lastTransitionTime":"2025-12-05T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.511941 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.512016 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.512034 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.512054 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.512068 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:01Z","lastTransitionTime":"2025-12-05T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.614803 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.614873 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.614891 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.614914 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.614928 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:01Z","lastTransitionTime":"2025-12-05T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.717782 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.717893 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.717912 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.717986 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.718007 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:01Z","lastTransitionTime":"2025-12-05T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.748886 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.749017 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:01 crc kubenswrapper[4997]: E1205 06:56:01.749072 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.749085 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.748886 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:01 crc kubenswrapper[4997]: E1205 06:56:01.749168 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:01 crc kubenswrapper[4997]: E1205 06:56:01.749243 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:01 crc kubenswrapper[4997]: E1205 06:56:01.749312 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.821038 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.821117 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.821140 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.821175 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.821195 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:01Z","lastTransitionTime":"2025-12-05T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.923825 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.923891 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.923913 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.923946 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:01 crc kubenswrapper[4997]: I1205 06:56:01.923969 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:01Z","lastTransitionTime":"2025-12-05T06:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.026914 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.026960 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.026973 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.026995 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.027011 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:02Z","lastTransitionTime":"2025-12-05T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.129586 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.129676 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.129715 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.129753 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.129778 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:02Z","lastTransitionTime":"2025-12-05T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.232974 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.233017 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.233026 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.233039 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.233049 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:02Z","lastTransitionTime":"2025-12-05T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.335106 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.335207 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.335217 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.335231 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.335240 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:02Z","lastTransitionTime":"2025-12-05T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.439151 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.439190 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.439199 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.439218 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.439228 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:02Z","lastTransitionTime":"2025-12-05T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.497206 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.497265 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.497275 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.497291 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.497301 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:02Z","lastTransitionTime":"2025-12-05T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:02 crc kubenswrapper[4997]: E1205 06:56:02.519478 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:02Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.523149 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.523289 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.523307 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.523320 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.523330 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:02Z","lastTransitionTime":"2025-12-05T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:02 crc kubenswrapper[4997]: E1205 06:56:02.538424 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:02Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.543045 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.543081 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.543091 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.543111 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.543123 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:02Z","lastTransitionTime":"2025-12-05T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:02 crc kubenswrapper[4997]: E1205 06:56:02.563738 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:02Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.568865 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.568901 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.568917 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.568938 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.568953 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:02Z","lastTransitionTime":"2025-12-05T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:02 crc kubenswrapper[4997]: E1205 06:56:02.583331 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:02Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.588848 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.588884 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.588895 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.588912 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.588923 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:02Z","lastTransitionTime":"2025-12-05T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:02 crc kubenswrapper[4997]: E1205 06:56:02.606524 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:02Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:02 crc kubenswrapper[4997]: E1205 06:56:02.606755 4997 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.608451 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.608482 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.608493 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.608515 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.608528 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:02Z","lastTransitionTime":"2025-12-05T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.711605 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.711696 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.711708 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.711725 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.711737 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:02Z","lastTransitionTime":"2025-12-05T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.821973 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.822043 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.822052 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.822071 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.822081 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:02Z","lastTransitionTime":"2025-12-05T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.924128 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.924167 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.924176 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.924191 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:02 crc kubenswrapper[4997]: I1205 06:56:02.924201 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:02Z","lastTransitionTime":"2025-12-05T06:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.027766 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.027813 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.027823 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.027840 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.027848 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:03Z","lastTransitionTime":"2025-12-05T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.130428 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.130473 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.130483 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.130499 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.130508 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:03Z","lastTransitionTime":"2025-12-05T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.232917 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.232952 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.232960 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.232973 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.232983 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:03Z","lastTransitionTime":"2025-12-05T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.334969 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.335035 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.335074 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.335092 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.335103 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:03Z","lastTransitionTime":"2025-12-05T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.438386 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.438461 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.438481 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.438510 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.438528 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:03Z","lastTransitionTime":"2025-12-05T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.540840 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.540901 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.540917 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.540933 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.540947 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:03Z","lastTransitionTime":"2025-12-05T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.643869 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.643926 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.643940 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.643959 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.643970 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:03Z","lastTransitionTime":"2025-12-05T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.746806 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.746886 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.746898 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.746914 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.746923 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:03Z","lastTransitionTime":"2025-12-05T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.748741 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.748862 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:03 crc kubenswrapper[4997]: E1205 06:56:03.748995 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.749053 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.749159 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:03 crc kubenswrapper[4997]: E1205 06:56:03.749199 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:03 crc kubenswrapper[4997]: E1205 06:56:03.749419 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:03 crc kubenswrapper[4997]: E1205 06:56:03.749536 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.750290 4997 scope.go:117] "RemoveContainer" containerID="5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f" Dec 05 06:56:03 crc kubenswrapper[4997]: E1205 06:56:03.750439 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.851078 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.851167 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.851190 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.851227 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.851266 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:03Z","lastTransitionTime":"2025-12-05T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.954405 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.954456 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.954467 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.954484 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:03 crc kubenswrapper[4997]: I1205 06:56:03.954498 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:03Z","lastTransitionTime":"2025-12-05T06:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.056699 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.056734 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.056743 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.056757 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.056765 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:04Z","lastTransitionTime":"2025-12-05T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.158933 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.158957 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.158965 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.158978 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.158987 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:04Z","lastTransitionTime":"2025-12-05T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.208591 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs\") pod \"network-metrics-daemon-b2vlg\" (UID: \"81f6790f-312b-42ce-9bd4-58bac09ce615\") " pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:04 crc kubenswrapper[4997]: E1205 06:56:04.208817 4997 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 06:56:04 crc kubenswrapper[4997]: E1205 06:56:04.208884 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs podName:81f6790f-312b-42ce-9bd4-58bac09ce615 nodeName:}" failed. No retries permitted until 2025-12-05 06:56:36.20886679 +0000 UTC m=+96.737774051 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs") pod "network-metrics-daemon-b2vlg" (UID: "81f6790f-312b-42ce-9bd4-58bac09ce615") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.262725 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.262762 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.262771 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.262792 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.262802 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:04Z","lastTransitionTime":"2025-12-05T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.365048 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.365090 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.365101 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.365121 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.365134 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:04Z","lastTransitionTime":"2025-12-05T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.467047 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.467085 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.467095 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.467112 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.467123 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:04Z","lastTransitionTime":"2025-12-05T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.569009 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.569043 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.569052 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.569065 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.569073 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:04Z","lastTransitionTime":"2025-12-05T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.672227 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.672273 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.672284 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.672300 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.672309 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:04Z","lastTransitionTime":"2025-12-05T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.775400 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.775450 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.775459 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.775474 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.775483 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:04Z","lastTransitionTime":"2025-12-05T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.877642 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.877675 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.877683 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.877699 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.877710 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:04Z","lastTransitionTime":"2025-12-05T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.980420 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.980472 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.980481 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.980504 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:04 crc kubenswrapper[4997]: I1205 06:56:04.980514 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:04Z","lastTransitionTime":"2025-12-05T06:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.083591 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.083667 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.083684 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.083725 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.083735 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:05Z","lastTransitionTime":"2025-12-05T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.186247 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-th9dq_85f05eb3-df7f-4c46-a0e2-451766bb3297/kube-multus/0.log" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.186307 4997 generic.go:334] "Generic (PLEG): container finished" podID="85f05eb3-df7f-4c46-a0e2-451766bb3297" containerID="7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6" exitCode=1 Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.186248 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.186336 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-th9dq" event={"ID":"85f05eb3-df7f-4c46-a0e2-451766bb3297","Type":"ContainerDied","Data":"7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6"} Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.186353 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.186365 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.186382 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.186393 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:05Z","lastTransitionTime":"2025-12-05T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.186920 4997 scope.go:117] "RemoveContainer" containerID="7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.199933 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:05Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.213718 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:05Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.226731 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:05Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.239906 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:05Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.252187 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:05Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.264878 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:05Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.277097 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:56:04Z\\\",\\\"message\\\":\\\"2025-12-05T06:55:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0334bbb5-cde0-49ba-9aaf-2cf74f71ef63\\\\n2025-12-05T06:55:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0334bbb5-cde0-49ba-9aaf-2cf74f71ef63 to /host/opt/cni/bin/\\\\n2025-12-05T06:55:19Z [verbose] multus-daemon started\\\\n2025-12-05T06:55:19Z [verbose] Readiness Indicator file check\\\\n2025-12-05T06:56:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:05Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.289406 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.289443 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.289450 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.289465 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.289473 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:05Z","lastTransitionTime":"2025-12-05T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.291863 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:05Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.313791 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:44Z\\\",\\\"message\\\":\\\"nil\\\\u003e UUID: UUIDName:}]\\\\nI1205 06:55:44.568703 6645 ovnkube.go:599] Stopped ovnkube\\\\nI1205 06:55:44.568739 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1205 06:55:44.568170 6645 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-l5p2z in node crc\\\\nI1205 06:55:44.568775 6645 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-l5p2z after 0 failed attempt(s)\\\\nI1205 06:55:44.568784 6645 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-l5p2z\\\\nI1205 06:55:44.568688 6645 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/api]} name:Service_openshift-apiserver/api_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.37:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {88e20c31-5b8d-4d44-bbd8-dba87b7dbaf0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1205 06:55:44.568837 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:05Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.326415 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:05Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.338365 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec78c08f-8482-4c7f-b321-903b61c52319\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4ff6e25dbcddad8fb0ae5ead4aaad30ce69d09dfc8b477217a7dd52bb4d3862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03a8cf621e0221d9dc37843dd6dff27479bda2bbd5919953b6ec2cdd11f088c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98126b01cc3216f78df3301553e52376d7dc895cb69ebec44c5681e7b9262d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:05Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.352287 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:05Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.374412 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:05Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.388219 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:05Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.398705 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.398743 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.398753 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.398774 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.398787 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:05Z","lastTransitionTime":"2025-12-05T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.402286 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:05Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.412046 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:05Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.423289 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:05Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.501750 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.501806 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.501823 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.501846 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.501859 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:05Z","lastTransitionTime":"2025-12-05T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.604554 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.604653 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.604669 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.604691 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.604728 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:05Z","lastTransitionTime":"2025-12-05T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.708253 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.708298 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.708308 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.708327 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.708338 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:05Z","lastTransitionTime":"2025-12-05T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.748976 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.749030 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.749131 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:05 crc kubenswrapper[4997]: E1205 06:56:05.749240 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.749265 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:05 crc kubenswrapper[4997]: E1205 06:56:05.749498 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:05 crc kubenswrapper[4997]: E1205 06:56:05.749651 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:05 crc kubenswrapper[4997]: E1205 06:56:05.749775 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.810079 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.810111 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.810120 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.810133 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.810142 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:05Z","lastTransitionTime":"2025-12-05T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.912646 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.912681 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.912690 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.912703 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:05 crc kubenswrapper[4997]: I1205 06:56:05.912711 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:05Z","lastTransitionTime":"2025-12-05T06:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.015350 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.015403 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.015415 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.015432 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.015442 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:06Z","lastTransitionTime":"2025-12-05T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.117965 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.118014 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.118023 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.118038 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.118046 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:06Z","lastTransitionTime":"2025-12-05T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.191735 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-th9dq_85f05eb3-df7f-4c46-a0e2-451766bb3297/kube-multus/0.log" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.191790 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-th9dq" event={"ID":"85f05eb3-df7f-4c46-a0e2-451766bb3297","Type":"ContainerStarted","Data":"ba8817aab2ae020cc998813edb099030cc3c3c6eb11e8f175e3cfd82b18686b2"} Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.204467 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:06Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.215234 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:06Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.223161 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.223192 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.223202 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.223216 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.223225 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:06Z","lastTransitionTime":"2025-12-05T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.225534 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:06Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.241055 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:06Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.255374 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:06Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.276191 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:06Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.293825 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:06Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.309668 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:06Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.323712 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:06Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.325527 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.325565 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.325578 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.325602 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.325632 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:06Z","lastTransitionTime":"2025-12-05T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.343896 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:06Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.370194 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:44Z\\\",\\\"message\\\":\\\"nil\\\\u003e UUID: UUIDName:}]\\\\nI1205 06:55:44.568703 6645 ovnkube.go:599] Stopped ovnkube\\\\nI1205 06:55:44.568739 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1205 06:55:44.568170 6645 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-l5p2z in node crc\\\\nI1205 06:55:44.568775 6645 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-l5p2z after 0 failed attempt(s)\\\\nI1205 06:55:44.568784 6645 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-l5p2z\\\\nI1205 06:55:44.568688 6645 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/api]} name:Service_openshift-apiserver/api_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.37:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {88e20c31-5b8d-4d44-bbd8-dba87b7dbaf0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1205 06:55:44.568837 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:06Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.384002 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:06Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.400147 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8817aab2ae020cc998813edb099030cc3c3c6eb11e8f175e3cfd82b18686b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:56:04Z\\\",\\\"message\\\":\\\"2025-12-05T06:55:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0334bbb5-cde0-49ba-9aaf-2cf74f71ef63\\\\n2025-12-05T06:55:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0334bbb5-cde0-49ba-9aaf-2cf74f71ef63 to /host/opt/cni/bin/\\\\n2025-12-05T06:55:19Z [verbose] multus-daemon started\\\\n2025-12-05T06:55:19Z [verbose] Readiness Indicator file check\\\\n2025-12-05T06:56:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:06Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.415490 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:06Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.425345 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:06Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.428206 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.428247 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.428264 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.428286 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.428300 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:06Z","lastTransitionTime":"2025-12-05T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.438068 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:06Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.448593 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec78c08f-8482-4c7f-b321-903b61c52319\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4ff6e25dbcddad8fb0ae5ead4aaad30ce69d09dfc8b477217a7dd52bb4d3862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03a8cf621e0221d9dc37843dd6dff27479bda2bbd5919953b6ec2cdd11f088c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98126b01cc3216f78df3301553e52376d7dc895cb69ebec44c5681e7b9262d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:06Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.530198 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.530229 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.530239 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.530260 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.530271 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:06Z","lastTransitionTime":"2025-12-05T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.632755 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.632787 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.632799 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.632817 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.632828 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:06Z","lastTransitionTime":"2025-12-05T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.734977 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.735019 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.735030 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.735047 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.735056 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:06Z","lastTransitionTime":"2025-12-05T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.837109 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.837187 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.837326 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.837370 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.837395 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:06Z","lastTransitionTime":"2025-12-05T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.939525 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.939588 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.939609 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.939674 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:06 crc kubenswrapper[4997]: I1205 06:56:06.939707 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:06Z","lastTransitionTime":"2025-12-05T06:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.042073 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.042147 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.042170 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.042202 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.042228 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:07Z","lastTransitionTime":"2025-12-05T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.144093 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.144136 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.144149 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.144166 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.144176 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:07Z","lastTransitionTime":"2025-12-05T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.245909 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.245945 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.245955 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.245970 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.245981 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:07Z","lastTransitionTime":"2025-12-05T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.352068 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.352099 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.352107 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.352120 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.352128 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:07Z","lastTransitionTime":"2025-12-05T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.455063 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.455102 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.455116 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.455134 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.455147 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:07Z","lastTransitionTime":"2025-12-05T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.557293 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.557328 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.557337 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.557353 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.557362 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:07Z","lastTransitionTime":"2025-12-05T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.659427 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.659472 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.659483 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.659499 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.659514 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:07Z","lastTransitionTime":"2025-12-05T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.748203 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.748269 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.748226 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.748226 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:07 crc kubenswrapper[4997]: E1205 06:56:07.748347 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:07 crc kubenswrapper[4997]: E1205 06:56:07.748425 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:07 crc kubenswrapper[4997]: E1205 06:56:07.748485 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:07 crc kubenswrapper[4997]: E1205 06:56:07.748572 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.762184 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.762226 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.762262 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.762281 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.762292 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:07Z","lastTransitionTime":"2025-12-05T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.864550 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.864603 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.864656 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.864678 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.864693 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:07Z","lastTransitionTime":"2025-12-05T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.966893 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.966936 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.966946 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.966964 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:07 crc kubenswrapper[4997]: I1205 06:56:07.966975 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:07Z","lastTransitionTime":"2025-12-05T06:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.068948 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.068988 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.068999 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.069017 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.069029 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:08Z","lastTransitionTime":"2025-12-05T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.171132 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.171186 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.171198 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.171214 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.171225 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:08Z","lastTransitionTime":"2025-12-05T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.274045 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.274082 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.274090 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.274104 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.274113 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:08Z","lastTransitionTime":"2025-12-05T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.377166 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.377210 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.377219 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.377236 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.377248 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:08Z","lastTransitionTime":"2025-12-05T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.480209 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.480251 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.480262 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.480278 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.480290 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:08Z","lastTransitionTime":"2025-12-05T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.582593 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.582654 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.582663 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.582679 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.582694 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:08Z","lastTransitionTime":"2025-12-05T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.684868 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.684923 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.684939 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.684961 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.684976 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:08Z","lastTransitionTime":"2025-12-05T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.787238 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.787278 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.787290 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.787307 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.787318 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:08Z","lastTransitionTime":"2025-12-05T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.890002 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.890029 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.890039 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.890056 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.890066 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:08Z","lastTransitionTime":"2025-12-05T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.992179 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.992224 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.992236 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.992254 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:08 crc kubenswrapper[4997]: I1205 06:56:08.992269 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:08Z","lastTransitionTime":"2025-12-05T06:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.094259 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.094307 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.094321 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.094337 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.094348 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:09Z","lastTransitionTime":"2025-12-05T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.196334 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.196536 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.196544 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.196556 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.196565 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:09Z","lastTransitionTime":"2025-12-05T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.298657 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.298717 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.298740 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.298769 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.298791 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:09Z","lastTransitionTime":"2025-12-05T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.401054 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.401096 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.401108 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.401123 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.401131 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:09Z","lastTransitionTime":"2025-12-05T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.503310 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.503344 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.503353 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.503368 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.503376 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:09Z","lastTransitionTime":"2025-12-05T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.605258 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.605296 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.605304 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.605318 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.605329 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:09Z","lastTransitionTime":"2025-12-05T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.707190 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.707224 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.707236 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.707251 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.707262 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:09Z","lastTransitionTime":"2025-12-05T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.748737 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.748783 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:09 crc kubenswrapper[4997]: E1205 06:56:09.748874 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:09 crc kubenswrapper[4997]: E1205 06:56:09.748958 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.749013 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.749017 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:09 crc kubenswrapper[4997]: E1205 06:56:09.749157 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:09 crc kubenswrapper[4997]: E1205 06:56:09.749206 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.766239 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:09Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.776423 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec78c08f-8482-4c7f-b321-903b61c52319\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4ff6e25dbcddad8fb0ae5ead4aaad30ce69d09dfc8b477217a7dd52bb4d3862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03a8cf621e0221d9dc37843dd6dff27479bda2bbd5919953b6ec2cdd11f088c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98126b01cc3216f78df3301553e52376d7dc895cb69ebec44c5681e7b9262d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:09Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.788913 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:09Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.801097 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:09Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.809396 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.809429 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.809437 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.809452 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.809465 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:09Z","lastTransitionTime":"2025-12-05T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.815982 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:09Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.826050 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:09Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.836729 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:09Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.844778 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:09Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.854522 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:09Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.867764 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:09Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.879076 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:09Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.888013 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:09Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.898191 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:09Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.908229 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:09Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.916555 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.916598 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.916641 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.916872 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.916885 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:09Z","lastTransitionTime":"2025-12-05T06:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.924774 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8817aab2ae020cc998813edb099030cc3c3c6eb11e8f175e3cfd82b18686b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:56:04Z\\\",\\\"message\\\":\\\"2025-12-05T06:55:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0334bbb5-cde0-49ba-9aaf-2cf74f71ef63\\\\n2025-12-05T06:55:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0334bbb5-cde0-49ba-9aaf-2cf74f71ef63 to /host/opt/cni/bin/\\\\n2025-12-05T06:55:19Z [verbose] multus-daemon started\\\\n2025-12-05T06:55:19Z [verbose] Readiness Indicator file check\\\\n2025-12-05T06:56:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:09Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.937525 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:09Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:09 crc kubenswrapper[4997]: I1205 06:56:09.953424 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:44Z\\\",\\\"message\\\":\\\"nil\\\\u003e UUID: UUIDName:}]\\\\nI1205 06:55:44.568703 6645 ovnkube.go:599] Stopped ovnkube\\\\nI1205 06:55:44.568739 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1205 06:55:44.568170 6645 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-l5p2z in node crc\\\\nI1205 06:55:44.568775 6645 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-l5p2z after 0 failed attempt(s)\\\\nI1205 06:55:44.568784 6645 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-l5p2z\\\\nI1205 06:55:44.568688 6645 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/api]} name:Service_openshift-apiserver/api_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.37:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {88e20c31-5b8d-4d44-bbd8-dba87b7dbaf0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1205 06:55:44.568837 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:09Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.018913 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.018940 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.018947 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.018960 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.018968 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:10Z","lastTransitionTime":"2025-12-05T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.121959 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.122001 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.122014 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.122032 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.122043 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:10Z","lastTransitionTime":"2025-12-05T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.225895 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.226007 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.226025 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.226119 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.226139 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:10Z","lastTransitionTime":"2025-12-05T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.328419 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.328671 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.328756 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.328849 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.328916 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:10Z","lastTransitionTime":"2025-12-05T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.431773 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.431813 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.431823 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.431844 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.431857 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:10Z","lastTransitionTime":"2025-12-05T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.534187 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.534235 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.534247 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.534267 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.534279 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:10Z","lastTransitionTime":"2025-12-05T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.637441 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.637708 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.637785 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.637850 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.637905 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:10Z","lastTransitionTime":"2025-12-05T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.740267 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.740307 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.740316 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.740331 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.740341 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:10Z","lastTransitionTime":"2025-12-05T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.842674 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.842941 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.843032 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.843128 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.843215 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:10Z","lastTransitionTime":"2025-12-05T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.945563 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.945604 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.945631 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.945646 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:10 crc kubenswrapper[4997]: I1205 06:56:10.945654 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:10Z","lastTransitionTime":"2025-12-05T06:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.048760 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.048843 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.048858 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.048876 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.048888 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:11Z","lastTransitionTime":"2025-12-05T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.152009 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.152048 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.152057 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.152076 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.152087 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:11Z","lastTransitionTime":"2025-12-05T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.254889 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.254924 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.254945 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.254963 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.254976 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:11Z","lastTransitionTime":"2025-12-05T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.357131 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.357162 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.357170 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.357191 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.357207 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:11Z","lastTransitionTime":"2025-12-05T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.459427 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.459490 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.459506 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.459530 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.459548 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:11Z","lastTransitionTime":"2025-12-05T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.561738 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.561781 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.561790 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.561804 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.561815 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:11Z","lastTransitionTime":"2025-12-05T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.664704 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.664750 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.664761 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.664776 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.664787 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:11Z","lastTransitionTime":"2025-12-05T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.748097 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.748143 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.748144 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.748144 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:11 crc kubenswrapper[4997]: E1205 06:56:11.748220 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:11 crc kubenswrapper[4997]: E1205 06:56:11.748293 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:11 crc kubenswrapper[4997]: E1205 06:56:11.748332 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:11 crc kubenswrapper[4997]: E1205 06:56:11.748440 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.766264 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.766310 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.766327 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.766347 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.766365 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:11Z","lastTransitionTime":"2025-12-05T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.868381 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.868423 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.868432 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.868447 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.868455 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:11Z","lastTransitionTime":"2025-12-05T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.971402 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.971451 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.971463 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.971479 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:11 crc kubenswrapper[4997]: I1205 06:56:11.971489 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:11Z","lastTransitionTime":"2025-12-05T06:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.073961 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.074001 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.074009 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.074025 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.074034 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:12Z","lastTransitionTime":"2025-12-05T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.175926 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.175964 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.175974 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.175993 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.176004 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:12Z","lastTransitionTime":"2025-12-05T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.278214 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.278278 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.278292 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.278314 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.278330 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:12Z","lastTransitionTime":"2025-12-05T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.380875 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.380920 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.380928 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.380945 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.380954 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:12Z","lastTransitionTime":"2025-12-05T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.482999 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.483035 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.483043 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.483056 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.483065 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:12Z","lastTransitionTime":"2025-12-05T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.585771 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.586074 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.586177 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.586286 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.586374 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:12Z","lastTransitionTime":"2025-12-05T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.688175 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.688416 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.688514 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.688632 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.688776 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:12Z","lastTransitionTime":"2025-12-05T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.790642 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.790683 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.790692 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.790705 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.790714 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:12Z","lastTransitionTime":"2025-12-05T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.892761 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.892809 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.892818 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.892832 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.892859 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:12Z","lastTransitionTime":"2025-12-05T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.926769 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.926820 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.926838 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.926863 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.926880 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:12Z","lastTransitionTime":"2025-12-05T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:12 crc kubenswrapper[4997]: E1205 06:56:12.946561 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:12Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.950417 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.950459 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.950470 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.950486 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.950496 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:12Z","lastTransitionTime":"2025-12-05T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:12 crc kubenswrapper[4997]: E1205 06:56:12.963627 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:12Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.967701 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.967735 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.967744 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.967760 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.967769 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:12Z","lastTransitionTime":"2025-12-05T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:12 crc kubenswrapper[4997]: E1205 06:56:12.978468 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:12Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.981208 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.981245 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.981255 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.981272 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.981284 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:12Z","lastTransitionTime":"2025-12-05T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:12 crc kubenswrapper[4997]: E1205 06:56:12.995876 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:12Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.998905 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.998941 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.998950 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.998964 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:12 crc kubenswrapper[4997]: I1205 06:56:12.999009 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:12Z","lastTransitionTime":"2025-12-05T06:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:13 crc kubenswrapper[4997]: E1205 06:56:13.009541 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:13Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:13 crc kubenswrapper[4997]: E1205 06:56:13.009662 4997 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.010700 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.010756 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.010766 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.010779 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.010787 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:13Z","lastTransitionTime":"2025-12-05T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.113183 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.113223 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.113232 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.113250 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.113261 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:13Z","lastTransitionTime":"2025-12-05T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.216230 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.216269 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.216280 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.216297 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.216311 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:13Z","lastTransitionTime":"2025-12-05T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.318287 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.318341 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.318356 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.318372 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.318386 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:13Z","lastTransitionTime":"2025-12-05T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.420669 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.420707 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.420715 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.420733 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.420743 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:13Z","lastTransitionTime":"2025-12-05T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.522960 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.523004 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.523018 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.523037 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.523050 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:13Z","lastTransitionTime":"2025-12-05T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.625787 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.625885 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.625903 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.625925 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.625980 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:13Z","lastTransitionTime":"2025-12-05T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.728028 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.728070 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.728081 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.728097 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.728108 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:13Z","lastTransitionTime":"2025-12-05T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.749217 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.749314 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:13 crc kubenswrapper[4997]: E1205 06:56:13.749659 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.749365 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.749327 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:13 crc kubenswrapper[4997]: E1205 06:56:13.749759 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:13 crc kubenswrapper[4997]: E1205 06:56:13.749558 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:13 crc kubenswrapper[4997]: E1205 06:56:13.750048 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.830014 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.830062 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.830075 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.830090 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.830119 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:13Z","lastTransitionTime":"2025-12-05T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.932364 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.932647 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.932721 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.932792 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:13 crc kubenswrapper[4997]: I1205 06:56:13.932848 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:13Z","lastTransitionTime":"2025-12-05T06:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.035447 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.035489 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.035501 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.035518 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.035531 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:14Z","lastTransitionTime":"2025-12-05T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.137447 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.137473 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.137481 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.137494 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.137502 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:14Z","lastTransitionTime":"2025-12-05T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.239872 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.239912 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.239921 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.239935 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.239945 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:14Z","lastTransitionTime":"2025-12-05T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.342484 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.342515 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.342526 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.342543 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.342553 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:14Z","lastTransitionTime":"2025-12-05T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.445460 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.445496 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.445507 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.445526 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.445538 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:14Z","lastTransitionTime":"2025-12-05T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.548605 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.548693 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.548711 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.548734 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.548755 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:14Z","lastTransitionTime":"2025-12-05T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.651367 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.651461 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.651475 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.651496 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.651508 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:14Z","lastTransitionTime":"2025-12-05T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.754204 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.754427 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.754493 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.754560 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.754660 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:14Z","lastTransitionTime":"2025-12-05T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.856952 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.856998 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.857009 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.857025 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.857035 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:14Z","lastTransitionTime":"2025-12-05T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.959970 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.960008 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.960017 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.960031 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:14 crc kubenswrapper[4997]: I1205 06:56:14.960041 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:14Z","lastTransitionTime":"2025-12-05T06:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.062699 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.062740 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.062751 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.062776 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.062790 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:15Z","lastTransitionTime":"2025-12-05T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.166147 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.166225 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.166234 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.166251 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.166260 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:15Z","lastTransitionTime":"2025-12-05T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.268541 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.269114 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.269182 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.269254 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.269311 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:15Z","lastTransitionTime":"2025-12-05T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.371911 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.372346 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.372410 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.372481 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.372553 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:15Z","lastTransitionTime":"2025-12-05T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.475158 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.475198 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.475209 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.475227 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.475239 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:15Z","lastTransitionTime":"2025-12-05T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.577919 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.578164 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.578263 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.578330 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.578394 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:15Z","lastTransitionTime":"2025-12-05T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.682386 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.682788 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.682862 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.682975 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.683047 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:15Z","lastTransitionTime":"2025-12-05T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.748718 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.748745 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:15 crc kubenswrapper[4997]: E1205 06:56:15.749536 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.748962 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:15 crc kubenswrapper[4997]: E1205 06:56:15.749368 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.748954 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:15 crc kubenswrapper[4997]: E1205 06:56:15.749980 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:15 crc kubenswrapper[4997]: E1205 06:56:15.749898 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.786014 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.786071 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.786081 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.786099 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.786111 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:15Z","lastTransitionTime":"2025-12-05T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.888533 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.888578 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.888590 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.888624 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:15 crc kubenswrapper[4997]: I1205 06:56:15.888638 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:15Z","lastTransitionTime":"2025-12-05T06:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.004228 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.004316 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.004345 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.004387 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.004410 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:16Z","lastTransitionTime":"2025-12-05T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.107942 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.107995 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.108008 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.108028 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.108038 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:16Z","lastTransitionTime":"2025-12-05T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.210815 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.210900 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.210931 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.210994 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.211024 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:16Z","lastTransitionTime":"2025-12-05T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.314029 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.314077 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.314088 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.314110 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.314121 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:16Z","lastTransitionTime":"2025-12-05T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.417919 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.417985 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.417997 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.418019 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.418031 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:16Z","lastTransitionTime":"2025-12-05T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.520900 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.520970 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.520990 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.521017 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.521032 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:16Z","lastTransitionTime":"2025-12-05T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.624509 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.624596 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.624608 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.624649 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.624664 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:16Z","lastTransitionTime":"2025-12-05T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.727410 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.727440 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.727450 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.727464 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.727473 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:16Z","lastTransitionTime":"2025-12-05T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.830710 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.830757 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.830774 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.830791 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.830801 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:16Z","lastTransitionTime":"2025-12-05T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.933868 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.933921 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.933932 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.933954 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:16 crc kubenswrapper[4997]: I1205 06:56:16.933970 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:16Z","lastTransitionTime":"2025-12-05T06:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.036096 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.036139 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.036150 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.036165 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.036174 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:17Z","lastTransitionTime":"2025-12-05T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.139980 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.140060 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.140086 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.140123 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.140147 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:17Z","lastTransitionTime":"2025-12-05T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.242765 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.242821 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.242835 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.242855 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.242869 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:17Z","lastTransitionTime":"2025-12-05T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.345430 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.345578 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.345679 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.345725 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.345770 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:17Z","lastTransitionTime":"2025-12-05T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.448802 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.448845 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.448857 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.448873 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.448885 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:17Z","lastTransitionTime":"2025-12-05T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.550755 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.550802 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.550831 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.550866 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.550878 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:17Z","lastTransitionTime":"2025-12-05T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.653010 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.653050 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.653059 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.653074 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.653084 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:17Z","lastTransitionTime":"2025-12-05T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.748856 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.748891 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.748940 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.748991 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:17 crc kubenswrapper[4997]: E1205 06:56:17.749130 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:17 crc kubenswrapper[4997]: E1205 06:56:17.749243 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:17 crc kubenswrapper[4997]: E1205 06:56:17.749329 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:17 crc kubenswrapper[4997]: E1205 06:56:17.749394 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.750172 4997 scope.go:117] "RemoveContainer" containerID="5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.754415 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.754449 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.754460 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.754478 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.754493 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:17Z","lastTransitionTime":"2025-12-05T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.856962 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.857596 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.857671 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.857697 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.857709 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:17Z","lastTransitionTime":"2025-12-05T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.959713 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.959754 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.959765 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.959781 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:17 crc kubenswrapper[4997]: I1205 06:56:17.959793 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:17Z","lastTransitionTime":"2025-12-05T06:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.061638 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.061671 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.061678 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.061691 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.061700 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:18Z","lastTransitionTime":"2025-12-05T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.163787 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.163830 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.163839 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.163855 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.163865 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:18Z","lastTransitionTime":"2025-12-05T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.266365 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.266414 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.266423 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.266436 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.266445 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:18Z","lastTransitionTime":"2025-12-05T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.368473 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.368514 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.368523 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.368539 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.368550 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:18Z","lastTransitionTime":"2025-12-05T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.471886 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.471928 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.471939 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.471958 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.471969 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:18Z","lastTransitionTime":"2025-12-05T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.574774 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.574818 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.574829 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.574843 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.574852 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:18Z","lastTransitionTime":"2025-12-05T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.677393 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.677489 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.677506 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.677533 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.677550 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:18Z","lastTransitionTime":"2025-12-05T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.780676 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.780740 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.780757 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.780782 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.780799 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:18Z","lastTransitionTime":"2025-12-05T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.882970 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.883011 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.883018 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.883032 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.883042 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:18Z","lastTransitionTime":"2025-12-05T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.987803 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.987851 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.987862 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.987879 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:18 crc kubenswrapper[4997]: I1205 06:56:18.987895 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:18Z","lastTransitionTime":"2025-12-05T06:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.091320 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.091367 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.091378 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.091395 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.091406 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:19Z","lastTransitionTime":"2025-12-05T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.193717 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.193788 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.193803 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.193823 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.193836 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:19Z","lastTransitionTime":"2025-12-05T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.296176 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.296226 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.296238 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.296258 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.296270 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:19Z","lastTransitionTime":"2025-12-05T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.399511 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.399539 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.399547 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.399560 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.399569 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:19Z","lastTransitionTime":"2025-12-05T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.502015 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.502063 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.502076 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.502094 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.502105 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:19Z","lastTransitionTime":"2025-12-05T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.604647 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.604676 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.604684 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.604716 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.604728 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:19Z","lastTransitionTime":"2025-12-05T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.707036 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.707071 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.707079 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.707092 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.707100 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:19Z","lastTransitionTime":"2025-12-05T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.748840 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:19 crc kubenswrapper[4997]: E1205 06:56:19.748987 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.749059 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.749190 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:19 crc kubenswrapper[4997]: E1205 06:56:19.749258 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:19 crc kubenswrapper[4997]: E1205 06:56:19.749473 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.748890 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:19 crc kubenswrapper[4997]: E1205 06:56:19.749788 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.760475 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.769717 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.781968 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.792137 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.802913 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.809180 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.809217 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.809231 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.809254 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.809266 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:19Z","lastTransitionTime":"2025-12-05T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.810751 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.820584 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.831705 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.845902 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.865836 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:44Z\\\",\\\"message\\\":\\\"nil\\\\u003e UUID: UUIDName:}]\\\\nI1205 06:55:44.568703 6645 ovnkube.go:599] Stopped ovnkube\\\\nI1205 06:55:44.568739 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1205 06:55:44.568170 6645 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-l5p2z in node crc\\\\nI1205 06:55:44.568775 6645 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-l5p2z after 0 failed attempt(s)\\\\nI1205 06:55:44.568784 6645 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-l5p2z\\\\nI1205 06:55:44.568688 6645 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/api]} name:Service_openshift-apiserver/api_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.37:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {88e20c31-5b8d-4d44-bbd8-dba87b7dbaf0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1205 06:55:44.568837 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.879971 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.895858 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8817aab2ae020cc998813edb099030cc3c3c6eb11e8f175e3cfd82b18686b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:56:04Z\\\",\\\"message\\\":\\\"2025-12-05T06:55:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0334bbb5-cde0-49ba-9aaf-2cf74f71ef63\\\\n2025-12-05T06:55:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0334bbb5-cde0-49ba-9aaf-2cf74f71ef63 to /host/opt/cni/bin/\\\\n2025-12-05T06:55:19Z [verbose] multus-daemon started\\\\n2025-12-05T06:55:19Z [verbose] Readiness Indicator file check\\\\n2025-12-05T06:56:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.911455 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.911485 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.911495 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.911510 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.911521 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:19Z","lastTransitionTime":"2025-12-05T06:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.918116 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.932869 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.945351 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.959955 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec78c08f-8482-4c7f-b321-903b61c52319\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4ff6e25dbcddad8fb0ae5ead4aaad30ce69d09dfc8b477217a7dd52bb4d3862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03a8cf621e0221d9dc37843dd6dff27479bda2bbd5919953b6ec2cdd11f088c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98126b01cc3216f78df3301553e52376d7dc895cb69ebec44c5681e7b9262d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:19 crc kubenswrapper[4997]: I1205 06:56:19.972950 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:19Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.013928 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.013956 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.013964 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.013978 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.013990 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:20Z","lastTransitionTime":"2025-12-05T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.116543 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.116582 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.116593 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.116607 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.116635 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:20Z","lastTransitionTime":"2025-12-05T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.218535 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.218578 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.218589 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.218605 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.218634 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:20Z","lastTransitionTime":"2025-12-05T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.239996 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovnkube-controller/2.log" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.242548 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerStarted","Data":"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57"} Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.242975 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.256490 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.279630 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:44Z\\\",\\\"message\\\":\\\"nil\\\\u003e UUID: UUIDName:}]\\\\nI1205 06:55:44.568703 6645 ovnkube.go:599] Stopped ovnkube\\\\nI1205 06:55:44.568739 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1205 06:55:44.568170 6645 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-l5p2z in node crc\\\\nI1205 06:55:44.568775 6645 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-l5p2z after 0 failed attempt(s)\\\\nI1205 06:55:44.568784 6645 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-l5p2z\\\\nI1205 06:55:44.568688 6645 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/api]} name:Service_openshift-apiserver/api_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.37:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {88e20c31-5b8d-4d44-bbd8-dba87b7dbaf0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1205 06:55:44.568837 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.291631 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.303082 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8817aab2ae020cc998813edb099030cc3c3c6eb11e8f175e3cfd82b18686b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:56:04Z\\\",\\\"message\\\":\\\"2025-12-05T06:55:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0334bbb5-cde0-49ba-9aaf-2cf74f71ef63\\\\n2025-12-05T06:55:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0334bbb5-cde0-49ba-9aaf-2cf74f71ef63 to /host/opt/cni/bin/\\\\n2025-12-05T06:55:19Z [verbose] multus-daemon started\\\\n2025-12-05T06:55:19Z [verbose] Readiness Indicator file check\\\\n2025-12-05T06:56:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.315782 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.320599 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.320642 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.320650 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.320663 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.320674 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:20Z","lastTransitionTime":"2025-12-05T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.328498 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.339779 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.349432 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec78c08f-8482-4c7f-b321-903b61c52319\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4ff6e25dbcddad8fb0ae5ead4aaad30ce69d09dfc8b477217a7dd52bb4d3862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03a8cf621e0221d9dc37843dd6dff27479bda2bbd5919953b6ec2cdd11f088c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98126b01cc3216f78df3301553e52376d7dc895cb69ebec44c5681e7b9262d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.359472 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.370135 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.380267 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.390758 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.401386 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.412064 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.421254 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.422749 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.422784 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.422798 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.422827 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.422841 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:20Z","lastTransitionTime":"2025-12-05T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.430796 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.443324 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.524912 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.524950 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.524957 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.524972 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.524983 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:20Z","lastTransitionTime":"2025-12-05T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.626984 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.627027 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.627038 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.627055 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.627065 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:20Z","lastTransitionTime":"2025-12-05T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.729003 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.729058 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.729077 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.729099 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.729115 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:20Z","lastTransitionTime":"2025-12-05T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.831226 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.831262 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.831272 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.831292 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.831303 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:20Z","lastTransitionTime":"2025-12-05T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.933514 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.933573 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.933591 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.933646 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:20 crc kubenswrapper[4997]: I1205 06:56:20.933664 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:20Z","lastTransitionTime":"2025-12-05T06:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.036231 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.036263 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.036271 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.036283 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.036292 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:21Z","lastTransitionTime":"2025-12-05T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.138138 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.138180 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.138189 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.138205 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.138216 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:21Z","lastTransitionTime":"2025-12-05T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.240249 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.240281 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.240290 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.240302 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.240311 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:21Z","lastTransitionTime":"2025-12-05T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.246746 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovnkube-controller/3.log" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.247366 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovnkube-controller/2.log" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.249633 4997 generic.go:334] "Generic (PLEG): container finished" podID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerID="efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57" exitCode=1 Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.249752 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerDied","Data":"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57"} Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.249846 4997 scope.go:117] "RemoveContainer" containerID="5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.250223 4997 scope.go:117] "RemoveContainer" containerID="efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57" Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.250363 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.262806 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.273199 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.286236 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.298935 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.309963 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.320858 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.330315 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.338123 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.342194 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.342229 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.342240 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.342255 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.342267 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:21Z","lastTransitionTime":"2025-12-05T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.350566 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.363953 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.375898 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8817aab2ae020cc998813edb099030cc3c3c6eb11e8f175e3cfd82b18686b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:56:04Z\\\",\\\"message\\\":\\\"2025-12-05T06:55:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0334bbb5-cde0-49ba-9aaf-2cf74f71ef63\\\\n2025-12-05T06:55:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0334bbb5-cde0-49ba-9aaf-2cf74f71ef63 to /host/opt/cni/bin/\\\\n2025-12-05T06:55:19Z [verbose] multus-daemon started\\\\n2025-12-05T06:55:19Z [verbose] Readiness Indicator file check\\\\n2025-12-05T06:56:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.390000 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.406912 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5201cd27b1c43fd71565234e4f1fa44a6352aaeba7a887a6808708e5e3e69d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:55:44Z\\\",\\\"message\\\":\\\"nil\\\\u003e UUID: UUIDName:}]\\\\nI1205 06:55:44.568703 6645 ovnkube.go:599] Stopped ovnkube\\\\nI1205 06:55:44.568739 6645 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1205 06:55:44.568170 6645 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-l5p2z in node crc\\\\nI1205 06:55:44.568775 6645 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-l5p2z after 0 failed attempt(s)\\\\nI1205 06:55:44.568784 6645 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-l5p2z\\\\nI1205 06:55:44.568688 6645 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/api]} name:Service_openshift-apiserver/api_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.37:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {88e20c31-5b8d-4d44-bbd8-dba87b7dbaf0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1205 06:55:44.568837 6645 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:56:20Z\\\",\\\"message\\\":\\\"Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1205 06:56:20.285133 7041 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z]\\\\nI1205 06:56:20.285142 7041 services_controller.go:452] Built service openshift-kube-apiserver-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1205 06:56:20.285144 7041 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1205 06:56:20.285153 7041 services_controller.go:453] Built serv\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:56:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.422003 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.434276 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec78c08f-8482-4c7f-b321-903b61c52319\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4ff6e25dbcddad8fb0ae5ead4aaad30ce69d09dfc8b477217a7dd52bb4d3862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03a8cf621e0221d9dc37843dd6dff27479bda2bbd5919953b6ec2cdd11f088c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98126b01cc3216f78df3301553e52376d7dc895cb69ebec44c5681e7b9262d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.444045 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.444074 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.444083 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.444098 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.444142 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:21Z","lastTransitionTime":"2025-12-05T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.448937 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.460527 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:21Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.545851 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.545887 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.545898 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.545920 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.545936 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:21Z","lastTransitionTime":"2025-12-05T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.594790 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.594895 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:25.594875674 +0000 UTC m=+146.123782935 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.594943 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.595009 4997 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.595062 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 06:57:25.595046958 +0000 UTC m=+146.123954219 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.648387 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.648423 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.648434 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.648450 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.648462 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:21Z","lastTransitionTime":"2025-12-05T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.696316 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.696422 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.696477 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.696490 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.696511 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.696524 4997 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.696573 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 06:57:25.696557209 +0000 UTC m=+146.225464470 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.696662 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.696697 4997 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.696692 4997 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.696792 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 06:57:25.696766135 +0000 UTC m=+146.225673436 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.696710 4997 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.696852 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 06:57:25.696843137 +0000 UTC m=+146.225750398 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.748408 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.748412 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.748568 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.748433 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.748682 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.748434 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.748768 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:21 crc kubenswrapper[4997]: E1205 06:56:21.749136 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.750144 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.750172 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.750184 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.750198 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.750208 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:21Z","lastTransitionTime":"2025-12-05T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.851992 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.852042 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.852053 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.852072 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.852085 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:21Z","lastTransitionTime":"2025-12-05T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.954993 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.955060 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.955087 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.955117 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:21 crc kubenswrapper[4997]: I1205 06:56:21.955138 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:21Z","lastTransitionTime":"2025-12-05T06:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.058197 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.058238 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.058249 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.058264 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.058276 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:22Z","lastTransitionTime":"2025-12-05T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.160839 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.160871 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.160883 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.160947 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.160961 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:22Z","lastTransitionTime":"2025-12-05T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.254453 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovnkube-controller/3.log" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.263653 4997 scope.go:117] "RemoveContainer" containerID="efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57" Dec 05 06:56:22 crc kubenswrapper[4997]: E1205 06:56:22.264268 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.265845 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.265880 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.265897 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.265915 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.265930 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:22Z","lastTransitionTime":"2025-12-05T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.282041 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-th9dq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f05eb3-df7f-4c46-a0e2-451766bb3297\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8817aab2ae020cc998813edb099030cc3c3c6eb11e8f175e3cfd82b18686b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:56:04Z\\\",\\\"message\\\":\\\"2025-12-05T06:55:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0334bbb5-cde0-49ba-9aaf-2cf74f71ef63\\\\n2025-12-05T06:55:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0334bbb5-cde0-49ba-9aaf-2cf74f71ef63 to /host/opt/cni/bin/\\\\n2025-12-05T06:55:19Z [verbose] multus-daemon started\\\\n2025-12-05T06:55:19Z [verbose] Readiness Indicator file check\\\\n2025-12-05T06:56:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-th9dq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.302696 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76d5e7c0-c9d4-47c8-8a9e-d96565189197\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0e7dc1bbd4483518995884621490509dd1042c5dab70ca98af67daa141b37bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51ea7e38b1ab614e8902c79448cb68839a9afdef9518457130991483bdc5f55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880aca11b54ef7cce5d7ff153842bb1a697c8e733863ac2eebba80d453515ec5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1243170e268c87e30763b9285e8e41f945c05687959feabac1c87748f97ae550\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90e55a1fcef6fb761c3bea31f70b0efc543abf7bc43f13365c5ff9d300753c98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871bc463a63d16b089c1fb80706ec365b7a84635a58b21372cf56d231e064a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a1d40f6f8ae5766d255a5e634da3cd87cb508074d5ae0e834a29f9a06df35c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qmw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5m9fs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.325739 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28f65c70-f59b-4393-b9c9-8163bb596f6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T06:56:20Z\\\",\\\"message\\\":\\\"Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1205 06:56:20.285133 7041 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:20Z is after 2025-08-24T17:21:41Z]\\\\nI1205 06:56:20.285142 7041 services_controller.go:452] Built service openshift-kube-apiserver-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1205 06:56:20.285144 7041 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1205 06:56:20.285153 7041 services_controller.go:453] Built serv\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:56:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vmkd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bchpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.339015 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6772ab9788df6d01a469896ea90fdcd7debbaf3fdcb91eccc9045afe8d55942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.358735 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec78c08f-8482-4c7f-b321-903b61c52319\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4ff6e25dbcddad8fb0ae5ead4aaad30ce69d09dfc8b477217a7dd52bb4d3862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03a8cf621e0221d9dc37843dd6dff27479bda2bbd5919953b6ec2cdd11f088c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98126b01cc3216f78df3301553e52376d7dc895cb69ebec44c5681e7b9262d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02eeb706114d43201816a6778f933457c3bedb964de9f16936affc8b32e400f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.367942 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.368169 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.368254 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.368351 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.368435 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:22Z","lastTransitionTime":"2025-12-05T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.375308 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd550f2aadb8f3b08ec56e1cf49fa84e71d7a623e2b51bad3071aec50c71a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.388569 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.403568 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"072cc9c7-9e48-446b-99ec-c03892c7e66c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 06:55:12.071430 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 06:55:12.072774 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2746846165/tls.crt::/tmp/serving-cert-2746846165/tls.key\\\\\\\"\\\\nI1205 06:55:17.629810 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 06:55:17.635053 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 06:55:17.635106 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 06:55:17.635167 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 06:55:17.635177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 06:55:17.643469 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1205 06:55:17.643495 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1205 06:55:17.643511 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643520 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 06:55:17.643525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 06:55:17.643529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 06:55:17.643532 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 06:55:17.643535 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1205 06:55:17.651230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T06:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.417363 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ed874a5-1318-4b81-a286-0293c003815f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63ad78131201de33e73d7520cd91471b8e6ee395afe99bade438885dde95d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a694b18ee0bd64268e70dd1075223e63a1bb573c72bacdcba6d3dbc1d2d70c38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcd47c533eddcccd89f2e7c35f2b65bec43d6c87e6e7eb42bb5071269cf0d73a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:54:59Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.431974 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jj4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2df2c8f1-3f95-4b85-825f-d79671f90630\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44f799d7e2e6cc3abd11c847a43034f983a9f1b9201177a1469c9094bee3c2b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2m8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jj4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.445602 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.457044 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.469724 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.471644 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.471678 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.471689 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.471706 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.471718 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:22Z","lastTransitionTime":"2025-12-05T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.481063 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://481e724f47ae9f93a23a794fca0a581ba22bfe35a774cdbf0c256d0392e0a918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b3d6c3aa927773a41b68055a365a9fce734a67a8c2f844c2f4716401a88cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.492819 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.504922 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l5p2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44799328-a533-4c95-9712-683b649c8ebf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8fb0098416c7e9e4a3792fc1a11fc6f92ad3cf60d0593468d53cd661d672114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-frk5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l5p2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.515860 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28cde6da-1de5-455a-93c8-dd250f9ea4d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cb1d9ee4664d005f368c9aa294a4b9c9c4a38c13b7c895ba997f835751d7d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4v4j8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l7hsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:22Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.574453 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.574486 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.574495 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.574509 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.574520 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:22Z","lastTransitionTime":"2025-12-05T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.677076 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.677117 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.677129 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.677146 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.677158 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:22Z","lastTransitionTime":"2025-12-05T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.778910 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.779135 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.779197 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.779300 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.779366 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:22Z","lastTransitionTime":"2025-12-05T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.882351 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.882405 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.882424 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.882447 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.882463 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:22Z","lastTransitionTime":"2025-12-05T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.984976 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.985021 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.985032 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.985046 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:22 crc kubenswrapper[4997]: I1205 06:56:22.985055 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:22Z","lastTransitionTime":"2025-12-05T06:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.087884 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.087943 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.087956 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.088005 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.088020 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:23Z","lastTransitionTime":"2025-12-05T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.190422 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.190465 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.190480 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.190498 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.190513 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:23Z","lastTransitionTime":"2025-12-05T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.221506 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.221548 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.221564 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.221578 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.221587 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:23Z","lastTransitionTime":"2025-12-05T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:23 crc kubenswrapper[4997]: E1205 06:56:23.234036 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:23Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.237889 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.237921 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.237933 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.237950 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.237961 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:23Z","lastTransitionTime":"2025-12-05T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:23 crc kubenswrapper[4997]: E1205 06:56:23.283156 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:23Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.287286 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.287322 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.287331 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.287347 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.287357 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:23Z","lastTransitionTime":"2025-12-05T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:23 crc kubenswrapper[4997]: E1205 06:56:23.300740 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:23Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.303566 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.303598 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.303607 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.303646 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.303657 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:23Z","lastTransitionTime":"2025-12-05T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:23 crc kubenswrapper[4997]: E1205 06:56:23.315702 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:23Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.320739 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.320774 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.320783 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.320798 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.320810 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:23Z","lastTransitionTime":"2025-12-05T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:23 crc kubenswrapper[4997]: E1205 06:56:23.331156 4997 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T06:56:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e3043b93-5986-433d-b39e-eae93f14d589\\\",\\\"systemUUID\\\":\\\"35ecf2fa-8290-45a9-ad37-8d68d8e2ec69\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:23Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:23 crc kubenswrapper[4997]: E1205 06:56:23.331272 4997 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.332444 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.332461 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.332471 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.332482 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.332490 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:23Z","lastTransitionTime":"2025-12-05T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.434222 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.434276 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.434283 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.434297 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.434306 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:23Z","lastTransitionTime":"2025-12-05T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.536583 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.536689 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.536707 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.536731 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.536748 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:23Z","lastTransitionTime":"2025-12-05T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.638959 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.638989 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.638998 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.639010 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.639018 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:23Z","lastTransitionTime":"2025-12-05T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.743037 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.743096 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.743118 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.743146 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.743170 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:23Z","lastTransitionTime":"2025-12-05T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.748423 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.748428 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.748462 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.748514 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:23 crc kubenswrapper[4997]: E1205 06:56:23.748690 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:23 crc kubenswrapper[4997]: E1205 06:56:23.748858 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:23 crc kubenswrapper[4997]: E1205 06:56:23.748925 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:23 crc kubenswrapper[4997]: E1205 06:56:23.749022 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.845848 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.845888 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.845897 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.845913 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.845923 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:23Z","lastTransitionTime":"2025-12-05T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.948408 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.948454 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.948465 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.948485 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:23 crc kubenswrapper[4997]: I1205 06:56:23.948504 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:23Z","lastTransitionTime":"2025-12-05T06:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.051692 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.051724 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.051732 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.051745 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.051754 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:24Z","lastTransitionTime":"2025-12-05T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.153767 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.153815 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.153830 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.153847 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.153859 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:24Z","lastTransitionTime":"2025-12-05T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.255932 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.255972 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.255983 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.255998 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.256008 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:24Z","lastTransitionTime":"2025-12-05T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.358787 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.358831 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.358844 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.358859 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.358870 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:24Z","lastTransitionTime":"2025-12-05T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.461580 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.461652 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.461665 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.461682 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.461693 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:24Z","lastTransitionTime":"2025-12-05T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.564251 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.564291 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.564304 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.564319 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.564329 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:24Z","lastTransitionTime":"2025-12-05T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.666688 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.666753 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.666764 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.666780 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.666792 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:24Z","lastTransitionTime":"2025-12-05T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.771072 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.771135 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.771153 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.771178 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.771198 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:24Z","lastTransitionTime":"2025-12-05T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.874174 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.874221 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.874232 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.874252 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.874263 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:24Z","lastTransitionTime":"2025-12-05T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.976365 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.976397 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.976409 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.976424 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:24 crc kubenswrapper[4997]: I1205 06:56:24.976435 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:24Z","lastTransitionTime":"2025-12-05T06:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.078492 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.078538 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.078559 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.078576 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.078587 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:25Z","lastTransitionTime":"2025-12-05T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.181324 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.181366 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.181382 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.181403 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.181417 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:25Z","lastTransitionTime":"2025-12-05T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.283886 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.283927 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.283938 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.283956 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.283968 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:25Z","lastTransitionTime":"2025-12-05T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.386012 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.386047 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.386061 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.386076 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.386085 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:25Z","lastTransitionTime":"2025-12-05T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.488469 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.488534 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.488547 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.488565 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.488579 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:25Z","lastTransitionTime":"2025-12-05T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.590605 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.590692 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.590708 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.590725 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.590735 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:25Z","lastTransitionTime":"2025-12-05T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.693590 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.693666 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.693678 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.693695 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.693705 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:25Z","lastTransitionTime":"2025-12-05T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.748982 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.749016 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.749093 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.749328 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:25 crc kubenswrapper[4997]: E1205 06:56:25.749430 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:25 crc kubenswrapper[4997]: E1205 06:56:25.749494 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:25 crc kubenswrapper[4997]: E1205 06:56:25.749564 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:25 crc kubenswrapper[4997]: E1205 06:56:25.749653 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.795875 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.795919 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.795930 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.795946 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.795957 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:25Z","lastTransitionTime":"2025-12-05T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.898166 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.898211 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.898223 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.898240 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:25 crc kubenswrapper[4997]: I1205 06:56:25.898250 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:25Z","lastTransitionTime":"2025-12-05T06:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.000338 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.000377 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.000385 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.000400 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.000420 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:26Z","lastTransitionTime":"2025-12-05T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.102583 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.102656 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.102669 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.102684 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.102695 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:26Z","lastTransitionTime":"2025-12-05T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.205211 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.205254 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.205264 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.205282 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.205292 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:26Z","lastTransitionTime":"2025-12-05T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.308242 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.308298 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.308317 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.308340 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.308357 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:26Z","lastTransitionTime":"2025-12-05T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.410730 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.410789 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.410809 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.410835 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.410853 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:26Z","lastTransitionTime":"2025-12-05T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.513916 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.513954 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.513965 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.513982 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.513991 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:26Z","lastTransitionTime":"2025-12-05T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.617079 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.617129 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.617139 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.617156 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.617167 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:26Z","lastTransitionTime":"2025-12-05T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.719603 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.719689 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.719704 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.719731 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.719749 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:26Z","lastTransitionTime":"2025-12-05T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.822639 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.822693 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.822704 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.822748 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.822763 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:26Z","lastTransitionTime":"2025-12-05T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.925318 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.925392 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.925418 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.925456 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:26 crc kubenswrapper[4997]: I1205 06:56:26.925481 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:26Z","lastTransitionTime":"2025-12-05T06:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.028868 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.028952 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.028971 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.029009 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.029031 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:27Z","lastTransitionTime":"2025-12-05T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.131755 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.131795 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.131806 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.131822 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.131832 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:27Z","lastTransitionTime":"2025-12-05T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.233829 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.233894 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.233912 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.233935 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.233953 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:27Z","lastTransitionTime":"2025-12-05T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.336733 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.336788 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.336797 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.336811 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.336820 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:27Z","lastTransitionTime":"2025-12-05T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.439520 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.439688 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.439721 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.439795 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.439819 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:27Z","lastTransitionTime":"2025-12-05T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.542919 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.542974 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.542991 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.543019 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.543042 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:27Z","lastTransitionTime":"2025-12-05T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.645210 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.645294 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.645314 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.645339 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.645356 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:27Z","lastTransitionTime":"2025-12-05T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.747843 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.747921 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.747954 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.747982 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.748002 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:27Z","lastTransitionTime":"2025-12-05T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.748238 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.748301 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.748342 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.748407 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:27 crc kubenswrapper[4997]: E1205 06:56:27.748322 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:27 crc kubenswrapper[4997]: E1205 06:56:27.748491 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:27 crc kubenswrapper[4997]: E1205 06:56:27.748488 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:27 crc kubenswrapper[4997]: E1205 06:56:27.748533 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.763263 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.850580 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.850633 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.850644 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.850659 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.850671 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:27Z","lastTransitionTime":"2025-12-05T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.952419 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.952453 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.952464 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.952479 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:27 crc kubenswrapper[4997]: I1205 06:56:27.952491 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:27Z","lastTransitionTime":"2025-12-05T06:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.053984 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.054027 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.054038 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.054080 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.054093 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:28Z","lastTransitionTime":"2025-12-05T06:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.156114 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.156150 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.156162 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.156178 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.156189 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:28Z","lastTransitionTime":"2025-12-05T06:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.258433 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.258493 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.258506 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.258523 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.258556 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:28Z","lastTransitionTime":"2025-12-05T06:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.361290 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.361333 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.361342 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.361356 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.361364 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:28Z","lastTransitionTime":"2025-12-05T06:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.464014 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.464056 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.464064 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.464080 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.464091 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:28Z","lastTransitionTime":"2025-12-05T06:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.567435 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.567638 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.567668 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.567699 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.567721 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:28Z","lastTransitionTime":"2025-12-05T06:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.670626 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.670852 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.670860 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.670876 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.670887 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:28Z","lastTransitionTime":"2025-12-05T06:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.773454 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.773514 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.773527 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.773630 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.773649 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:28Z","lastTransitionTime":"2025-12-05T06:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.876873 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.876907 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.876915 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.876927 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.876942 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:28Z","lastTransitionTime":"2025-12-05T06:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.979013 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.979056 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.979069 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.979087 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:28 crc kubenswrapper[4997]: I1205 06:56:28.979099 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:28Z","lastTransitionTime":"2025-12-05T06:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.081011 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.081062 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.081073 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.081092 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.081108 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:29Z","lastTransitionTime":"2025-12-05T06:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.184490 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.184766 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.184776 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.184790 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.184799 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:29Z","lastTransitionTime":"2025-12-05T06:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.287029 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.287076 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.287091 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.287112 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.287129 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:29Z","lastTransitionTime":"2025-12-05T06:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.389232 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.389290 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.389299 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.389312 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.389321 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:29Z","lastTransitionTime":"2025-12-05T06:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.492761 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.492807 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.492817 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.492833 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.492844 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:29Z","lastTransitionTime":"2025-12-05T06:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.595825 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.595859 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.595869 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.595882 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.595892 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:29Z","lastTransitionTime":"2025-12-05T06:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.698799 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.698834 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.698875 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.698890 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.698899 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:29Z","lastTransitionTime":"2025-12-05T06:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.748396 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.748457 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.748599 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:29 crc kubenswrapper[4997]: E1205 06:56:29.748744 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.748773 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:29 crc kubenswrapper[4997]: E1205 06:56:29.748876 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:29 crc kubenswrapper[4997]: E1205 06:56:29.749010 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:29 crc kubenswrapper[4997]: E1205 06:56:29.749080 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.764874 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1d2f43a-f27f-4739-9f27-095ef2ff7ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d212927277df4e60fcd5c06b8dbbdcd07b154a261d645ee8408e4f75780f40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e66ce92ed41ad231f19d0e91f9796c1d1405f8f15b3612d0d5a6a26418b6066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T06:55:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2kb4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k85r5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.776036 4997 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f6790f-312b-42ce-9bd4-58bac09ce615\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T06:55:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bk5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T06:55:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b2vlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T06:56:29Z is after 2025-08-24T17:21:41Z" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.801646 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.802070 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.802226 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.802331 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.802415 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:29Z","lastTransitionTime":"2025-12-05T06:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.810932 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=72.810906091 podStartE2EDuration="1m12.810906091s" podCreationTimestamp="2025-12-05 06:55:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:56:29.810595233 +0000 UTC m=+90.339502534" watchObservedRunningTime="2025-12-05 06:56:29.810906091 +0000 UTC m=+90.339813362" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.847362 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-2jj4c" podStartSLOduration=70.847337211 podStartE2EDuration="1m10.847337211s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:56:29.824549553 +0000 UTC m=+90.353456814" watchObservedRunningTime="2025-12-05 06:56:29.847337211 +0000 UTC m=+90.376244472" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.861606 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-l5p2z" podStartSLOduration=71.861576289 podStartE2EDuration="1m11.861576289s" podCreationTimestamp="2025-12-05 06:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:56:29.861021605 +0000 UTC m=+90.389928936" watchObservedRunningTime="2025-12-05 06:56:29.861576289 +0000 UTC m=+90.390483590" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.885920 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podStartSLOduration=71.885899337 podStartE2EDuration="1m11.885899337s" podCreationTimestamp="2025-12-05 06:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:56:29.885265121 +0000 UTC m=+90.414172402" watchObservedRunningTime="2025-12-05 06:56:29.885899337 +0000 UTC m=+90.414806608" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.897896 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=2.897856246 podStartE2EDuration="2.897856246s" podCreationTimestamp="2025-12-05 06:56:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:56:29.897556688 +0000 UTC m=+90.426463999" watchObservedRunningTime="2025-12-05 06:56:29.897856246 +0000 UTC m=+90.426763517" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.904744 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.904787 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.904802 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.904820 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.904834 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:29Z","lastTransitionTime":"2025-12-05T06:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:29 crc kubenswrapper[4997]: I1205 06:56:29.992625 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-th9dq" podStartSLOduration=71.992587072 podStartE2EDuration="1m11.992587072s" podCreationTimestamp="2025-12-05 06:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:56:29.991921014 +0000 UTC m=+90.520828285" watchObservedRunningTime="2025-12-05 06:56:29.992587072 +0000 UTC m=+90.521494333" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.007358 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.007412 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.007422 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.007442 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.007455 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:30Z","lastTransitionTime":"2025-12-05T06:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.027139 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-5m9fs" podStartSLOduration=72.027113683 podStartE2EDuration="1m12.027113683s" podCreationTimestamp="2025-12-05 06:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:56:30.010214797 +0000 UTC m=+90.539122058" watchObservedRunningTime="2025-12-05 06:56:30.027113683 +0000 UTC m=+90.556020954" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.044905 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=73.044870882 podStartE2EDuration="1m13.044870882s" podCreationTimestamp="2025-12-05 06:55:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:56:30.044159763 +0000 UTC m=+90.573067034" watchObservedRunningTime="2025-12-05 06:56:30.044870882 +0000 UTC m=+90.573778143" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.057732 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=36.057701043 podStartE2EDuration="36.057701043s" podCreationTimestamp="2025-12-05 06:55:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:56:30.057139139 +0000 UTC m=+90.586046400" watchObservedRunningTime="2025-12-05 06:56:30.057701043 +0000 UTC m=+90.586608304" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.110007 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.110052 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.110061 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.110075 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.110085 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:30Z","lastTransitionTime":"2025-12-05T06:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.212113 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.212145 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.212158 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.212175 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.212189 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:30Z","lastTransitionTime":"2025-12-05T06:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.313806 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.313843 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.313853 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.313866 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.313876 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:30Z","lastTransitionTime":"2025-12-05T06:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.416738 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.416790 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.416802 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.416822 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.416837 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:30Z","lastTransitionTime":"2025-12-05T06:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.518942 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.518981 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.518989 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.519003 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.519014 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:30Z","lastTransitionTime":"2025-12-05T06:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.621554 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.621594 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.621603 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.621639 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.621652 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:30Z","lastTransitionTime":"2025-12-05T06:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.724293 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.724347 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.724356 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.724374 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.724383 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:30Z","lastTransitionTime":"2025-12-05T06:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.827106 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.827147 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.827155 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.827170 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.827181 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:30Z","lastTransitionTime":"2025-12-05T06:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.930498 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.930538 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.930549 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.930569 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:30 crc kubenswrapper[4997]: I1205 06:56:30.930582 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:30Z","lastTransitionTime":"2025-12-05T06:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.032589 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.032646 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.032655 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.032670 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.032681 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:31Z","lastTransitionTime":"2025-12-05T06:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.135352 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.135383 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.135392 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.135404 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.135412 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:31Z","lastTransitionTime":"2025-12-05T06:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.237758 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.237793 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.237801 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.237816 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.237825 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:31Z","lastTransitionTime":"2025-12-05T06:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.340965 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.341178 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.341278 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.341355 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.341426 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:31Z","lastTransitionTime":"2025-12-05T06:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.444235 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.444276 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.444285 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.444301 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.444310 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:31Z","lastTransitionTime":"2025-12-05T06:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.546606 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.546671 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.546682 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.546698 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.546708 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:31Z","lastTransitionTime":"2025-12-05T06:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.648731 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.649213 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.649302 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.649372 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.649427 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:31Z","lastTransitionTime":"2025-12-05T06:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.751059 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:31 crc kubenswrapper[4997]: E1205 06:56:31.751164 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.751335 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:31 crc kubenswrapper[4997]: E1205 06:56:31.751376 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.751754 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.751873 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:31 crc kubenswrapper[4997]: E1205 06:56:31.752001 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:31 crc kubenswrapper[4997]: E1205 06:56:31.752141 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.752877 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.752899 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.752909 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.752926 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.752937 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:31Z","lastTransitionTime":"2025-12-05T06:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.856046 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.856086 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.856098 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.856118 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.856129 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:31Z","lastTransitionTime":"2025-12-05T06:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.958867 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.958910 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.958920 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.958937 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:31 crc kubenswrapper[4997]: I1205 06:56:31.958950 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:31Z","lastTransitionTime":"2025-12-05T06:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.061534 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.061806 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.061897 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.061985 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.062059 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:32Z","lastTransitionTime":"2025-12-05T06:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.164564 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.164597 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.164608 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.164640 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.164650 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:32Z","lastTransitionTime":"2025-12-05T06:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.268542 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.268582 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.268594 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.268628 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.268641 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:32Z","lastTransitionTime":"2025-12-05T06:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.370500 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.370534 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.370542 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.370556 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.370565 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:32Z","lastTransitionTime":"2025-12-05T06:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.473327 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.473631 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.473710 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.473800 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.473895 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:32Z","lastTransitionTime":"2025-12-05T06:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.576493 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.576760 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.576833 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.576904 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.576967 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:32Z","lastTransitionTime":"2025-12-05T06:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.679078 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.679678 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.679781 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.679847 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.679914 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:32Z","lastTransitionTime":"2025-12-05T06:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.782124 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.782170 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.782179 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.782193 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.782202 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:32Z","lastTransitionTime":"2025-12-05T06:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.884967 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.885005 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.885014 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.885028 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.885040 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:32Z","lastTransitionTime":"2025-12-05T06:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.987666 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.987720 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.987730 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.987748 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:32 crc kubenswrapper[4997]: I1205 06:56:32.987759 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:32Z","lastTransitionTime":"2025-12-05T06:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.091124 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.091173 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.091184 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.091203 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.091217 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:33Z","lastTransitionTime":"2025-12-05T06:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.194122 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.194156 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.194166 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.194183 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.194194 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:33Z","lastTransitionTime":"2025-12-05T06:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.296716 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.296985 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.297129 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.297238 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.297331 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:33Z","lastTransitionTime":"2025-12-05T06:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.399358 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.399403 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.399415 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.399432 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.399443 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:33Z","lastTransitionTime":"2025-12-05T06:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.501512 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.501545 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.501555 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.501569 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.501579 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:33Z","lastTransitionTime":"2025-12-05T06:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.603588 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.603670 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.603681 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.603697 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.603707 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:33Z","lastTransitionTime":"2025-12-05T06:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.614266 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.614394 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.614499 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.614638 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.614742 4997 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T06:56:33Z","lastTransitionTime":"2025-12-05T06:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.653654 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws"] Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.654460 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.656685 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.656713 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.656993 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.657296 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.668364 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k85r5" podStartSLOduration=74.668320484 podStartE2EDuration="1m14.668320484s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:56:33.666938689 +0000 UTC m=+94.195845950" watchObservedRunningTime="2025-12-05 06:56:33.668320484 +0000 UTC m=+94.197227765" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.712084 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8465e806-dc7a-4b59-8889-390425f2a3e3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6zmws\" (UID: \"8465e806-dc7a-4b59-8889-390425f2a3e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.712126 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8465e806-dc7a-4b59-8889-390425f2a3e3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6zmws\" (UID: \"8465e806-dc7a-4b59-8889-390425f2a3e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.712155 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8465e806-dc7a-4b59-8889-390425f2a3e3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6zmws\" (UID: \"8465e806-dc7a-4b59-8889-390425f2a3e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.712189 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8465e806-dc7a-4b59-8889-390425f2a3e3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6zmws\" (UID: \"8465e806-dc7a-4b59-8889-390425f2a3e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.712205 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8465e806-dc7a-4b59-8889-390425f2a3e3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6zmws\" (UID: \"8465e806-dc7a-4b59-8889-390425f2a3e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.748157 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:33 crc kubenswrapper[4997]: E1205 06:56:33.748272 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.748334 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.748544 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.748678 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:33 crc kubenswrapper[4997]: E1205 06:56:33.748766 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:33 crc kubenswrapper[4997]: E1205 06:56:33.748849 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:33 crc kubenswrapper[4997]: E1205 06:56:33.748888 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.748939 4997 scope.go:117] "RemoveContainer" containerID="efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57" Dec 05 06:56:33 crc kubenswrapper[4997]: E1205 06:56:33.749090 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.812928 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8465e806-dc7a-4b59-8889-390425f2a3e3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6zmws\" (UID: \"8465e806-dc7a-4b59-8889-390425f2a3e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.812967 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8465e806-dc7a-4b59-8889-390425f2a3e3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6zmws\" (UID: \"8465e806-dc7a-4b59-8889-390425f2a3e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.813015 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8465e806-dc7a-4b59-8889-390425f2a3e3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6zmws\" (UID: \"8465e806-dc7a-4b59-8889-390425f2a3e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.813032 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8465e806-dc7a-4b59-8889-390425f2a3e3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6zmws\" (UID: \"8465e806-dc7a-4b59-8889-390425f2a3e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.813060 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8465e806-dc7a-4b59-8889-390425f2a3e3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6zmws\" (UID: \"8465e806-dc7a-4b59-8889-390425f2a3e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.813212 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8465e806-dc7a-4b59-8889-390425f2a3e3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6zmws\" (UID: \"8465e806-dc7a-4b59-8889-390425f2a3e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.813325 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8465e806-dc7a-4b59-8889-390425f2a3e3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6zmws\" (UID: \"8465e806-dc7a-4b59-8889-390425f2a3e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.813945 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8465e806-dc7a-4b59-8889-390425f2a3e3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6zmws\" (UID: \"8465e806-dc7a-4b59-8889-390425f2a3e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.825527 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8465e806-dc7a-4b59-8889-390425f2a3e3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6zmws\" (UID: \"8465e806-dc7a-4b59-8889-390425f2a3e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.828401 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8465e806-dc7a-4b59-8889-390425f2a3e3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6zmws\" (UID: \"8465e806-dc7a-4b59-8889-390425f2a3e3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" Dec 05 06:56:33 crc kubenswrapper[4997]: I1205 06:56:33.967347 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" Dec 05 06:56:34 crc kubenswrapper[4997]: I1205 06:56:34.296585 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" event={"ID":"8465e806-dc7a-4b59-8889-390425f2a3e3","Type":"ContainerStarted","Data":"1406e6b85e4127a6c174b125b3b759962cc89f1423b6b3d22e1d03fca5e8e377"} Dec 05 06:56:34 crc kubenswrapper[4997]: I1205 06:56:34.296710 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" event={"ID":"8465e806-dc7a-4b59-8889-390425f2a3e3","Type":"ContainerStarted","Data":"7ca8facbd8ef313f886a18bddf78dde657871c0d23e9820be4a560fa8ddf8516"} Dec 05 06:56:34 crc kubenswrapper[4997]: I1205 06:56:34.314586 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6zmws" podStartSLOduration=76.314560879 podStartE2EDuration="1m16.314560879s" podCreationTimestamp="2025-12-05 06:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:56:34.312989279 +0000 UTC m=+94.841896560" watchObservedRunningTime="2025-12-05 06:56:34.314560879 +0000 UTC m=+94.843468170" Dec 05 06:56:35 crc kubenswrapper[4997]: I1205 06:56:35.748136 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:35 crc kubenswrapper[4997]: I1205 06:56:35.748192 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:35 crc kubenswrapper[4997]: E1205 06:56:35.748576 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:35 crc kubenswrapper[4997]: I1205 06:56:35.748309 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:35 crc kubenswrapper[4997]: E1205 06:56:35.748732 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:35 crc kubenswrapper[4997]: I1205 06:56:35.748256 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:35 crc kubenswrapper[4997]: E1205 06:56:35.748815 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:35 crc kubenswrapper[4997]: E1205 06:56:35.749021 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:36 crc kubenswrapper[4997]: I1205 06:56:36.238382 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs\") pod \"network-metrics-daemon-b2vlg\" (UID: \"81f6790f-312b-42ce-9bd4-58bac09ce615\") " pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:36 crc kubenswrapper[4997]: E1205 06:56:36.238528 4997 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 06:56:36 crc kubenswrapper[4997]: E1205 06:56:36.238578 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs podName:81f6790f-312b-42ce-9bd4-58bac09ce615 nodeName:}" failed. No retries permitted until 2025-12-05 06:57:40.23856386 +0000 UTC m=+160.767471121 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs") pod "network-metrics-daemon-b2vlg" (UID: "81f6790f-312b-42ce-9bd4-58bac09ce615") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 06:56:37 crc kubenswrapper[4997]: I1205 06:56:37.748325 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:37 crc kubenswrapper[4997]: I1205 06:56:37.748472 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:37 crc kubenswrapper[4997]: I1205 06:56:37.749504 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:37 crc kubenswrapper[4997]: I1205 06:56:37.749875 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:37 crc kubenswrapper[4997]: E1205 06:56:37.749951 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:37 crc kubenswrapper[4997]: E1205 06:56:37.750216 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:37 crc kubenswrapper[4997]: E1205 06:56:37.750326 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:37 crc kubenswrapper[4997]: E1205 06:56:37.750425 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:39 crc kubenswrapper[4997]: I1205 06:56:39.748895 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:39 crc kubenswrapper[4997]: I1205 06:56:39.748956 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:39 crc kubenswrapper[4997]: I1205 06:56:39.748931 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:39 crc kubenswrapper[4997]: I1205 06:56:39.749881 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:39 crc kubenswrapper[4997]: E1205 06:56:39.750054 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:39 crc kubenswrapper[4997]: E1205 06:56:39.750304 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:39 crc kubenswrapper[4997]: E1205 06:56:39.750355 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:39 crc kubenswrapper[4997]: E1205 06:56:39.750435 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:41 crc kubenswrapper[4997]: I1205 06:56:41.748855 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:41 crc kubenswrapper[4997]: I1205 06:56:41.749039 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:41 crc kubenswrapper[4997]: I1205 06:56:41.749743 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:41 crc kubenswrapper[4997]: E1205 06:56:41.749839 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:41 crc kubenswrapper[4997]: I1205 06:56:41.749902 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:41 crc kubenswrapper[4997]: E1205 06:56:41.750017 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:41 crc kubenswrapper[4997]: E1205 06:56:41.750132 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:41 crc kubenswrapper[4997]: E1205 06:56:41.750228 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:43 crc kubenswrapper[4997]: I1205 06:56:43.748950 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:43 crc kubenswrapper[4997]: I1205 06:56:43.749084 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:43 crc kubenswrapper[4997]: E1205 06:56:43.749130 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:43 crc kubenswrapper[4997]: I1205 06:56:43.749274 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:43 crc kubenswrapper[4997]: E1205 06:56:43.749546 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:43 crc kubenswrapper[4997]: I1205 06:56:43.750150 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:43 crc kubenswrapper[4997]: E1205 06:56:43.750353 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:43 crc kubenswrapper[4997]: E1205 06:56:43.750955 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:44 crc kubenswrapper[4997]: I1205 06:56:44.750032 4997 scope.go:117] "RemoveContainer" containerID="efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57" Dec 05 06:56:44 crc kubenswrapper[4997]: E1205 06:56:44.750429 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" Dec 05 06:56:44 crc kubenswrapper[4997]: I1205 06:56:44.770583 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 05 06:56:45 crc kubenswrapper[4997]: I1205 06:56:45.748317 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:45 crc kubenswrapper[4997]: I1205 06:56:45.748424 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:45 crc kubenswrapper[4997]: I1205 06:56:45.748426 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:45 crc kubenswrapper[4997]: I1205 06:56:45.748570 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:45 crc kubenswrapper[4997]: E1205 06:56:45.748576 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:45 crc kubenswrapper[4997]: E1205 06:56:45.748711 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:45 crc kubenswrapper[4997]: E1205 06:56:45.748754 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:45 crc kubenswrapper[4997]: E1205 06:56:45.748826 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:47 crc kubenswrapper[4997]: I1205 06:56:47.748694 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:47 crc kubenswrapper[4997]: I1205 06:56:47.748698 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:47 crc kubenswrapper[4997]: I1205 06:56:47.748738 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:47 crc kubenswrapper[4997]: I1205 06:56:47.749845 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:47 crc kubenswrapper[4997]: E1205 06:56:47.750000 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:47 crc kubenswrapper[4997]: E1205 06:56:47.750406 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:47 crc kubenswrapper[4997]: E1205 06:56:47.750533 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:47 crc kubenswrapper[4997]: E1205 06:56:47.750630 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:49 crc kubenswrapper[4997]: I1205 06:56:49.748889 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:49 crc kubenswrapper[4997]: I1205 06:56:49.748889 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:49 crc kubenswrapper[4997]: I1205 06:56:49.749803 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:49 crc kubenswrapper[4997]: I1205 06:56:49.749878 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:49 crc kubenswrapper[4997]: E1205 06:56:49.750045 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:49 crc kubenswrapper[4997]: E1205 06:56:49.750129 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:49 crc kubenswrapper[4997]: E1205 06:56:49.750234 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:49 crc kubenswrapper[4997]: E1205 06:56:49.750863 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:49 crc kubenswrapper[4997]: I1205 06:56:49.780359 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=5.780341627 podStartE2EDuration="5.780341627s" podCreationTimestamp="2025-12-05 06:56:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:56:49.777835442 +0000 UTC m=+110.306742713" watchObservedRunningTime="2025-12-05 06:56:49.780341627 +0000 UTC m=+110.309248888" Dec 05 06:56:51 crc kubenswrapper[4997]: I1205 06:56:51.350747 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-th9dq_85f05eb3-df7f-4c46-a0e2-451766bb3297/kube-multus/1.log" Dec 05 06:56:51 crc kubenswrapper[4997]: I1205 06:56:51.351504 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-th9dq_85f05eb3-df7f-4c46-a0e2-451766bb3297/kube-multus/0.log" Dec 05 06:56:51 crc kubenswrapper[4997]: I1205 06:56:51.351563 4997 generic.go:334] "Generic (PLEG): container finished" podID="85f05eb3-df7f-4c46-a0e2-451766bb3297" containerID="ba8817aab2ae020cc998813edb099030cc3c3c6eb11e8f175e3cfd82b18686b2" exitCode=1 Dec 05 06:56:51 crc kubenswrapper[4997]: I1205 06:56:51.351604 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-th9dq" event={"ID":"85f05eb3-df7f-4c46-a0e2-451766bb3297","Type":"ContainerDied","Data":"ba8817aab2ae020cc998813edb099030cc3c3c6eb11e8f175e3cfd82b18686b2"} Dec 05 06:56:51 crc kubenswrapper[4997]: I1205 06:56:51.351687 4997 scope.go:117] "RemoveContainer" containerID="7f1f4de1568febce74289a08bd758381dad0025a8c79fbbd00256cd4381744d6" Dec 05 06:56:51 crc kubenswrapper[4997]: I1205 06:56:51.352230 4997 scope.go:117] "RemoveContainer" containerID="ba8817aab2ae020cc998813edb099030cc3c3c6eb11e8f175e3cfd82b18686b2" Dec 05 06:56:51 crc kubenswrapper[4997]: E1205 06:56:51.352452 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-th9dq_openshift-multus(85f05eb3-df7f-4c46-a0e2-451766bb3297)\"" pod="openshift-multus/multus-th9dq" podUID="85f05eb3-df7f-4c46-a0e2-451766bb3297" Dec 05 06:56:51 crc kubenswrapper[4997]: I1205 06:56:51.748848 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:51 crc kubenswrapper[4997]: I1205 06:56:51.748905 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:51 crc kubenswrapper[4997]: I1205 06:56:51.748866 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:51 crc kubenswrapper[4997]: E1205 06:56:51.748980 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:51 crc kubenswrapper[4997]: E1205 06:56:51.749205 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:51 crc kubenswrapper[4997]: E1205 06:56:51.749471 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:51 crc kubenswrapper[4997]: I1205 06:56:51.749689 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:51 crc kubenswrapper[4997]: E1205 06:56:51.749777 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:52 crc kubenswrapper[4997]: I1205 06:56:52.356260 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-th9dq_85f05eb3-df7f-4c46-a0e2-451766bb3297/kube-multus/1.log" Dec 05 06:56:53 crc kubenswrapper[4997]: I1205 06:56:53.748802 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:53 crc kubenswrapper[4997]: I1205 06:56:53.748870 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:53 crc kubenswrapper[4997]: I1205 06:56:53.748803 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:53 crc kubenswrapper[4997]: E1205 06:56:53.748929 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:53 crc kubenswrapper[4997]: E1205 06:56:53.749010 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:53 crc kubenswrapper[4997]: E1205 06:56:53.749096 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:53 crc kubenswrapper[4997]: I1205 06:56:53.749915 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:53 crc kubenswrapper[4997]: E1205 06:56:53.750080 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:55 crc kubenswrapper[4997]: I1205 06:56:55.748989 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:55 crc kubenswrapper[4997]: I1205 06:56:55.749034 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:55 crc kubenswrapper[4997]: E1205 06:56:55.749133 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:55 crc kubenswrapper[4997]: I1205 06:56:55.749006 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:55 crc kubenswrapper[4997]: I1205 06:56:55.749260 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:55 crc kubenswrapper[4997]: E1205 06:56:55.749361 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:55 crc kubenswrapper[4997]: E1205 06:56:55.749523 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:55 crc kubenswrapper[4997]: E1205 06:56:55.749672 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:56 crc kubenswrapper[4997]: I1205 06:56:56.750399 4997 scope.go:117] "RemoveContainer" containerID="efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57" Dec 05 06:56:56 crc kubenswrapper[4997]: E1205 06:56:56.750947 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bchpk_openshift-ovn-kubernetes(28f65c70-f59b-4393-b9c9-8163bb596f6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" Dec 05 06:56:57 crc kubenswrapper[4997]: I1205 06:56:57.748184 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:57 crc kubenswrapper[4997]: I1205 06:56:57.748237 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:57 crc kubenswrapper[4997]: E1205 06:56:57.748289 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:57 crc kubenswrapper[4997]: I1205 06:56:57.748314 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:57 crc kubenswrapper[4997]: E1205 06:56:57.748441 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:57 crc kubenswrapper[4997]: I1205 06:56:57.748462 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:57 crc kubenswrapper[4997]: E1205 06:56:57.748509 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:57 crc kubenswrapper[4997]: E1205 06:56:57.748571 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:59 crc kubenswrapper[4997]: E1205 06:56:59.694107 4997 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 05 06:56:59 crc kubenswrapper[4997]: I1205 06:56:59.748427 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:56:59 crc kubenswrapper[4997]: I1205 06:56:59.748430 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:56:59 crc kubenswrapper[4997]: I1205 06:56:59.748526 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:56:59 crc kubenswrapper[4997]: E1205 06:56:59.749767 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:56:59 crc kubenswrapper[4997]: I1205 06:56:59.749822 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:56:59 crc kubenswrapper[4997]: E1205 06:56:59.750005 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:56:59 crc kubenswrapper[4997]: E1205 06:56:59.750100 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:56:59 crc kubenswrapper[4997]: E1205 06:56:59.750136 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:56:59 crc kubenswrapper[4997]: E1205 06:56:59.835230 4997 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 06:57:01 crc kubenswrapper[4997]: I1205 06:57:01.748835 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:57:01 crc kubenswrapper[4997]: I1205 06:57:01.748885 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:57:01 crc kubenswrapper[4997]: I1205 06:57:01.748885 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:57:01 crc kubenswrapper[4997]: I1205 06:57:01.749019 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:57:01 crc kubenswrapper[4997]: E1205 06:57:01.749001 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:57:01 crc kubenswrapper[4997]: E1205 06:57:01.749105 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:57:01 crc kubenswrapper[4997]: E1205 06:57:01.749168 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:57:01 crc kubenswrapper[4997]: E1205 06:57:01.749210 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:57:03 crc kubenswrapper[4997]: I1205 06:57:03.748680 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:57:03 crc kubenswrapper[4997]: I1205 06:57:03.748972 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:57:03 crc kubenswrapper[4997]: I1205 06:57:03.748786 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:57:03 crc kubenswrapper[4997]: I1205 06:57:03.748782 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:57:03 crc kubenswrapper[4997]: E1205 06:57:03.749071 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:57:03 crc kubenswrapper[4997]: E1205 06:57:03.749196 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:57:03 crc kubenswrapper[4997]: E1205 06:57:03.749265 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:57:03 crc kubenswrapper[4997]: E1205 06:57:03.749366 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:57:04 crc kubenswrapper[4997]: E1205 06:57:04.836952 4997 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 06:57:05 crc kubenswrapper[4997]: I1205 06:57:05.748411 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:57:05 crc kubenswrapper[4997]: I1205 06:57:05.748434 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:57:05 crc kubenswrapper[4997]: I1205 06:57:05.748438 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:57:05 crc kubenswrapper[4997]: I1205 06:57:05.748535 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:57:05 crc kubenswrapper[4997]: E1205 06:57:05.748537 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:57:05 crc kubenswrapper[4997]: E1205 06:57:05.748629 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:57:05 crc kubenswrapper[4997]: E1205 06:57:05.748892 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:57:05 crc kubenswrapper[4997]: E1205 06:57:05.748864 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:57:06 crc kubenswrapper[4997]: I1205 06:57:06.748980 4997 scope.go:117] "RemoveContainer" containerID="ba8817aab2ae020cc998813edb099030cc3c3c6eb11e8f175e3cfd82b18686b2" Dec 05 06:57:07 crc kubenswrapper[4997]: I1205 06:57:07.410040 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-th9dq_85f05eb3-df7f-4c46-a0e2-451766bb3297/kube-multus/1.log" Dec 05 06:57:07 crc kubenswrapper[4997]: I1205 06:57:07.410097 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-th9dq" event={"ID":"85f05eb3-df7f-4c46-a0e2-451766bb3297","Type":"ContainerStarted","Data":"41601dcf153b1e00253366b960707c5217ee9e19312ec6eb9f65a38e45cbcdd7"} Dec 05 06:57:07 crc kubenswrapper[4997]: I1205 06:57:07.748296 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:57:07 crc kubenswrapper[4997]: I1205 06:57:07.748328 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:57:07 crc kubenswrapper[4997]: I1205 06:57:07.748307 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:57:07 crc kubenswrapper[4997]: E1205 06:57:07.748419 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:57:07 crc kubenswrapper[4997]: I1205 06:57:07.748453 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:57:07 crc kubenswrapper[4997]: E1205 06:57:07.748523 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:57:07 crc kubenswrapper[4997]: E1205 06:57:07.748558 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:57:07 crc kubenswrapper[4997]: E1205 06:57:07.748668 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:57:08 crc kubenswrapper[4997]: I1205 06:57:08.749262 4997 scope.go:117] "RemoveContainer" containerID="efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57" Dec 05 06:57:09 crc kubenswrapper[4997]: I1205 06:57:09.417024 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovnkube-controller/3.log" Dec 05 06:57:09 crc kubenswrapper[4997]: I1205 06:57:09.420044 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerStarted","Data":"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa"} Dec 05 06:57:09 crc kubenswrapper[4997]: I1205 06:57:09.420770 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:57:09 crc kubenswrapper[4997]: I1205 06:57:09.445425 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podStartSLOduration=111.445404398 podStartE2EDuration="1m51.445404398s" podCreationTimestamp="2025-12-05 06:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:09.444632399 +0000 UTC m=+129.973539670" watchObservedRunningTime="2025-12-05 06:57:09.445404398 +0000 UTC m=+129.974311659" Dec 05 06:57:09 crc kubenswrapper[4997]: I1205 06:57:09.629426 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-b2vlg"] Dec 05 06:57:09 crc kubenswrapper[4997]: I1205 06:57:09.629552 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:57:09 crc kubenswrapper[4997]: E1205 06:57:09.629673 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:57:09 crc kubenswrapper[4997]: I1205 06:57:09.748451 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:57:09 crc kubenswrapper[4997]: I1205 06:57:09.748470 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:57:09 crc kubenswrapper[4997]: E1205 06:57:09.749314 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:57:09 crc kubenswrapper[4997]: I1205 06:57:09.749419 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:57:09 crc kubenswrapper[4997]: E1205 06:57:09.749481 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:57:09 crc kubenswrapper[4997]: E1205 06:57:09.749532 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:57:09 crc kubenswrapper[4997]: E1205 06:57:09.859324 4997 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 06:57:11 crc kubenswrapper[4997]: I1205 06:57:11.748164 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:57:11 crc kubenswrapper[4997]: I1205 06:57:11.748164 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:57:11 crc kubenswrapper[4997]: I1205 06:57:11.748235 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:57:11 crc kubenswrapper[4997]: I1205 06:57:11.748346 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:57:11 crc kubenswrapper[4997]: E1205 06:57:11.748597 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:57:11 crc kubenswrapper[4997]: E1205 06:57:11.748684 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:57:11 crc kubenswrapper[4997]: E1205 06:57:11.748751 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:57:11 crc kubenswrapper[4997]: E1205 06:57:11.748826 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:57:13 crc kubenswrapper[4997]: I1205 06:57:13.749405 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:57:13 crc kubenswrapper[4997]: I1205 06:57:13.749561 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:57:13 crc kubenswrapper[4997]: I1205 06:57:13.749423 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:57:13 crc kubenswrapper[4997]: E1205 06:57:13.749740 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 06:57:13 crc kubenswrapper[4997]: E1205 06:57:13.750051 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b2vlg" podUID="81f6790f-312b-42ce-9bd4-58bac09ce615" Dec 05 06:57:13 crc kubenswrapper[4997]: I1205 06:57:13.750132 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:57:13 crc kubenswrapper[4997]: E1205 06:57:13.750231 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 06:57:13 crc kubenswrapper[4997]: E1205 06:57:13.750336 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 06:57:15 crc kubenswrapper[4997]: I1205 06:57:15.748800 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:57:15 crc kubenswrapper[4997]: I1205 06:57:15.749298 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:57:15 crc kubenswrapper[4997]: I1205 06:57:15.749403 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:57:15 crc kubenswrapper[4997]: I1205 06:57:15.752806 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:57:15 crc kubenswrapper[4997]: I1205 06:57:15.755026 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 05 06:57:15 crc kubenswrapper[4997]: I1205 06:57:15.760266 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 05 06:57:15 crc kubenswrapper[4997]: I1205 06:57:15.761644 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 05 06:57:15 crc kubenswrapper[4997]: I1205 06:57:15.761644 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 05 06:57:15 crc kubenswrapper[4997]: I1205 06:57:15.762405 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 05 06:57:15 crc kubenswrapper[4997]: I1205 06:57:15.762428 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 05 06:57:18 crc kubenswrapper[4997]: I1205 06:57:18.757247 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 06:57:19 crc kubenswrapper[4997]: I1205 06:57:19.770155 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 06:57:19 crc kubenswrapper[4997]: I1205 06:57:19.770549 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.257195 4997 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.313912 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-57ksd"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.314349 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.314733 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wpl2k"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.315362 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.315366 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5b8mm"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.316824 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.317306 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-649kx"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.317451 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.317551 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.318284 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-jktft"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.318659 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9k5xc"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.319092 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9k5xc" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.319101 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.319717 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.320087 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jktft" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.320567 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.322906 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-lskhh"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.323036 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.323568 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.324199 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qz7ht"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.324855 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-67z92"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.325217 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zxx6b"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.325802 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-zxx6b" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.326257 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qz7ht" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.326548 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-67z92" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.328191 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.331578 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.331652 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.333698 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.333803 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.334893 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-57ksd"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.336699 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wxs8l"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.337338 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.337743 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qd6rv"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.338427 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qd6rv" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.343205 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.343424 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.343667 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.343667 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.357400 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.358430 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.358480 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.358765 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.359082 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.359395 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.360299 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.362745 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.363034 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.363359 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.364116 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.364324 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.364430 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.371821 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.372271 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.372654 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.372892 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.373087 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.373311 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.373799 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.373995 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.372278 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.374566 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.375214 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.375666 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.375878 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.376075 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.376330 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.376389 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.391574 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zm2tt"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.391814 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.392135 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.392176 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.392536 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.393954 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.394014 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.394053 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.394118 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.394187 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.394194 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.394318 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.394365 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.393961 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.394120 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.394522 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.394580 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.394732 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.394872 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.394948 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.395109 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.395138 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.394525 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.395336 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.395434 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.395568 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.395673 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.395801 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.395938 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.396027 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qj4lx"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.396749 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.396866 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.397019 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.397123 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.397315 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qj4lx" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.399381 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.399544 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.399632 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5xbd9"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.400057 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.400201 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.400301 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.400328 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.400452 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.400548 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.401036 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.402100 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.402245 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.402332 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.402469 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.402580 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.402711 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.403019 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.403018 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.404750 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.405450 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.410331 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.423152 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.423788 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wpl2k"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.423986 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.426327 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.428491 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9k5xc"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.428893 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.437891 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.439968 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.440240 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.445655 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.449947 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.450135 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.451131 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.453416 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.457601 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.460289 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.462581 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.481567 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-v4mxn"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.487922 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5msxl"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.488272 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-lskhh"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.488396 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.488937 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.492849 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-service-ca\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.492902 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7face06-9e66-4375-8b67-974b8936289f-serving-cert\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.492935 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a412ad65-d072-433a-8d6f-7e5390195f89-audit-dir\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.492958 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxvjd\" (UniqueName: \"kubernetes.io/projected/a412ad65-d072-433a-8d6f-7e5390195f89-kube-api-access-vxvjd\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.492983 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493003 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkmkk\" (UniqueName: \"kubernetes.io/projected/f4c37dd5-377e-4ed4-99f5-b179110153af-kube-api-access-gkmkk\") pod \"console-operator-58897d9998-qz7ht\" (UID: \"f4c37dd5-377e-4ed4-99f5-b179110153af\") " pod="openshift-console-operator/console-operator-58897d9998-qz7ht" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493024 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r72j7\" (UniqueName: \"kubernetes.io/projected/8c6d7e0f-f064-48a0-b05b-287b09867b45-kube-api-access-r72j7\") pod \"machine-approver-56656f9798-649kx\" (UID: \"8c6d7e0f-f064-48a0-b05b-287b09867b45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493043 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c7face06-9e66-4375-8b67-974b8936289f-node-pullsecrets\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493062 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-oauth-serving-cert\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493081 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/821da4d3-7e2a-480e-8290-a7080b916b08-audit-dir\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493101 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9370d84-0bfb-4e61-a5b7-bcb43e903593-serving-cert\") pod \"authentication-operator-69f744f599-zm2tt\" (UID: \"e9370d84-0bfb-4e61-a5b7-bcb43e903593\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493121 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzhn6\" (UniqueName: \"kubernetes.io/projected/e9370d84-0bfb-4e61-a5b7-bcb43e903593-kube-api-access-wzhn6\") pod \"authentication-operator-69f744f599-zm2tt\" (UID: \"e9370d84-0bfb-4e61-a5b7-bcb43e903593\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493142 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7face06-9e66-4375-8b67-974b8936289f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493163 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vs7j\" (UniqueName: \"kubernetes.io/projected/0160154e-35ae-471c-b848-214b9f33ec93-kube-api-access-6vs7j\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493184 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/821da4d3-7e2a-480e-8290-a7080b916b08-etcd-client\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493216 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3993123-571c-4be4-a525-d8e0481b1aa5-config\") pod \"machine-api-operator-5694c8668f-wpl2k\" (UID: \"f3993123-571c-4be4-a525-d8e0481b1aa5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493236 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f3993123-571c-4be4-a525-d8e0481b1aa5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wpl2k\" (UID: \"f3993123-571c-4be4-a525-d8e0481b1aa5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493267 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/821da4d3-7e2a-480e-8290-a7080b916b08-serving-cert\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493286 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493305 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4c37dd5-377e-4ed4-99f5-b179110153af-serving-cert\") pod \"console-operator-58897d9998-qz7ht\" (UID: \"f4c37dd5-377e-4ed4-99f5-b179110153af\") " pod="openshift-console-operator/console-operator-58897d9998-qz7ht" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493324 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-57ksd\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493344 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-trusted-ca-bundle\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493367 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7face06-9e66-4375-8b67-974b8936289f-audit-dir\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493390 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c78711c-d17c-4b1c-b00e-f8ac8972f1d9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qd6rv\" (UID: \"4c78711c-d17c-4b1c-b00e-f8ac8972f1d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qd6rv" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493413 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/821da4d3-7e2a-480e-8290-a7080b916b08-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493436 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493457 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4c37dd5-377e-4ed4-99f5-b179110153af-config\") pod \"console-operator-58897d9998-qz7ht\" (UID: \"f4c37dd5-377e-4ed4-99f5-b179110153af\") " pod="openshift-console-operator/console-operator-58897d9998-qz7ht" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493478 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c78711c-d17c-4b1c-b00e-f8ac8972f1d9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qd6rv\" (UID: \"4c78711c-d17c-4b1c-b00e-f8ac8972f1d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qd6rv" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493511 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/821da4d3-7e2a-480e-8290-a7080b916b08-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493533 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493554 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdc72754-1951-4312-9dd0-3458357f2caa-serving-cert\") pod \"openshift-config-operator-7777fb866f-jktft\" (UID: \"bdc72754-1951-4312-9dd0-3458357f2caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jktft" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493576 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzjz6\" (UniqueName: \"kubernetes.io/projected/0a38ab13-190d-40c5-9a69-389e2fae6884-kube-api-access-rzjz6\") pod \"route-controller-manager-6576b87f9c-db5rh\" (UID: \"0a38ab13-190d-40c5-9a69-389e2fae6884\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493604 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0160154e-35ae-471c-b848-214b9f33ec93-console-oauth-config\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493668 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x7lm\" (UniqueName: \"kubernetes.io/projected/800c31da-5f11-4029-a43f-ad96ac33b062-kube-api-access-5x7lm\") pod \"cluster-samples-operator-665b6dd947-9k5xc\" (UID: \"800c31da-5f11-4029-a43f-ad96ac33b062\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9k5xc" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493689 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq6hc\" (UniqueName: \"kubernetes.io/projected/f3993123-571c-4be4-a525-d8e0481b1aa5-kube-api-access-wq6hc\") pod \"machine-api-operator-5694c8668f-wpl2k\" (UID: \"f3993123-571c-4be4-a525-d8e0481b1aa5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493714 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p5st\" (UniqueName: \"kubernetes.io/projected/18378604-4db4-4da8-94c4-c52452ee6805-kube-api-access-7p5st\") pod \"dns-operator-744455d44c-zxx6b\" (UID: \"18378604-4db4-4da8-94c4-c52452ee6805\") " pod="openshift-dns-operator/dns-operator-744455d44c-zxx6b" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493738 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49f5s\" (UniqueName: \"kubernetes.io/projected/bdc72754-1951-4312-9dd0-3458357f2caa-kube-api-access-49f5s\") pod \"openshift-config-operator-7777fb866f-jktft\" (UID: \"bdc72754-1951-4312-9dd0-3458357f2caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jktft" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493760 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a38ab13-190d-40c5-9a69-389e2fae6884-config\") pod \"route-controller-manager-6576b87f9c-db5rh\" (UID: \"0a38ab13-190d-40c5-9a69-389e2fae6884\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493792 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493811 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8c6d7e0f-f064-48a0-b05b-287b09867b45-auth-proxy-config\") pod \"machine-approver-56656f9798-649kx\" (UID: \"8c6d7e0f-f064-48a0-b05b-287b09867b45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493836 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f3993123-571c-4be4-a525-d8e0481b1aa5-images\") pod \"machine-api-operator-5694c8668f-wpl2k\" (UID: \"f3993123-571c-4be4-a525-d8e0481b1aa5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493855 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/18378604-4db4-4da8-94c4-c52452ee6805-metrics-tls\") pod \"dns-operator-744455d44c-zxx6b\" (UID: \"18378604-4db4-4da8-94c4-c52452ee6805\") " pod="openshift-dns-operator/dns-operator-744455d44c-zxx6b" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493876 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-console-config\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493899 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9370d84-0bfb-4e61-a5b7-bcb43e903593-config\") pod \"authentication-operator-69f744f599-zm2tt\" (UID: \"e9370d84-0bfb-4e61-a5b7-bcb43e903593\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493922 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/800c31da-5f11-4029-a43f-ad96ac33b062-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9k5xc\" (UID: \"800c31da-5f11-4029-a43f-ad96ac33b062\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9k5xc" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493944 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwcmm\" (UniqueName: \"kubernetes.io/projected/821da4d3-7e2a-480e-8290-a7080b916b08-kube-api-access-hwcmm\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493966 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9370d84-0bfb-4e61-a5b7-bcb43e903593-service-ca-bundle\") pod \"authentication-operator-69f744f599-zm2tt\" (UID: \"e9370d84-0bfb-4e61-a5b7-bcb43e903593\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.493989 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494013 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0a38ab13-190d-40c5-9a69-389e2fae6884-client-ca\") pod \"route-controller-manager-6576b87f9c-db5rh\" (UID: \"0a38ab13-190d-40c5-9a69-389e2fae6884\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494037 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a38ab13-190d-40c5-9a69-389e2fae6884-serving-cert\") pod \"route-controller-manager-6576b87f9c-db5rh\" (UID: \"0a38ab13-190d-40c5-9a69-389e2fae6884\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494058 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/821da4d3-7e2a-480e-8290-a7080b916b08-encryption-config\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494082 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6a372ba-59d9-4fcc-ba21-7b9de5316239-serving-cert\") pod \"controller-manager-879f6c89f-57ksd\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494103 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c7face06-9e66-4375-8b67-974b8936289f-etcd-client\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494125 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494148 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494170 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/821da4d3-7e2a-480e-8290-a7080b916b08-audit-policies\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494191 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c7face06-9e66-4375-8b67-974b8936289f-encryption-config\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494213 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c6d7e0f-f064-48a0-b05b-287b09867b45-config\") pod \"machine-approver-56656f9798-649kx\" (UID: \"8c6d7e0f-f064-48a0-b05b-287b09867b45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494239 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc5np\" (UniqueName: \"kubernetes.io/projected/0836455d-d750-4a39-a7cc-6e95fe3e0abd-kube-api-access-kc5np\") pod \"downloads-7954f5f757-67z92\" (UID: \"0836455d-d750-4a39-a7cc-6e95fe3e0abd\") " pod="openshift-console/downloads-7954f5f757-67z92" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494262 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsqbk\" (UniqueName: \"kubernetes.io/projected/e6a372ba-59d9-4fcc-ba21-7b9de5316239-kube-api-access-jsqbk\") pod \"controller-manager-879f6c89f-57ksd\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494378 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb22l\" (UniqueName: \"kubernetes.io/projected/4c78711c-d17c-4b1c-b00e-f8ac8972f1d9-kube-api-access-mb22l\") pod \"openshift-controller-manager-operator-756b6f6bc6-qd6rv\" (UID: \"4c78711c-d17c-4b1c-b00e-f8ac8972f1d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qd6rv" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494606 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-client-ca\") pod \"controller-manager-879f6c89f-57ksd\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494671 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bdc72754-1951-4312-9dd0-3458357f2caa-available-featuregates\") pod \"openshift-config-operator-7777fb866f-jktft\" (UID: \"bdc72754-1951-4312-9dd0-3458357f2caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jktft" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494696 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0160154e-35ae-471c-b848-214b9f33ec93-console-serving-cert\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494791 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7face06-9e66-4375-8b67-974b8936289f-config\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494837 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4c37dd5-377e-4ed4-99f5-b179110153af-trusted-ca\") pod \"console-operator-58897d9998-qz7ht\" (UID: \"f4c37dd5-377e-4ed4-99f5-b179110153af\") " pod="openshift-console-operator/console-operator-58897d9998-qz7ht" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494908 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-config\") pod \"controller-manager-879f6c89f-57ksd\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494938 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9370d84-0bfb-4e61-a5b7-bcb43e903593-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zm2tt\" (UID: \"e9370d84-0bfb-4e61-a5b7-bcb43e903593\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.494977 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c7face06-9e66-4375-8b67-974b8936289f-etcd-serving-ca\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.495005 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c7face06-9e66-4375-8b67-974b8936289f-audit\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.495102 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.495181 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c7face06-9e66-4375-8b67-974b8936289f-image-import-ca\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.495252 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.495277 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.495332 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrbsm\" (UniqueName: \"kubernetes.io/projected/c7face06-9e66-4375-8b67-974b8936289f-kube-api-access-zrbsm\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.495363 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8c6d7e0f-f064-48a0-b05b-287b09867b45-machine-approver-tls\") pod \"machine-approver-56656f9798-649kx\" (UID: \"8c6d7e0f-f064-48a0-b05b-287b09867b45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.495414 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-audit-policies\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.500594 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5b8mm"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.510733 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4fgv6"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.511302 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f8k76"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.511641 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.512101 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.509236 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.512603 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4fgv6" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.512784 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f8k76" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.509247 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.509300 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.509433 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.509465 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.517540 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-brhqt"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.518302 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wl94z"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.518722 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wl94z" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.518775 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-brhqt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.519048 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fws6m"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.525951 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.526598 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.527004 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.527655 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fws6m" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.528261 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.532279 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.532511 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.533579 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.536994 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k9twj"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.537834 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-s8cx6"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.538377 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7snv7"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.538431 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.538399 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-s8cx6" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.539373 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7snv7" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.548725 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.550707 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.551449 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.551927 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.552179 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.553197 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs9km"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.553998 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7jtjl"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.554510 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7jtjl" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.555083 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs9km" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.555182 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-hbt6g"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.555861 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hbt6g" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.559424 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-67z92"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.559520 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.562170 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.563592 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.565728 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vbxst"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.566482 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qz7ht"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.566503 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-vbxst" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.569710 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wxs8l"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.573674 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qj4lx"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.574763 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.575305 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zxx6b"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.577002 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.577863 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.582887 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.586640 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.589794 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.592353 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5xbd9"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.594077 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-jktft"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.595242 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zm2tt"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596207 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3993123-571c-4be4-a525-d8e0481b1aa5-config\") pod \"machine-api-operator-5694c8668f-wpl2k\" (UID: \"f3993123-571c-4be4-a525-d8e0481b1aa5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596245 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f3993123-571c-4be4-a525-d8e0481b1aa5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wpl2k\" (UID: \"f3993123-571c-4be4-a525-d8e0481b1aa5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596280 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4c37dd5-377e-4ed4-99f5-b179110153af-serving-cert\") pod \"console-operator-58897d9998-qz7ht\" (UID: \"f4c37dd5-377e-4ed4-99f5-b179110153af\") " pod="openshift-console-operator/console-operator-58897d9998-qz7ht" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596298 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-57ksd\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596317 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-trusted-ca-bundle\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596334 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/821da4d3-7e2a-480e-8290-a7080b916b08-serving-cert\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596355 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596375 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c78711c-d17c-4b1c-b00e-f8ac8972f1d9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qd6rv\" (UID: \"4c78711c-d17c-4b1c-b00e-f8ac8972f1d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qd6rv" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596393 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/821da4d3-7e2a-480e-8290-a7080b916b08-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596412 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7face06-9e66-4375-8b67-974b8936289f-audit-dir\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596432 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596450 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4c37dd5-377e-4ed4-99f5-b179110153af-config\") pod \"console-operator-58897d9998-qz7ht\" (UID: \"f4c37dd5-377e-4ed4-99f5-b179110153af\") " pod="openshift-console-operator/console-operator-58897d9998-qz7ht" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596465 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c78711c-d17c-4b1c-b00e-f8ac8972f1d9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qd6rv\" (UID: \"4c78711c-d17c-4b1c-b00e-f8ac8972f1d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qd6rv" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596490 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/821da4d3-7e2a-480e-8290-a7080b916b08-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596508 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596526 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdc72754-1951-4312-9dd0-3458357f2caa-serving-cert\") pod \"openshift-config-operator-7777fb866f-jktft\" (UID: \"bdc72754-1951-4312-9dd0-3458357f2caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jktft" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596547 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzjz6\" (UniqueName: \"kubernetes.io/projected/0a38ab13-190d-40c5-9a69-389e2fae6884-kube-api-access-rzjz6\") pod \"route-controller-manager-6576b87f9c-db5rh\" (UID: \"0a38ab13-190d-40c5-9a69-389e2fae6884\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596565 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0160154e-35ae-471c-b848-214b9f33ec93-console-oauth-config\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596583 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p5st\" (UniqueName: \"kubernetes.io/projected/18378604-4db4-4da8-94c4-c52452ee6805-kube-api-access-7p5st\") pod \"dns-operator-744455d44c-zxx6b\" (UID: \"18378604-4db4-4da8-94c4-c52452ee6805\") " pod="openshift-dns-operator/dns-operator-744455d44c-zxx6b" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596601 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49f5s\" (UniqueName: \"kubernetes.io/projected/bdc72754-1951-4312-9dd0-3458357f2caa-kube-api-access-49f5s\") pod \"openshift-config-operator-7777fb866f-jktft\" (UID: \"bdc72754-1951-4312-9dd0-3458357f2caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jktft" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596638 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x7lm\" (UniqueName: \"kubernetes.io/projected/800c31da-5f11-4029-a43f-ad96ac33b062-kube-api-access-5x7lm\") pod \"cluster-samples-operator-665b6dd947-9k5xc\" (UID: \"800c31da-5f11-4029-a43f-ad96ac33b062\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9k5xc" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596656 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq6hc\" (UniqueName: \"kubernetes.io/projected/f3993123-571c-4be4-a525-d8e0481b1aa5-kube-api-access-wq6hc\") pod \"machine-api-operator-5694c8668f-wpl2k\" (UID: \"f3993123-571c-4be4-a525-d8e0481b1aa5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596674 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a38ab13-190d-40c5-9a69-389e2fae6884-config\") pod \"route-controller-manager-6576b87f9c-db5rh\" (UID: \"0a38ab13-190d-40c5-9a69-389e2fae6884\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596695 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596721 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8c6d7e0f-f064-48a0-b05b-287b09867b45-auth-proxy-config\") pod \"machine-approver-56656f9798-649kx\" (UID: \"8c6d7e0f-f064-48a0-b05b-287b09867b45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596753 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/18378604-4db4-4da8-94c4-c52452ee6805-metrics-tls\") pod \"dns-operator-744455d44c-zxx6b\" (UID: \"18378604-4db4-4da8-94c4-c52452ee6805\") " pod="openshift-dns-operator/dns-operator-744455d44c-zxx6b" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596772 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-console-config\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596791 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9370d84-0bfb-4e61-a5b7-bcb43e903593-config\") pod \"authentication-operator-69f744f599-zm2tt\" (UID: \"e9370d84-0bfb-4e61-a5b7-bcb43e903593\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596809 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f3993123-571c-4be4-a525-d8e0481b1aa5-images\") pod \"machine-api-operator-5694c8668f-wpl2k\" (UID: \"f3993123-571c-4be4-a525-d8e0481b1aa5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596826 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596848 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0a38ab13-190d-40c5-9a69-389e2fae6884-client-ca\") pod \"route-controller-manager-6576b87f9c-db5rh\" (UID: \"0a38ab13-190d-40c5-9a69-389e2fae6884\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596878 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a38ab13-190d-40c5-9a69-389e2fae6884-serving-cert\") pod \"route-controller-manager-6576b87f9c-db5rh\" (UID: \"0a38ab13-190d-40c5-9a69-389e2fae6884\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596901 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/800c31da-5f11-4029-a43f-ad96ac33b062-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9k5xc\" (UID: \"800c31da-5f11-4029-a43f-ad96ac33b062\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9k5xc" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596919 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwcmm\" (UniqueName: \"kubernetes.io/projected/821da4d3-7e2a-480e-8290-a7080b916b08-kube-api-access-hwcmm\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596936 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9370d84-0bfb-4e61-a5b7-bcb43e903593-service-ca-bundle\") pod \"authentication-operator-69f744f599-zm2tt\" (UID: \"e9370d84-0bfb-4e61-a5b7-bcb43e903593\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596955 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/821da4d3-7e2a-480e-8290-a7080b916b08-encryption-config\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596974 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.596991 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6a372ba-59d9-4fcc-ba21-7b9de5316239-serving-cert\") pod \"controller-manager-879f6c89f-57ksd\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597010 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c7face06-9e66-4375-8b67-974b8936289f-etcd-client\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597027 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597048 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc5np\" (UniqueName: \"kubernetes.io/projected/0836455d-d750-4a39-a7cc-6e95fe3e0abd-kube-api-access-kc5np\") pod \"downloads-7954f5f757-67z92\" (UID: \"0836455d-d750-4a39-a7cc-6e95fe3e0abd\") " pod="openshift-console/downloads-7954f5f757-67z92" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597065 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/821da4d3-7e2a-480e-8290-a7080b916b08-audit-policies\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597079 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c7face06-9e66-4375-8b67-974b8936289f-encryption-config\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597098 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c6d7e0f-f064-48a0-b05b-287b09867b45-config\") pod \"machine-approver-56656f9798-649kx\" (UID: \"8c6d7e0f-f064-48a0-b05b-287b09867b45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597114 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb22l\" (UniqueName: \"kubernetes.io/projected/4c78711c-d17c-4b1c-b00e-f8ac8972f1d9-kube-api-access-mb22l\") pod \"openshift-controller-manager-operator-756b6f6bc6-qd6rv\" (UID: \"4c78711c-d17c-4b1c-b00e-f8ac8972f1d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qd6rv" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597130 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-client-ca\") pod \"controller-manager-879f6c89f-57ksd\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597148 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsqbk\" (UniqueName: \"kubernetes.io/projected/e6a372ba-59d9-4fcc-ba21-7b9de5316239-kube-api-access-jsqbk\") pod \"controller-manager-879f6c89f-57ksd\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597168 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bdc72754-1951-4312-9dd0-3458357f2caa-available-featuregates\") pod \"openshift-config-operator-7777fb866f-jktft\" (UID: \"bdc72754-1951-4312-9dd0-3458357f2caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jktft" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597189 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0160154e-35ae-471c-b848-214b9f33ec93-console-serving-cert\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597191 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3993123-571c-4be4-a525-d8e0481b1aa5-config\") pod \"machine-api-operator-5694c8668f-wpl2k\" (UID: \"f3993123-571c-4be4-a525-d8e0481b1aa5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597210 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7face06-9e66-4375-8b67-974b8936289f-config\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597231 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-config\") pod \"controller-manager-879f6c89f-57ksd\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597229 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597255 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9370d84-0bfb-4e61-a5b7-bcb43e903593-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zm2tt\" (UID: \"e9370d84-0bfb-4e61-a5b7-bcb43e903593\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597276 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c7face06-9e66-4375-8b67-974b8936289f-etcd-serving-ca\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597298 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4c37dd5-377e-4ed4-99f5-b179110153af-trusted-ca\") pod \"console-operator-58897d9998-qz7ht\" (UID: \"f4c37dd5-377e-4ed4-99f5-b179110153af\") " pod="openshift-console-operator/console-operator-58897d9998-qz7ht" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597317 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c7face06-9e66-4375-8b67-974b8936289f-audit\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597344 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597542 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c7face06-9e66-4375-8b67-974b8936289f-image-import-ca\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597566 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-audit-policies\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597585 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597602 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597640 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrbsm\" (UniqueName: \"kubernetes.io/projected/c7face06-9e66-4375-8b67-974b8936289f-kube-api-access-zrbsm\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597658 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8c6d7e0f-f064-48a0-b05b-287b09867b45-machine-approver-tls\") pod \"machine-approver-56656f9798-649kx\" (UID: \"8c6d7e0f-f064-48a0-b05b-287b09867b45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597677 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a412ad65-d072-433a-8d6f-7e5390195f89-audit-dir\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597694 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxvjd\" (UniqueName: \"kubernetes.io/projected/a412ad65-d072-433a-8d6f-7e5390195f89-kube-api-access-vxvjd\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597713 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-service-ca\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597753 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7face06-9e66-4375-8b67-974b8936289f-serving-cert\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597774 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597798 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkmkk\" (UniqueName: \"kubernetes.io/projected/f4c37dd5-377e-4ed4-99f5-b179110153af-kube-api-access-gkmkk\") pod \"console-operator-58897d9998-qz7ht\" (UID: \"f4c37dd5-377e-4ed4-99f5-b179110153af\") " pod="openshift-console-operator/console-operator-58897d9998-qz7ht" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597823 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r72j7\" (UniqueName: \"kubernetes.io/projected/8c6d7e0f-f064-48a0-b05b-287b09867b45-kube-api-access-r72j7\") pod \"machine-approver-56656f9798-649kx\" (UID: \"8c6d7e0f-f064-48a0-b05b-287b09867b45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597840 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c7face06-9e66-4375-8b67-974b8936289f-node-pullsecrets\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597858 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-oauth-serving-cert\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597875 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/821da4d3-7e2a-480e-8290-a7080b916b08-audit-dir\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597875 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/821da4d3-7e2a-480e-8290-a7080b916b08-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597829 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.597996 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7face06-9e66-4375-8b67-974b8936289f-audit-dir\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.598069 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vs7j\" (UniqueName: \"kubernetes.io/projected/0160154e-35ae-471c-b848-214b9f33ec93-kube-api-access-6vs7j\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.598094 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/821da4d3-7e2a-480e-8290-a7080b916b08-etcd-client\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.598139 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9370d84-0bfb-4e61-a5b7-bcb43e903593-serving-cert\") pod \"authentication-operator-69f744f599-zm2tt\" (UID: \"e9370d84-0bfb-4e61-a5b7-bcb43e903593\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.598252 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzhn6\" (UniqueName: \"kubernetes.io/projected/e9370d84-0bfb-4e61-a5b7-bcb43e903593-kube-api-access-wzhn6\") pod \"authentication-operator-69f744f599-zm2tt\" (UID: \"e9370d84-0bfb-4e61-a5b7-bcb43e903593\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.598270 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7face06-9e66-4375-8b67-974b8936289f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.598990 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-57ksd\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.599844 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-trusted-ca-bundle\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.600076 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c78711c-d17c-4b1c-b00e-f8ac8972f1d9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qd6rv\" (UID: \"4c78711c-d17c-4b1c-b00e-f8ac8972f1d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qd6rv" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.600452 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.600544 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7face06-9e66-4375-8b67-974b8936289f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.600993 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4c37dd5-377e-4ed4-99f5-b179110153af-config\") pod \"console-operator-58897d9998-qz7ht\" (UID: \"f4c37dd5-377e-4ed4-99f5-b179110153af\") " pod="openshift-console-operator/console-operator-58897d9998-qz7ht" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.602569 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdc72754-1951-4312-9dd0-3458357f2caa-serving-cert\") pod \"openshift-config-operator-7777fb866f-jktft\" (UID: \"bdc72754-1951-4312-9dd0-3458357f2caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jktft" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.602899 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f3993123-571c-4be4-a525-d8e0481b1aa5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wpl2k\" (UID: \"f3993123-571c-4be4-a525-d8e0481b1aa5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.603038 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.603658 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qd6rv"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.603775 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.603816 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9370d84-0bfb-4e61-a5b7-bcb43e903593-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zm2tt\" (UID: \"e9370d84-0bfb-4e61-a5b7-bcb43e903593\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.603853 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0160154e-35ae-471c-b848-214b9f33ec93-console-oauth-config\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.604154 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a38ab13-190d-40c5-9a69-389e2fae6884-config\") pod \"route-controller-manager-6576b87f9c-db5rh\" (UID: \"0a38ab13-190d-40c5-9a69-389e2fae6884\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.604316 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4c37dd5-377e-4ed4-99f5-b179110153af-serving-cert\") pod \"console-operator-58897d9998-qz7ht\" (UID: \"f4c37dd5-377e-4ed4-99f5-b179110153af\") " pod="openshift-console-operator/console-operator-58897d9998-qz7ht" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.605044 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c7face06-9e66-4375-8b67-974b8936289f-etcd-client\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.605270 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c7face06-9e66-4375-8b67-974b8936289f-node-pullsecrets\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.605421 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wl94z"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.605750 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6a372ba-59d9-4fcc-ba21-7b9de5316239-serving-cert\") pod \"controller-manager-879f6c89f-57ksd\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.605895 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-oauth-serving-cert\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.605934 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/821da4d3-7e2a-480e-8290-a7080b916b08-audit-dir\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.606008 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c78711c-d17c-4b1c-b00e-f8ac8972f1d9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qd6rv\" (UID: \"4c78711c-d17c-4b1c-b00e-f8ac8972f1d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qd6rv" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.606482 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0a38ab13-190d-40c5-9a69-389e2fae6884-client-ca\") pod \"route-controller-manager-6576b87f9c-db5rh\" (UID: \"0a38ab13-190d-40c5-9a69-389e2fae6884\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.606694 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7face06-9e66-4375-8b67-974b8936289f-serving-cert\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.606745 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a412ad65-d072-433a-8d6f-7e5390195f89-audit-dir\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.606885 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.607127 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-client-ca\") pod \"controller-manager-879f6c89f-57ksd\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.607242 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bdc72754-1951-4312-9dd0-3458357f2caa-available-featuregates\") pod \"openshift-config-operator-7777fb866f-jktft\" (UID: \"bdc72754-1951-4312-9dd0-3458357f2caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jktft" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.607545 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/821da4d3-7e2a-480e-8290-a7080b916b08-serving-cert\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.607588 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/821da4d3-7e2a-480e-8290-a7080b916b08-audit-policies\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.608131 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/821da4d3-7e2a-480e-8290-a7080b916b08-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.608519 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-service-ca\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.608900 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k9twj"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.609313 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7face06-9e66-4375-8b67-974b8936289f-config\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.609341 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9370d84-0bfb-4e61-a5b7-bcb43e903593-service-ca-bundle\") pod \"authentication-operator-69f744f599-zm2tt\" (UID: \"e9370d84-0bfb-4e61-a5b7-bcb43e903593\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.609348 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vbxst"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.609559 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c6d7e0f-f064-48a0-b05b-287b09867b45-config\") pod \"machine-approver-56656f9798-649kx\" (UID: \"8c6d7e0f-f064-48a0-b05b-287b09867b45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.609628 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8c6d7e0f-f064-48a0-b05b-287b09867b45-machine-approver-tls\") pod \"machine-approver-56656f9798-649kx\" (UID: \"8c6d7e0f-f064-48a0-b05b-287b09867b45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.610550 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9370d84-0bfb-4e61-a5b7-bcb43e903593-config\") pod \"authentication-operator-69f744f599-zm2tt\" (UID: \"e9370d84-0bfb-4e61-a5b7-bcb43e903593\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.610555 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-console-config\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.610697 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c7face06-9e66-4375-8b67-974b8936289f-audit\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.610850 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c7face06-9e66-4375-8b67-974b8936289f-etcd-serving-ca\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.610949 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/821da4d3-7e2a-480e-8290-a7080b916b08-encryption-config\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.610995 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.611333 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f3993123-571c-4be4-a525-d8e0481b1aa5-images\") pod \"machine-api-operator-5694c8668f-wpl2k\" (UID: \"f3993123-571c-4be4-a525-d8e0481b1aa5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.611656 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a38ab13-190d-40c5-9a69-389e2fae6884-serving-cert\") pod \"route-controller-manager-6576b87f9c-db5rh\" (UID: \"0a38ab13-190d-40c5-9a69-389e2fae6884\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.611821 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4c37dd5-377e-4ed4-99f5-b179110153af-trusted-ca\") pod \"console-operator-58897d9998-qz7ht\" (UID: \"f4c37dd5-377e-4ed4-99f5-b179110153af\") " pod="openshift-console-operator/console-operator-58897d9998-qz7ht" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.612261 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/821da4d3-7e2a-480e-8290-a7080b916b08-etcd-client\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.612361 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c7face06-9e66-4375-8b67-974b8936289f-image-import-ca\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.612381 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0160154e-35ae-471c-b848-214b9f33ec93-console-serving-cert\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.612454 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5msxl"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.613111 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-config\") pod \"controller-manager-879f6c89f-57ksd\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.613379 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/18378604-4db4-4da8-94c4-c52452ee6805-metrics-tls\") pod \"dns-operator-744455d44c-zxx6b\" (UID: \"18378604-4db4-4da8-94c4-c52452ee6805\") " pod="openshift-dns-operator/dns-operator-744455d44c-zxx6b" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.613927 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8c6d7e0f-f064-48a0-b05b-287b09867b45-auth-proxy-config\") pod \"machine-approver-56656f9798-649kx\" (UID: \"8c6d7e0f-f064-48a0-b05b-287b09867b45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.615116 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-brhqt"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.615237 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f8k76"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.615959 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.616074 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9370d84-0bfb-4e61-a5b7-bcb43e903593-serving-cert\") pod \"authentication-operator-69f744f599-zm2tt\" (UID: \"e9370d84-0bfb-4e61-a5b7-bcb43e903593\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.616557 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/800c31da-5f11-4029-a43f-ad96ac33b062-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9k5xc\" (UID: \"800c31da-5f11-4029-a43f-ad96ac33b062\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9k5xc" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.617204 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.617582 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.618269 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.618330 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c7face06-9e66-4375-8b67-974b8936289f-encryption-config\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.619439 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.619921 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.621041 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.621108 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-audit-policies\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.621146 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4fgv6"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.621560 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.622273 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.623232 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7snv7"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.624493 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.624707 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-zcnpm"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.625742 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zcnpm" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.626241 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs9km"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.627638 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-swdhl"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.628218 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-swdhl" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.629173 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7jtjl"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.630227 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hbt6g"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.631524 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.632554 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zcnpm"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.633675 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fws6m"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.635594 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-s8cx6"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.638194 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pjknx"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.643346 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pjknx"] Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.643506 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.649525 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.661913 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.676562 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.697024 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.718029 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.736875 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.757726 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.776947 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.796680 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.816940 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.838116 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.857585 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.876913 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.897280 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.918016 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.936471 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.957049 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 05 06:57:24 crc kubenswrapper[4997]: I1205 06:57:24.976957 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.017645 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.037874 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.057201 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.077601 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.098672 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.117858 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.140456 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.158329 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.177918 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.196597 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.217777 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.237493 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.257973 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.277040 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.297061 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.316836 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.336972 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.357016 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.377884 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.397859 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.454817 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.455216 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.456209 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.476894 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.497128 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.517949 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.535693 4997 request.go:700] Waited for 1.006930087s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager-operator/configmaps?fieldSelector=metadata.name%3Dkube-controller-manager-operator-config&limit=500&resourceVersion=0 Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.537933 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.557798 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.577721 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.596239 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.619336 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.637815 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.656031 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:25 crc kubenswrapper[4997]: E1205 06:57:25.656339 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:59:27.65625012 +0000 UTC m=+268.185157421 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.656919 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.657852 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.659892 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.678377 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.713851 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.721738 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.738979 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.757918 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.757996 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.758138 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.758475 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.764247 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.765097 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.765095 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.778761 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.797970 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.817665 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.837197 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.856894 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.877285 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.897489 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.917048 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.937834 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.956858 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.977559 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.980565 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.992071 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:57:25 crc kubenswrapper[4997]: I1205 06:57:25.997078 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.000905 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.017863 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.037435 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.057289 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.100172 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.118001 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.138096 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.157905 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.177807 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.223219 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq6hc\" (UniqueName: \"kubernetes.io/projected/f3993123-571c-4be4-a525-d8e0481b1aa5-kube-api-access-wq6hc\") pod \"machine-api-operator-5694c8668f-wpl2k\" (UID: \"f3993123-571c-4be4-a525-d8e0481b1aa5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" Dec 05 06:57:26 crc kubenswrapper[4997]: W1205 06:57:26.223669 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-23fa0ed829b8058ecc79d2a9a045046099e511fef647c270759e9ae24bd763c3 WatchSource:0}: Error finding container 23fa0ed829b8058ecc79d2a9a045046099e511fef647c270759e9ae24bd763c3: Status 404 returned error can't find the container with id 23fa0ed829b8058ecc79d2a9a045046099e511fef647c270759e9ae24bd763c3 Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.233563 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p5st\" (UniqueName: \"kubernetes.io/projected/18378604-4db4-4da8-94c4-c52452ee6805-kube-api-access-7p5st\") pod \"dns-operator-744455d44c-zxx6b\" (UID: \"18378604-4db4-4da8-94c4-c52452ee6805\") " pod="openshift-dns-operator/dns-operator-744455d44c-zxx6b" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.251691 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzjz6\" (UniqueName: \"kubernetes.io/projected/0a38ab13-190d-40c5-9a69-389e2fae6884-kube-api-access-rzjz6\") pod \"route-controller-manager-6576b87f9c-db5rh\" (UID: \"0a38ab13-190d-40c5-9a69-389e2fae6884\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 06:57:26 crc kubenswrapper[4997]: W1205 06:57:26.252796 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-25404e0effc5a63e13e9003c67368bae7cdee67602a6933f37eca096653c26a8 WatchSource:0}: Error finding container 25404e0effc5a63e13e9003c67368bae7cdee67602a6933f37eca096653c26a8: Status 404 returned error can't find the container with id 25404e0effc5a63e13e9003c67368bae7cdee67602a6933f37eca096653c26a8 Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.274833 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwcmm\" (UniqueName: \"kubernetes.io/projected/821da4d3-7e2a-480e-8290-a7080b916b08-kube-api-access-hwcmm\") pod \"apiserver-7bbb656c7d-f4kdd\" (UID: \"821da4d3-7e2a-480e-8290-a7080b916b08\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.290415 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vs7j\" (UniqueName: \"kubernetes.io/projected/0160154e-35ae-471c-b848-214b9f33ec93-kube-api-access-6vs7j\") pod \"console-f9d7485db-lskhh\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.312123 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkmkk\" (UniqueName: \"kubernetes.io/projected/f4c37dd5-377e-4ed4-99f5-b179110153af-kube-api-access-gkmkk\") pod \"console-operator-58897d9998-qz7ht\" (UID: \"f4c37dd5-377e-4ed4-99f5-b179110153af\") " pod="openshift-console-operator/console-operator-58897d9998-qz7ht" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.331259 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r72j7\" (UniqueName: \"kubernetes.io/projected/8c6d7e0f-f064-48a0-b05b-287b09867b45-kube-api-access-r72j7\") pod \"machine-approver-56656f9798-649kx\" (UID: \"8c6d7e0f-f064-48a0-b05b-287b09867b45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.332007 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.345228 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-zxx6b" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.352544 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qz7ht" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.359598 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxvjd\" (UniqueName: \"kubernetes.io/projected/a412ad65-d072-433a-8d6f-7e5390195f89-kube-api-access-vxvjd\") pod \"oauth-openshift-558db77b4-wxs8l\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.374041 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.377796 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsqbk\" (UniqueName: \"kubernetes.io/projected/e6a372ba-59d9-4fcc-ba21-7b9de5316239-kube-api-access-jsqbk\") pod \"controller-manager-879f6c89f-57ksd\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.395590 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc5np\" (UniqueName: \"kubernetes.io/projected/0836455d-d750-4a39-a7cc-6e95fe3e0abd-kube-api-access-kc5np\") pod \"downloads-7954f5f757-67z92\" (UID: \"0836455d-d750-4a39-a7cc-6e95fe3e0abd\") " pod="openshift-console/downloads-7954f5f757-67z92" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.423900 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb22l\" (UniqueName: \"kubernetes.io/projected/4c78711c-d17c-4b1c-b00e-f8ac8972f1d9-kube-api-access-mb22l\") pod \"openshift-controller-manager-operator-756b6f6bc6-qd6rv\" (UID: \"4c78711c-d17c-4b1c-b00e-f8ac8972f1d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qd6rv" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.443955 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.444193 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzhn6\" (UniqueName: \"kubernetes.io/projected/e9370d84-0bfb-4e61-a5b7-bcb43e903593-kube-api-access-wzhn6\") pod \"authentication-operator-69f744f599-zm2tt\" (UID: \"e9370d84-0bfb-4e61-a5b7-bcb43e903593\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.453604 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.454023 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrbsm\" (UniqueName: \"kubernetes.io/projected/c7face06-9e66-4375-8b67-974b8936289f-kube-api-access-zrbsm\") pod \"apiserver-76f77b778f-5b8mm\" (UID: \"c7face06-9e66-4375-8b67-974b8936289f\") " pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.468550 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.476999 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x7lm\" (UniqueName: \"kubernetes.io/projected/800c31da-5f11-4029-a43f-ad96ac33b062-kube-api-access-5x7lm\") pod \"cluster-samples-operator-665b6dd947-9k5xc\" (UID: \"800c31da-5f11-4029-a43f-ad96ac33b062\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9k5xc" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.497604 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.498806 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"471f98cbf412d4dd6aa7e39adb493b20c03cd0d203c8697fd40650f03b80d214"} Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.498876 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"10e914c5169a93a73368ca90effdfe54dc218da6846e733cf1bff630f79a613a"} Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.498936 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49f5s\" (UniqueName: \"kubernetes.io/projected/bdc72754-1951-4312-9dd0-3458357f2caa-kube-api-access-49f5s\") pod \"openshift-config-operator-7777fb866f-jktft\" (UID: \"bdc72754-1951-4312-9dd0-3458357f2caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jktft" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.501587 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ec92d4eda2961d9b0d4e3ffbff21f9bab0c848e1bf18d59f484aca0ec1c9e3be"} Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.501666 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"23fa0ed829b8058ecc79d2a9a045046099e511fef647c270759e9ae24bd763c3"} Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.501830 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.502854 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"20e2bb25d1e1c679c13228ec54021e4c1d0aa7c73db1be8c591fdbb5b2d84b33"} Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.502883 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"25404e0effc5a63e13e9003c67368bae7cdee67602a6933f37eca096653c26a8"} Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.509934 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.517102 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.523525 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9k5xc" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.536409 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.545353 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-lskhh"] Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.556359 4997 request.go:700] Waited for 1.927868855s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-tls&limit=500&resourceVersion=0 Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.558268 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.559051 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.565739 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jktft" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.578947 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.599963 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.609125 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.618788 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.637647 4997 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.658379 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.667812 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-67z92" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.687633 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qd6rv" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.700966 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.766305 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-57ksd"] Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781226 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0dc484d1-5411-48aa-81d8-c208d0c5e2a6-metrics-tls\") pod \"ingress-operator-5b745b69d9-n5wgm\" (UID: \"0dc484d1-5411-48aa-81d8-c208d0c5e2a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781256 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0dc484d1-5411-48aa-81d8-c208d0c5e2a6-trusted-ca\") pod \"ingress-operator-5b745b69d9-n5wgm\" (UID: \"0dc484d1-5411-48aa-81d8-c208d0c5e2a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781294 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/560926cd-d0fe-4823-b217-803ba02d2702-stats-auth\") pod \"router-default-5444994796-v4mxn\" (UID: \"560926cd-d0fe-4823-b217-803ba02d2702\") " pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781320 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781337 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cd09450e-798d-41d5-b925-90ab15df8dac-trusted-ca\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781375 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/560926cd-d0fe-4823-b217-803ba02d2702-service-ca-bundle\") pod \"router-default-5444994796-v4mxn\" (UID: \"560926cd-d0fe-4823-b217-803ba02d2702\") " pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781390 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/560926cd-d0fe-4823-b217-803ba02d2702-metrics-certs\") pod \"router-default-5444994796-v4mxn\" (UID: \"560926cd-d0fe-4823-b217-803ba02d2702\") " pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781416 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ff56ef31-a073-49ac-b00a-e9611e2e0025-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-7pzd7\" (UID: \"ff56ef31-a073-49ac-b00a-e9611e2e0025\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781465 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cd09450e-798d-41d5-b925-90ab15df8dac-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781489 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cd09450e-798d-41d5-b925-90ab15df8dac-registry-certificates\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781563 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65e6a96e-df08-432d-8319-93e3e483f126-config\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781640 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr7gj\" (UniqueName: \"kubernetes.io/projected/65e6a96e-df08-432d-8319-93e3e483f126-kube-api-access-zr7gj\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781707 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/65e6a96e-df08-432d-8319-93e3e483f126-etcd-client\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781732 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cd09450e-798d-41d5-b925-90ab15df8dac-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781778 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0dc484d1-5411-48aa-81d8-c208d0c5e2a6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n5wgm\" (UID: \"0dc484d1-5411-48aa-81d8-c208d0c5e2a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781794 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df171b06-9a20-42ed-8caa-c9f8aebd611c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qj4lx\" (UID: \"df171b06-9a20-42ed-8caa-c9f8aebd611c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qj4lx" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781859 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkrch\" (UniqueName: \"kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-kube-api-access-vkrch\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781883 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/65e6a96e-df08-432d-8319-93e3e483f126-etcd-service-ca\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781944 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-registry-tls\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.781962 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9b7r\" (UniqueName: \"kubernetes.io/projected/ff56ef31-a073-49ac-b00a-e9611e2e0025-kube-api-access-c9b7r\") pod \"cluster-image-registry-operator-dc59b4c8b-7pzd7\" (UID: \"ff56ef31-a073-49ac-b00a-e9611e2e0025\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.782025 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz8xx\" (UniqueName: \"kubernetes.io/projected/560926cd-d0fe-4823-b217-803ba02d2702-kube-api-access-bz8xx\") pod \"router-default-5444994796-v4mxn\" (UID: \"560926cd-d0fe-4823-b217-803ba02d2702\") " pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.782040 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df171b06-9a20-42ed-8caa-c9f8aebd611c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qj4lx\" (UID: \"df171b06-9a20-42ed-8caa-c9f8aebd611c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qj4lx" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.782061 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/65e6a96e-df08-432d-8319-93e3e483f126-etcd-ca\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.782096 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65e6a96e-df08-432d-8319-93e3e483f126-serving-cert\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.782112 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-bound-sa-token\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.782143 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45fw4\" (UniqueName: \"kubernetes.io/projected/df171b06-9a20-42ed-8caa-c9f8aebd611c-kube-api-access-45fw4\") pod \"openshift-apiserver-operator-796bbdcf4f-qj4lx\" (UID: \"df171b06-9a20-42ed-8caa-c9f8aebd611c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qj4lx" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.782183 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff56ef31-a073-49ac-b00a-e9611e2e0025-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-7pzd7\" (UID: \"ff56ef31-a073-49ac-b00a-e9611e2e0025\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.782198 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/560926cd-d0fe-4823-b217-803ba02d2702-default-certificate\") pod \"router-default-5444994796-v4mxn\" (UID: \"560926cd-d0fe-4823-b217-803ba02d2702\") " pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.782241 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ff56ef31-a073-49ac-b00a-e9611e2e0025-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-7pzd7\" (UID: \"ff56ef31-a073-49ac-b00a-e9611e2e0025\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.782279 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8nrl\" (UniqueName: \"kubernetes.io/projected/0dc484d1-5411-48aa-81d8-c208d0c5e2a6-kube-api-access-z8nrl\") pod \"ingress-operator-5b745b69d9-n5wgm\" (UID: \"0dc484d1-5411-48aa-81d8-c208d0c5e2a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" Dec 05 06:57:26 crc kubenswrapper[4997]: E1205 06:57:26.783888 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:27.283866879 +0000 UTC m=+147.812774140 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:26 crc kubenswrapper[4997]: W1205 06:57:26.817267 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6a372ba_59d9_4fcc_ba21_7b9de5316239.slice/crio-e6ef3c4bbe66efeeea728677ebda5fc005e6b7cb8728228bcc91270d9605c7f2 WatchSource:0}: Error finding container e6ef3c4bbe66efeeea728677ebda5fc005e6b7cb8728228bcc91270d9605c7f2: Status 404 returned error can't find the container with id e6ef3c4bbe66efeeea728677ebda5fc005e6b7cb8728228bcc91270d9605c7f2 Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.861026 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wxs8l"] Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.862424 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qz7ht"] Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.882889 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zxx6b"] Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.883465 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:26 crc kubenswrapper[4997]: E1205 06:57:26.883535 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:27.383517497 +0000 UTC m=+147.912424758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.885522 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ff56ef31-a073-49ac-b00a-e9611e2e0025-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-7pzd7\" (UID: \"ff56ef31-a073-49ac-b00a-e9611e2e0025\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.885624 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cd09450e-798d-41d5-b925-90ab15df8dac-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.885706 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3-config\") pod \"kube-apiserver-operator-766d6c64bb-f8k76\" (UID: \"33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f8k76" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.885790 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17fe9bd5-5973-47bd-945f-55bb8caa9d65-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k9twj\" (UID: \"17fe9bd5-5973-47bd-945f-55bb8caa9d65\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.885876 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65e6a96e-df08-432d-8319-93e3e483f126-config\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.885988 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr7gj\" (UniqueName: \"kubernetes.io/projected/65e6a96e-df08-432d-8319-93e3e483f126-kube-api-access-zr7gj\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.886112 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/65e6a96e-df08-432d-8319-93e3e483f126-etcd-client\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.887296 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jnjd\" (UniqueName: \"kubernetes.io/projected/e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6-kube-api-access-9jnjd\") pod \"olm-operator-6b444d44fb-8vkgq\" (UID: \"e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.887399 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1ea90520-ebc6-41e2-9854-54caf800fda7-tmpfs\") pod \"packageserver-d55dfcdfc-rdtmj\" (UID: \"1ea90520-ebc6-41e2-9854-54caf800fda7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.887468 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6008a937-1ce8-48ba-b1eb-bf915c452468-node-bootstrap-token\") pod \"machine-config-server-swdhl\" (UID: \"6008a937-1ce8-48ba-b1eb-bf915c452468\") " pod="openshift-machine-config-operator/machine-config-server-swdhl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.887563 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0dc484d1-5411-48aa-81d8-c208d0c5e2a6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n5wgm\" (UID: \"0dc484d1-5411-48aa-81d8-c208d0c5e2a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.887656 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2be0da6d-6565-4336-89e0-590353e31685-serving-cert\") pod \"service-ca-operator-777779d784-7jtjl\" (UID: \"2be0da6d-6565-4336-89e0-590353e31685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7jtjl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.887729 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28rbk\" (UniqueName: \"kubernetes.io/projected/49b1c6db-0638-4367-8264-eec97250ad0a-kube-api-access-28rbk\") pod \"dns-default-zcnpm\" (UID: \"49b1c6db-0638-4367-8264-eec97250ad0a\") " pod="openshift-dns/dns-default-zcnpm" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.887844 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df171b06-9a20-42ed-8caa-c9f8aebd611c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qj4lx\" (UID: \"df171b06-9a20-42ed-8caa-c9f8aebd611c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qj4lx" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.887925 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-csi-data-dir\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.888020 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgknv\" (UniqueName: \"kubernetes.io/projected/3c394687-7ae3-4500-bda2-a5de1cb7b8b1-kube-api-access-fgknv\") pod \"kube-storage-version-migrator-operator-b67b599dd-4fgv6\" (UID: \"3c394687-7ae3-4500-bda2-a5de1cb7b8b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4fgv6" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.888120 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1db3923d-a8bf-4ed6-b30e-4bdf15471401-proxy-tls\") pod \"machine-config-controller-84d6567774-wsvhz\" (UID: \"1db3923d-a8bf-4ed6-b30e-4bdf15471401\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.888207 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/49b1c6db-0638-4367-8264-eec97250ad0a-metrics-tls\") pod \"dns-default-zcnpm\" (UID: \"49b1c6db-0638-4367-8264-eec97250ad0a\") " pod="openshift-dns/dns-default-zcnpm" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.888285 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz6q4\" (UniqueName: \"kubernetes.io/projected/80dca7ef-76c4-4616-9d8a-81ff066f6464-kube-api-access-dz6q4\") pod \"machine-config-operator-74547568cd-2kfps\" (UID: \"80dca7ef-76c4-4616-9d8a-81ff066f6464\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.888353 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6-srv-cert\") pod \"olm-operator-6b444d44fb-8vkgq\" (UID: \"e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.888434 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/49b1c6db-0638-4367-8264-eec97250ad0a-config-volume\") pod \"dns-default-zcnpm\" (UID: \"49b1c6db-0638-4367-8264-eec97250ad0a\") " pod="openshift-dns/dns-default-zcnpm" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.889373 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s4gz\" (UniqueName: \"kubernetes.io/projected/e0d60887-d7ee-4e14-96f1-81d1283ffa2e-kube-api-access-2s4gz\") pod \"package-server-manager-789f6589d5-s8cx6\" (UID: \"e0d60887-d7ee-4e14-96f1-81d1283ffa2e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-s8cx6" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.889477 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f915acae-6293-415a-a2a4-266340eee2e9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fws6m\" (UID: \"f915acae-6293-415a-a2a4-266340eee2e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fws6m" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.889574 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg89t\" (UniqueName: \"kubernetes.io/projected/6008a937-1ce8-48ba-b1eb-bf915c452468-kube-api-access-fg89t\") pod \"machine-config-server-swdhl\" (UID: \"6008a937-1ce8-48ba-b1eb-bf915c452468\") " pod="openshift-machine-config-operator/machine-config-server-swdhl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.889702 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-f8k76\" (UID: \"33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f8k76" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.889802 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0d60887-d7ee-4e14-96f1-81d1283ffa2e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-s8cx6\" (UID: \"e0d60887-d7ee-4e14-96f1-81d1283ffa2e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-s8cx6" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.889885 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f692422b-da9f-4775-825e-ee198e5b61b8-config-volume\") pod \"collect-profiles-29415285-28q92\" (UID: \"f692422b-da9f-4775-825e-ee198e5b61b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.889970 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/487642f9-0987-4953-b369-f092392d0209-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wl94z\" (UID: \"487642f9-0987-4953-b369-f092392d0209\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wl94z" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.890045 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5pjp\" (UniqueName: \"kubernetes.io/projected/1ea90520-ebc6-41e2-9854-54caf800fda7-kube-api-access-c5pjp\") pod \"packageserver-d55dfcdfc-rdtmj\" (UID: \"1ea90520-ebc6-41e2-9854-54caf800fda7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.890142 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-bound-sa-token\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.890239 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45fw4\" (UniqueName: \"kubernetes.io/projected/df171b06-9a20-42ed-8caa-c9f8aebd611c-kube-api-access-45fw4\") pod \"openshift-apiserver-operator-796bbdcf4f-qj4lx\" (UID: \"df171b06-9a20-42ed-8caa-c9f8aebd611c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qj4lx" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.890493 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkfz8\" (UniqueName: \"kubernetes.io/projected/cf990560-e430-4c2b-8965-c5ea18f8b695-kube-api-access-bkfz8\") pod \"catalog-operator-68c6474976-d28rt\" (UID: \"cf990560-e430-4c2b-8965-c5ea18f8b695\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.891667 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1247bce5-d4bf-4a8b-a438-3b150109ef5f-signing-cabundle\") pod \"service-ca-9c57cc56f-vbxst\" (UID: \"1247bce5-d4bf-4a8b-a438-3b150109ef5f\") " pod="openshift-service-ca/service-ca-9c57cc56f-vbxst" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.891808 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f692422b-da9f-4775-825e-ee198e5b61b8-secret-volume\") pod \"collect-profiles-29415285-28q92\" (UID: \"f692422b-da9f-4775-825e-ee198e5b61b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.891897 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xkcb\" (UniqueName: \"kubernetes.io/projected/eba9ab59-079c-4512-b47c-b6c0e2a2886a-kube-api-access-9xkcb\") pod \"ingress-canary-hbt6g\" (UID: \"eba9ab59-079c-4512-b47c-b6c0e2a2886a\") " pod="openshift-ingress-canary/ingress-canary-hbt6g" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.891971 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c394687-7ae3-4500-bda2-a5de1cb7b8b1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4fgv6\" (UID: \"3c394687-7ae3-4500-bda2-a5de1cb7b8b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4fgv6" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.892072 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8nrl\" (UniqueName: \"kubernetes.io/projected/0dc484d1-5411-48aa-81d8-c208d0c5e2a6-kube-api-access-z8nrl\") pod \"ingress-operator-5b745b69d9-n5wgm\" (UID: \"0dc484d1-5411-48aa-81d8-c208d0c5e2a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.892168 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f915acae-6293-415a-a2a4-266340eee2e9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fws6m\" (UID: \"f915acae-6293-415a-a2a4-266340eee2e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fws6m" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.892270 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cf990560-e430-4c2b-8965-c5ea18f8b695-profile-collector-cert\") pod \"catalog-operator-68c6474976-d28rt\" (UID: \"cf990560-e430-4c2b-8965-c5ea18f8b695\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.892363 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/560926cd-d0fe-4823-b217-803ba02d2702-stats-auth\") pod \"router-default-5444994796-v4mxn\" (UID: \"560926cd-d0fe-4823-b217-803ba02d2702\") " pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.892446 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/487642f9-0987-4953-b369-f092392d0209-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wl94z\" (UID: \"487642f9-0987-4953-b369-f092392d0209\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wl94z" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.892569 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cd09450e-798d-41d5-b925-90ab15df8dac-trusted-ca\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.892667 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/560926cd-d0fe-4823-b217-803ba02d2702-service-ca-bundle\") pod \"router-default-5444994796-v4mxn\" (UID: \"560926cd-d0fe-4823-b217-803ba02d2702\") " pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.894408 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/560926cd-d0fe-4823-b217-803ba02d2702-metrics-certs\") pod \"router-default-5444994796-v4mxn\" (UID: \"560926cd-d0fe-4823-b217-803ba02d2702\") " pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.895795 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwbbp\" (UniqueName: \"kubernetes.io/projected/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-kube-api-access-jwbbp\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.896001 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.896167 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1db3923d-a8bf-4ed6-b30e-4bdf15471401-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wsvhz\" (UID: \"1db3923d-a8bf-4ed6-b30e-4bdf15471401\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.896368 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1247bce5-d4bf-4a8b-a438-3b150109ef5f-signing-key\") pod \"service-ca-9c57cc56f-vbxst\" (UID: \"1247bce5-d4bf-4a8b-a438-3b150109ef5f\") " pod="openshift-service-ca/service-ca-9c57cc56f-vbxst" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.896456 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-mountpoint-dir\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.896526 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmrmp\" (UniqueName: \"kubernetes.io/projected/17fe9bd5-5973-47bd-945f-55bb8caa9d65-kube-api-access-jmrmp\") pod \"marketplace-operator-79b997595-k9twj\" (UID: \"17fe9bd5-5973-47bd-945f-55bb8caa9d65\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.896606 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cd09450e-798d-41d5-b925-90ab15df8dac-registry-certificates\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.896709 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6008a937-1ce8-48ba-b1eb-bf915c452468-certs\") pod \"machine-config-server-swdhl\" (UID: \"6008a937-1ce8-48ba-b1eb-bf915c452468\") " pod="openshift-machine-config-operator/machine-config-server-swdhl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.895252 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65e6a96e-df08-432d-8319-93e3e483f126-config\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:26 crc kubenswrapper[4997]: E1205 06:57:26.896769 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:27.396748564 +0000 UTC m=+147.925655825 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.893965 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/560926cd-d0fe-4823-b217-803ba02d2702-service-ca-bundle\") pod \"router-default-5444994796-v4mxn\" (UID: \"560926cd-d0fe-4823-b217-803ba02d2702\") " pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.897695 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-registration-dir\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.897735 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cd09450e-798d-41d5-b925-90ab15df8dac-trusted-ca\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.897746 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6-profile-collector-cert\") pod \"olm-operator-6b444d44fb-8vkgq\" (UID: \"e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.897799 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4h54\" (UniqueName: \"kubernetes.io/projected/1db3923d-a8bf-4ed6-b30e-4bdf15471401-kube-api-access-q4h54\") pod \"machine-config-controller-84d6567774-wsvhz\" (UID: \"1db3923d-a8bf-4ed6-b30e-4bdf15471401\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.897943 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/80dca7ef-76c4-4616-9d8a-81ff066f6464-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2kfps\" (UID: \"80dca7ef-76c4-4616-9d8a-81ff066f6464\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898098 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cd09450e-798d-41d5-b925-90ab15df8dac-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898146 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df171b06-9a20-42ed-8caa-c9f8aebd611c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qj4lx\" (UID: \"df171b06-9a20-42ed-8caa-c9f8aebd611c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qj4lx" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898214 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eba9ab59-079c-4512-b47c-b6c0e2a2886a-cert\") pod \"ingress-canary-hbt6g\" (UID: \"eba9ab59-079c-4512-b47c-b6c0e2a2886a\") " pod="openshift-ingress-canary/ingress-canary-hbt6g" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898280 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1ea90520-ebc6-41e2-9854-54caf800fda7-webhook-cert\") pod \"packageserver-d55dfcdfc-rdtmj\" (UID: \"1ea90520-ebc6-41e2-9854-54caf800fda7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898301 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cd09450e-798d-41d5-b925-90ab15df8dac-registry-certificates\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898382 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/baf64b43-1f21-4b84-ac70-3e0d879244a5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7snv7\" (UID: \"baf64b43-1f21-4b84-ac70-3e0d879244a5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7snv7" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898449 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/80dca7ef-76c4-4616-9d8a-81ff066f6464-proxy-tls\") pod \"machine-config-operator-74547568cd-2kfps\" (UID: \"80dca7ef-76c4-4616-9d8a-81ff066f6464\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898499 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkrch\" (UniqueName: \"kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-kube-api-access-vkrch\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898527 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-f8k76\" (UID: \"33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f8k76" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898573 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/65e6a96e-df08-432d-8319-93e3e483f126-etcd-service-ca\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898642 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-registry-tls\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898678 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8v9t\" (UniqueName: \"kubernetes.io/projected/baf64b43-1f21-4b84-ac70-3e0d879244a5-kube-api-access-v8v9t\") pod \"multus-admission-controller-857f4d67dd-7snv7\" (UID: \"baf64b43-1f21-4b84-ac70-3e0d879244a5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7snv7" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898733 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68bfb\" (UniqueName: \"kubernetes.io/projected/a4b45170-1957-4ea8-b27e-e08dd68a56cd-kube-api-access-68bfb\") pod \"migrator-59844c95c7-brhqt\" (UID: \"a4b45170-1957-4ea8-b27e-e08dd68a56cd\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-brhqt" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898778 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9b7r\" (UniqueName: \"kubernetes.io/projected/ff56ef31-a073-49ac-b00a-e9611e2e0025-kube-api-access-c9b7r\") pod \"cluster-image-registry-operator-dc59b4c8b-7pzd7\" (UID: \"ff56ef31-a073-49ac-b00a-e9611e2e0025\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898833 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz8xx\" (UniqueName: \"kubernetes.io/projected/560926cd-d0fe-4823-b217-803ba02d2702-kube-api-access-bz8xx\") pod \"router-default-5444994796-v4mxn\" (UID: \"560926cd-d0fe-4823-b217-803ba02d2702\") " pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898858 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df171b06-9a20-42ed-8caa-c9f8aebd611c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qj4lx\" (UID: \"df171b06-9a20-42ed-8caa-c9f8aebd611c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qj4lx" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898923 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/80dca7ef-76c4-4616-9d8a-81ff066f6464-images\") pod \"machine-config-operator-74547568cd-2kfps\" (UID: \"80dca7ef-76c4-4616-9d8a-81ff066f6464\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.898979 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbng5\" (UniqueName: \"kubernetes.io/projected/f692422b-da9f-4775-825e-ee198e5b61b8-kube-api-access-kbng5\") pod \"collect-profiles-29415285-28q92\" (UID: \"f692422b-da9f-4775-825e-ee198e5b61b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899007 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1ea90520-ebc6-41e2-9854-54caf800fda7-apiservice-cert\") pod \"packageserver-d55dfcdfc-rdtmj\" (UID: \"1ea90520-ebc6-41e2-9854-54caf800fda7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899058 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/65e6a96e-df08-432d-8319-93e3e483f126-etcd-ca\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899080 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65e6a96e-df08-432d-8319-93e3e483f126-serving-cert\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899127 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2be0da6d-6565-4336-89e0-590353e31685-config\") pod \"service-ca-operator-777779d784-7jtjl\" (UID: \"2be0da6d-6565-4336-89e0-590353e31685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7jtjl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899151 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f915acae-6293-415a-a2a4-266340eee2e9-config\") pod \"kube-controller-manager-operator-78b949d7b-fws6m\" (UID: \"f915acae-6293-415a-a2a4-266340eee2e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fws6m" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899176 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/17fe9bd5-5973-47bd-945f-55bb8caa9d65-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k9twj\" (UID: \"17fe9bd5-5973-47bd-945f-55bb8caa9d65\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899228 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2t2h\" (UniqueName: \"kubernetes.io/projected/1247bce5-d4bf-4a8b-a438-3b150109ef5f-kube-api-access-q2t2h\") pod \"service-ca-9c57cc56f-vbxst\" (UID: \"1247bce5-d4bf-4a8b-a438-3b150109ef5f\") " pod="openshift-service-ca/service-ca-9c57cc56f-vbxst" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899249 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq96d\" (UniqueName: \"kubernetes.io/projected/2be0da6d-6565-4336-89e0-590353e31685-kube-api-access-gq96d\") pod \"service-ca-operator-777779d784-7jtjl\" (UID: \"2be0da6d-6565-4336-89e0-590353e31685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7jtjl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899294 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/65e6a96e-df08-432d-8319-93e3e483f126-etcd-service-ca\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899297 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cf990560-e430-4c2b-8965-c5ea18f8b695-srv-cert\") pod \"catalog-operator-68c6474976-d28rt\" (UID: \"cf990560-e430-4c2b-8965-c5ea18f8b695\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899410 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ab4d334f-b4a4-4d04-bf90-d2178f991f5f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zs9km\" (UID: \"ab4d334f-b4a4-4d04-bf90-d2178f991f5f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs9km" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899450 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff56ef31-a073-49ac-b00a-e9611e2e0025-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-7pzd7\" (UID: \"ff56ef31-a073-49ac-b00a-e9611e2e0025\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899477 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/560926cd-d0fe-4823-b217-803ba02d2702-default-certificate\") pod \"router-default-5444994796-v4mxn\" (UID: \"560926cd-d0fe-4823-b217-803ba02d2702\") " pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899504 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c394687-7ae3-4500-bda2-a5de1cb7b8b1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4fgv6\" (UID: \"3c394687-7ae3-4500-bda2-a5de1cb7b8b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4fgv6" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899534 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-socket-dir\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899563 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ff56ef31-a073-49ac-b00a-e9611e2e0025-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-7pzd7\" (UID: \"ff56ef31-a073-49ac-b00a-e9611e2e0025\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899629 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/487642f9-0987-4953-b369-f092392d0209-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wl94z\" (UID: \"487642f9-0987-4953-b369-f092392d0209\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wl94z" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899701 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0dc484d1-5411-48aa-81d8-c208d0c5e2a6-metrics-tls\") pod \"ingress-operator-5b745b69d9-n5wgm\" (UID: \"0dc484d1-5411-48aa-81d8-c208d0c5e2a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899733 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0dc484d1-5411-48aa-81d8-c208d0c5e2a6-trusted-ca\") pod \"ingress-operator-5b745b69d9-n5wgm\" (UID: \"0dc484d1-5411-48aa-81d8-c208d0c5e2a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899813 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-plugins-dir\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.899847 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z8ws\" (UniqueName: \"kubernetes.io/projected/ab4d334f-b4a4-4d04-bf90-d2178f991f5f-kube-api-access-2z8ws\") pod \"control-plane-machine-set-operator-78cbb6b69f-zs9km\" (UID: \"ab4d334f-b4a4-4d04-bf90-d2178f991f5f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs9km" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.900135 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cd09450e-798d-41d5-b925-90ab15df8dac-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.901020 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0dc484d1-5411-48aa-81d8-c208d0c5e2a6-trusted-ca\") pod \"ingress-operator-5b745b69d9-n5wgm\" (UID: \"0dc484d1-5411-48aa-81d8-c208d0c5e2a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.901497 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff56ef31-a073-49ac-b00a-e9611e2e0025-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-7pzd7\" (UID: \"ff56ef31-a073-49ac-b00a-e9611e2e0025\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.901922 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/560926cd-d0fe-4823-b217-803ba02d2702-stats-auth\") pod \"router-default-5444994796-v4mxn\" (UID: \"560926cd-d0fe-4823-b217-803ba02d2702\") " pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.902045 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/65e6a96e-df08-432d-8319-93e3e483f126-etcd-ca\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.907406 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0dc484d1-5411-48aa-81d8-c208d0c5e2a6-metrics-tls\") pod \"ingress-operator-5b745b69d9-n5wgm\" (UID: \"0dc484d1-5411-48aa-81d8-c208d0c5e2a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.907563 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ff56ef31-a073-49ac-b00a-e9611e2e0025-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-7pzd7\" (UID: \"ff56ef31-a073-49ac-b00a-e9611e2e0025\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.909776 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-registry-tls\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.910032 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/560926cd-d0fe-4823-b217-803ba02d2702-default-certificate\") pod \"router-default-5444994796-v4mxn\" (UID: \"560926cd-d0fe-4823-b217-803ba02d2702\") " pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.910256 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cd09450e-798d-41d5-b925-90ab15df8dac-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.910333 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df171b06-9a20-42ed-8caa-c9f8aebd611c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qj4lx\" (UID: \"df171b06-9a20-42ed-8caa-c9f8aebd611c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qj4lx" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.910725 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/65e6a96e-df08-432d-8319-93e3e483f126-etcd-client\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.911098 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65e6a96e-df08-432d-8319-93e3e483f126-serving-cert\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.912056 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/560926cd-d0fe-4823-b217-803ba02d2702-metrics-certs\") pod \"router-default-5444994796-v4mxn\" (UID: \"560926cd-d0fe-4823-b217-803ba02d2702\") " pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.952165 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8nrl\" (UniqueName: \"kubernetes.io/projected/0dc484d1-5411-48aa-81d8-c208d0c5e2a6-kube-api-access-z8nrl\") pod \"ingress-operator-5b745b69d9-n5wgm\" (UID: \"0dc484d1-5411-48aa-81d8-c208d0c5e2a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.963125 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-bound-sa-token\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:26 crc kubenswrapper[4997]: I1205 06:57:26.982995 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45fw4\" (UniqueName: \"kubernetes.io/projected/df171b06-9a20-42ed-8caa-c9f8aebd611c-kube-api-access-45fw4\") pod \"openshift-apiserver-operator-796bbdcf4f-qj4lx\" (UID: \"df171b06-9a20-42ed-8caa-c9f8aebd611c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qj4lx" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.004764 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.004970 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/baf64b43-1f21-4b84-ac70-3e0d879244a5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7snv7\" (UID: \"baf64b43-1f21-4b84-ac70-3e0d879244a5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7snv7" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.004995 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1ea90520-ebc6-41e2-9854-54caf800fda7-webhook-cert\") pod \"packageserver-d55dfcdfc-rdtmj\" (UID: \"1ea90520-ebc6-41e2-9854-54caf800fda7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005023 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-f8k76\" (UID: \"33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f8k76" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005041 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/80dca7ef-76c4-4616-9d8a-81ff066f6464-proxy-tls\") pod \"machine-config-operator-74547568cd-2kfps\" (UID: \"80dca7ef-76c4-4616-9d8a-81ff066f6464\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005062 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68bfb\" (UniqueName: \"kubernetes.io/projected/a4b45170-1957-4ea8-b27e-e08dd68a56cd-kube-api-access-68bfb\") pod \"migrator-59844c95c7-brhqt\" (UID: \"a4b45170-1957-4ea8-b27e-e08dd68a56cd\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-brhqt" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005114 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8v9t\" (UniqueName: \"kubernetes.io/projected/baf64b43-1f21-4b84-ac70-3e0d879244a5-kube-api-access-v8v9t\") pod \"multus-admission-controller-857f4d67dd-7snv7\" (UID: \"baf64b43-1f21-4b84-ac70-3e0d879244a5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7snv7" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005133 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/80dca7ef-76c4-4616-9d8a-81ff066f6464-images\") pod \"machine-config-operator-74547568cd-2kfps\" (UID: \"80dca7ef-76c4-4616-9d8a-81ff066f6464\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005164 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbng5\" (UniqueName: \"kubernetes.io/projected/f692422b-da9f-4775-825e-ee198e5b61b8-kube-api-access-kbng5\") pod \"collect-profiles-29415285-28q92\" (UID: \"f692422b-da9f-4775-825e-ee198e5b61b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005180 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1ea90520-ebc6-41e2-9854-54caf800fda7-apiservice-cert\") pod \"packageserver-d55dfcdfc-rdtmj\" (UID: \"1ea90520-ebc6-41e2-9854-54caf800fda7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005200 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2be0da6d-6565-4336-89e0-590353e31685-config\") pod \"service-ca-operator-777779d784-7jtjl\" (UID: \"2be0da6d-6565-4336-89e0-590353e31685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7jtjl" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005218 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f915acae-6293-415a-a2a4-266340eee2e9-config\") pod \"kube-controller-manager-operator-78b949d7b-fws6m\" (UID: \"f915acae-6293-415a-a2a4-266340eee2e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fws6m" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005233 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/17fe9bd5-5973-47bd-945f-55bb8caa9d65-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k9twj\" (UID: \"17fe9bd5-5973-47bd-945f-55bb8caa9d65\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005249 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2t2h\" (UniqueName: \"kubernetes.io/projected/1247bce5-d4bf-4a8b-a438-3b150109ef5f-kube-api-access-q2t2h\") pod \"service-ca-9c57cc56f-vbxst\" (UID: \"1247bce5-d4bf-4a8b-a438-3b150109ef5f\") " pod="openshift-service-ca/service-ca-9c57cc56f-vbxst" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005264 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq96d\" (UniqueName: \"kubernetes.io/projected/2be0da6d-6565-4336-89e0-590353e31685-kube-api-access-gq96d\") pod \"service-ca-operator-777779d784-7jtjl\" (UID: \"2be0da6d-6565-4336-89e0-590353e31685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7jtjl" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005278 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cf990560-e430-4c2b-8965-c5ea18f8b695-srv-cert\") pod \"catalog-operator-68c6474976-d28rt\" (UID: \"cf990560-e430-4c2b-8965-c5ea18f8b695\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005297 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ab4d334f-b4a4-4d04-bf90-d2178f991f5f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zs9km\" (UID: \"ab4d334f-b4a4-4d04-bf90-d2178f991f5f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs9km" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005316 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c394687-7ae3-4500-bda2-a5de1cb7b8b1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4fgv6\" (UID: \"3c394687-7ae3-4500-bda2-a5de1cb7b8b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4fgv6" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005340 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-socket-dir\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005355 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/487642f9-0987-4953-b369-f092392d0209-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wl94z\" (UID: \"487642f9-0987-4953-b369-f092392d0209\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wl94z" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005374 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-plugins-dir\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005388 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2z8ws\" (UniqueName: \"kubernetes.io/projected/ab4d334f-b4a4-4d04-bf90-d2178f991f5f-kube-api-access-2z8ws\") pod \"control-plane-machine-set-operator-78cbb6b69f-zs9km\" (UID: \"ab4d334f-b4a4-4d04-bf90-d2178f991f5f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs9km" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005414 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3-config\") pod \"kube-apiserver-operator-766d6c64bb-f8k76\" (UID: \"33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f8k76" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005430 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17fe9bd5-5973-47bd-945f-55bb8caa9d65-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k9twj\" (UID: \"17fe9bd5-5973-47bd-945f-55bb8caa9d65\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005458 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1ea90520-ebc6-41e2-9854-54caf800fda7-tmpfs\") pod \"packageserver-d55dfcdfc-rdtmj\" (UID: \"1ea90520-ebc6-41e2-9854-54caf800fda7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005476 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6008a937-1ce8-48ba-b1eb-bf915c452468-node-bootstrap-token\") pod \"machine-config-server-swdhl\" (UID: \"6008a937-1ce8-48ba-b1eb-bf915c452468\") " pod="openshift-machine-config-operator/machine-config-server-swdhl" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005493 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jnjd\" (UniqueName: \"kubernetes.io/projected/e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6-kube-api-access-9jnjd\") pod \"olm-operator-6b444d44fb-8vkgq\" (UID: \"e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005511 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2be0da6d-6565-4336-89e0-590353e31685-serving-cert\") pod \"service-ca-operator-777779d784-7jtjl\" (UID: \"2be0da6d-6565-4336-89e0-590353e31685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7jtjl" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005527 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28rbk\" (UniqueName: \"kubernetes.io/projected/49b1c6db-0638-4367-8264-eec97250ad0a-kube-api-access-28rbk\") pod \"dns-default-zcnpm\" (UID: \"49b1c6db-0638-4367-8264-eec97250ad0a\") " pod="openshift-dns/dns-default-zcnpm" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005548 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgknv\" (UniqueName: \"kubernetes.io/projected/3c394687-7ae3-4500-bda2-a5de1cb7b8b1-kube-api-access-fgknv\") pod \"kube-storage-version-migrator-operator-b67b599dd-4fgv6\" (UID: \"3c394687-7ae3-4500-bda2-a5de1cb7b8b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4fgv6" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005564 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-csi-data-dir\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005581 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1db3923d-a8bf-4ed6-b30e-4bdf15471401-proxy-tls\") pod \"machine-config-controller-84d6567774-wsvhz\" (UID: \"1db3923d-a8bf-4ed6-b30e-4bdf15471401\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005596 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/49b1c6db-0638-4367-8264-eec97250ad0a-metrics-tls\") pod \"dns-default-zcnpm\" (UID: \"49b1c6db-0638-4367-8264-eec97250ad0a\") " pod="openshift-dns/dns-default-zcnpm" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005626 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz6q4\" (UniqueName: \"kubernetes.io/projected/80dca7ef-76c4-4616-9d8a-81ff066f6464-kube-api-access-dz6q4\") pod \"machine-config-operator-74547568cd-2kfps\" (UID: \"80dca7ef-76c4-4616-9d8a-81ff066f6464\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005644 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/49b1c6db-0638-4367-8264-eec97250ad0a-config-volume\") pod \"dns-default-zcnpm\" (UID: \"49b1c6db-0638-4367-8264-eec97250ad0a\") " pod="openshift-dns/dns-default-zcnpm" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005659 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6-srv-cert\") pod \"olm-operator-6b444d44fb-8vkgq\" (UID: \"e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005674 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f915acae-6293-415a-a2a4-266340eee2e9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fws6m\" (UID: \"f915acae-6293-415a-a2a4-266340eee2e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fws6m" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005690 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s4gz\" (UniqueName: \"kubernetes.io/projected/e0d60887-d7ee-4e14-96f1-81d1283ffa2e-kube-api-access-2s4gz\") pod \"package-server-manager-789f6589d5-s8cx6\" (UID: \"e0d60887-d7ee-4e14-96f1-81d1283ffa2e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-s8cx6" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005706 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-f8k76\" (UID: \"33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f8k76" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005721 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg89t\" (UniqueName: \"kubernetes.io/projected/6008a937-1ce8-48ba-b1eb-bf915c452468-kube-api-access-fg89t\") pod \"machine-config-server-swdhl\" (UID: \"6008a937-1ce8-48ba-b1eb-bf915c452468\") " pod="openshift-machine-config-operator/machine-config-server-swdhl" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005737 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0d60887-d7ee-4e14-96f1-81d1283ffa2e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-s8cx6\" (UID: \"e0d60887-d7ee-4e14-96f1-81d1283ffa2e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-s8cx6" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005753 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f692422b-da9f-4775-825e-ee198e5b61b8-config-volume\") pod \"collect-profiles-29415285-28q92\" (UID: \"f692422b-da9f-4775-825e-ee198e5b61b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005768 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/487642f9-0987-4953-b369-f092392d0209-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wl94z\" (UID: \"487642f9-0987-4953-b369-f092392d0209\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wl94z" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005785 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5pjp\" (UniqueName: \"kubernetes.io/projected/1ea90520-ebc6-41e2-9854-54caf800fda7-kube-api-access-c5pjp\") pod \"packageserver-d55dfcdfc-rdtmj\" (UID: \"1ea90520-ebc6-41e2-9854-54caf800fda7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005801 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkfz8\" (UniqueName: \"kubernetes.io/projected/cf990560-e430-4c2b-8965-c5ea18f8b695-kube-api-access-bkfz8\") pod \"catalog-operator-68c6474976-d28rt\" (UID: \"cf990560-e430-4c2b-8965-c5ea18f8b695\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005818 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1247bce5-d4bf-4a8b-a438-3b150109ef5f-signing-cabundle\") pod \"service-ca-9c57cc56f-vbxst\" (UID: \"1247bce5-d4bf-4a8b-a438-3b150109ef5f\") " pod="openshift-service-ca/service-ca-9c57cc56f-vbxst" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005832 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f692422b-da9f-4775-825e-ee198e5b61b8-secret-volume\") pod \"collect-profiles-29415285-28q92\" (UID: \"f692422b-da9f-4775-825e-ee198e5b61b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005847 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xkcb\" (UniqueName: \"kubernetes.io/projected/eba9ab59-079c-4512-b47c-b6c0e2a2886a-kube-api-access-9xkcb\") pod \"ingress-canary-hbt6g\" (UID: \"eba9ab59-079c-4512-b47c-b6c0e2a2886a\") " pod="openshift-ingress-canary/ingress-canary-hbt6g" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005862 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c394687-7ae3-4500-bda2-a5de1cb7b8b1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4fgv6\" (UID: \"3c394687-7ae3-4500-bda2-a5de1cb7b8b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4fgv6" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005883 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cf990560-e430-4c2b-8965-c5ea18f8b695-profile-collector-cert\") pod \"catalog-operator-68c6474976-d28rt\" (UID: \"cf990560-e430-4c2b-8965-c5ea18f8b695\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005928 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f915acae-6293-415a-a2a4-266340eee2e9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fws6m\" (UID: \"f915acae-6293-415a-a2a4-266340eee2e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fws6m" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005960 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/487642f9-0987-4953-b369-f092392d0209-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wl94z\" (UID: \"487642f9-0987-4953-b369-f092392d0209\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wl94z" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.005977 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwbbp\" (UniqueName: \"kubernetes.io/projected/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-kube-api-access-jwbbp\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.006005 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1db3923d-a8bf-4ed6-b30e-4bdf15471401-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wsvhz\" (UID: \"1db3923d-a8bf-4ed6-b30e-4bdf15471401\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.006019 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmrmp\" (UniqueName: \"kubernetes.io/projected/17fe9bd5-5973-47bd-945f-55bb8caa9d65-kube-api-access-jmrmp\") pod \"marketplace-operator-79b997595-k9twj\" (UID: \"17fe9bd5-5973-47bd-945f-55bb8caa9d65\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.006033 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1247bce5-d4bf-4a8b-a438-3b150109ef5f-signing-key\") pod \"service-ca-9c57cc56f-vbxst\" (UID: \"1247bce5-d4bf-4a8b-a438-3b150109ef5f\") " pod="openshift-service-ca/service-ca-9c57cc56f-vbxst" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.006048 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-mountpoint-dir\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.006063 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6008a937-1ce8-48ba-b1eb-bf915c452468-certs\") pod \"machine-config-server-swdhl\" (UID: \"6008a937-1ce8-48ba-b1eb-bf915c452468\") " pod="openshift-machine-config-operator/machine-config-server-swdhl" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.006079 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-registration-dir\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.006093 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6-profile-collector-cert\") pod \"olm-operator-6b444d44fb-8vkgq\" (UID: \"e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.006110 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4h54\" (UniqueName: \"kubernetes.io/projected/1db3923d-a8bf-4ed6-b30e-4bdf15471401-kube-api-access-q4h54\") pod \"machine-config-controller-84d6567774-wsvhz\" (UID: \"1db3923d-a8bf-4ed6-b30e-4bdf15471401\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.006135 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/80dca7ef-76c4-4616-9d8a-81ff066f6464-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2kfps\" (UID: \"80dca7ef-76c4-4616-9d8a-81ff066f6464\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.006150 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eba9ab59-079c-4512-b47c-b6c0e2a2886a-cert\") pod \"ingress-canary-hbt6g\" (UID: \"eba9ab59-079c-4512-b47c-b6c0e2a2886a\") " pod="openshift-ingress-canary/ingress-canary-hbt6g" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.006882 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-csi-data-dir\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:27 crc kubenswrapper[4997]: E1205 06:57:27.007014 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:27.506991431 +0000 UTC m=+148.035898782 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.009390 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/487642f9-0987-4953-b369-f092392d0209-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wl94z\" (UID: \"487642f9-0987-4953-b369-f092392d0209\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wl94z" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.012817 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0dc484d1-5411-48aa-81d8-c208d0c5e2a6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n5wgm\" (UID: \"0dc484d1-5411-48aa-81d8-c208d0c5e2a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.013136 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-registration-dir\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.013425 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-f8k76\" (UID: \"33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f8k76" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.013452 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/baf64b43-1f21-4b84-ac70-3e0d879244a5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7snv7\" (UID: \"baf64b43-1f21-4b84-ac70-3e0d879244a5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7snv7" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.013639 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-mountpoint-dir\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.014855 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f692422b-da9f-4775-825e-ee198e5b61b8-config-volume\") pod \"collect-profiles-29415285-28q92\" (UID: \"f692422b-da9f-4775-825e-ee198e5b61b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.017532 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1247bce5-d4bf-4a8b-a438-3b150109ef5f-signing-cabundle\") pod \"service-ca-9c57cc56f-vbxst\" (UID: \"1247bce5-d4bf-4a8b-a438-3b150109ef5f\") " pod="openshift-service-ca/service-ca-9c57cc56f-vbxst" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.017571 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1db3923d-a8bf-4ed6-b30e-4bdf15471401-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wsvhz\" (UID: \"1db3923d-a8bf-4ed6-b30e-4bdf15471401\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.018523 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2be0da6d-6565-4336-89e0-590353e31685-config\") pod \"service-ca-operator-777779d784-7jtjl\" (UID: \"2be0da6d-6565-4336-89e0-590353e31685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7jtjl" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.019253 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/80dca7ef-76c4-4616-9d8a-81ff066f6464-images\") pod \"machine-config-operator-74547568cd-2kfps\" (UID: \"80dca7ef-76c4-4616-9d8a-81ff066f6464\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.020289 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c394687-7ae3-4500-bda2-a5de1cb7b8b1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4fgv6\" (UID: \"3c394687-7ae3-4500-bda2-a5de1cb7b8b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4fgv6" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.021298 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f915acae-6293-415a-a2a4-266340eee2e9-config\") pod \"kube-controller-manager-operator-78b949d7b-fws6m\" (UID: \"f915acae-6293-415a-a2a4-266340eee2e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fws6m" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.022731 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c394687-7ae3-4500-bda2-a5de1cb7b8b1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4fgv6\" (UID: \"3c394687-7ae3-4500-bda2-a5de1cb7b8b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4fgv6" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.023286 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f692422b-da9f-4775-825e-ee198e5b61b8-secret-volume\") pod \"collect-profiles-29415285-28q92\" (UID: \"f692422b-da9f-4775-825e-ee198e5b61b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.024173 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6-profile-collector-cert\") pod \"olm-operator-6b444d44fb-8vkgq\" (UID: \"e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.024848 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17fe9bd5-5973-47bd-945f-55bb8caa9d65-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k9twj\" (UID: \"17fe9bd5-5973-47bd-945f-55bb8caa9d65\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.025012 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-socket-dir\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.025312 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/80dca7ef-76c4-4616-9d8a-81ff066f6464-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2kfps\" (UID: \"80dca7ef-76c4-4616-9d8a-81ff066f6464\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" Dec 05 06:57:27 crc kubenswrapper[4997]: W1205 06:57:27.025583 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda412ad65_d072_433a_8d6f_7e5390195f89.slice/crio-75211d00b7ee261b58a1f03075a25f2f030695d8d9afbd33fdafb70a239de409 WatchSource:0}: Error finding container 75211d00b7ee261b58a1f03075a25f2f030695d8d9afbd33fdafb70a239de409: Status 404 returned error can't find the container with id 75211d00b7ee261b58a1f03075a25f2f030695d8d9afbd33fdafb70a239de409 Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.025829 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/49b1c6db-0638-4367-8264-eec97250ad0a-config-volume\") pod \"dns-default-zcnpm\" (UID: \"49b1c6db-0638-4367-8264-eec97250ad0a\") " pod="openshift-dns/dns-default-zcnpm" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.025937 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-plugins-dir\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.026533 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3-config\") pod \"kube-apiserver-operator-766d6c64bb-f8k76\" (UID: \"33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f8k76" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.027056 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1ea90520-ebc6-41e2-9854-54caf800fda7-tmpfs\") pod \"packageserver-d55dfcdfc-rdtmj\" (UID: \"1ea90520-ebc6-41e2-9854-54caf800fda7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.027059 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6008a937-1ce8-48ba-b1eb-bf915c452468-certs\") pod \"machine-config-server-swdhl\" (UID: \"6008a937-1ce8-48ba-b1eb-bf915c452468\") " pod="openshift-machine-config-operator/machine-config-server-swdhl" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.027348 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/49b1c6db-0638-4367-8264-eec97250ad0a-metrics-tls\") pod \"dns-default-zcnpm\" (UID: \"49b1c6db-0638-4367-8264-eec97250ad0a\") " pod="openshift-dns/dns-default-zcnpm" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.027788 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cf990560-e430-4c2b-8965-c5ea18f8b695-profile-collector-cert\") pod \"catalog-operator-68c6474976-d28rt\" (UID: \"cf990560-e430-4c2b-8965-c5ea18f8b695\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.027905 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1db3923d-a8bf-4ed6-b30e-4bdf15471401-proxy-tls\") pod \"machine-config-controller-84d6567774-wsvhz\" (UID: \"1db3923d-a8bf-4ed6-b30e-4bdf15471401\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.027921 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1ea90520-ebc6-41e2-9854-54caf800fda7-webhook-cert\") pod \"packageserver-d55dfcdfc-rdtmj\" (UID: \"1ea90520-ebc6-41e2-9854-54caf800fda7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.028295 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1ea90520-ebc6-41e2-9854-54caf800fda7-apiservice-cert\") pod \"packageserver-d55dfcdfc-rdtmj\" (UID: \"1ea90520-ebc6-41e2-9854-54caf800fda7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.029293 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1247bce5-d4bf-4a8b-a438-3b150109ef5f-signing-key\") pod \"service-ca-9c57cc56f-vbxst\" (UID: \"1247bce5-d4bf-4a8b-a438-3b150109ef5f\") " pod="openshift-service-ca/service-ca-9c57cc56f-vbxst" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.029736 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/80dca7ef-76c4-4616-9d8a-81ff066f6464-proxy-tls\") pod \"machine-config-operator-74547568cd-2kfps\" (UID: \"80dca7ef-76c4-4616-9d8a-81ff066f6464\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.029795 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/487642f9-0987-4953-b369-f092392d0209-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wl94z\" (UID: \"487642f9-0987-4953-b369-f092392d0209\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wl94z" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.031175 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eba9ab59-079c-4512-b47c-b6c0e2a2886a-cert\") pod \"ingress-canary-hbt6g\" (UID: \"eba9ab59-079c-4512-b47c-b6c0e2a2886a\") " pod="openshift-ingress-canary/ingress-canary-hbt6g" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.032197 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6008a937-1ce8-48ba-b1eb-bf915c452468-node-bootstrap-token\") pod \"machine-config-server-swdhl\" (UID: \"6008a937-1ce8-48ba-b1eb-bf915c452468\") " pod="openshift-machine-config-operator/machine-config-server-swdhl" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.032871 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0d60887-d7ee-4e14-96f1-81d1283ffa2e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-s8cx6\" (UID: \"e0d60887-d7ee-4e14-96f1-81d1283ffa2e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-s8cx6" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.033545 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cf990560-e430-4c2b-8965-c5ea18f8b695-srv-cert\") pod \"catalog-operator-68c6474976-d28rt\" (UID: \"cf990560-e430-4c2b-8965-c5ea18f8b695\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.034143 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ab4d334f-b4a4-4d04-bf90-d2178f991f5f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zs9km\" (UID: \"ab4d334f-b4a4-4d04-bf90-d2178f991f5f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs9km" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.035818 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6-srv-cert\") pod \"olm-operator-6b444d44fb-8vkgq\" (UID: \"e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.038713 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2be0da6d-6565-4336-89e0-590353e31685-serving-cert\") pod \"service-ca-operator-777779d784-7jtjl\" (UID: \"2be0da6d-6565-4336-89e0-590353e31685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7jtjl" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.042286 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr7gj\" (UniqueName: \"kubernetes.io/projected/65e6a96e-df08-432d-8319-93e3e483f126-kube-api-access-zr7gj\") pod \"etcd-operator-b45778765-5msxl\" (UID: \"65e6a96e-df08-432d-8319-93e3e483f126\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.042289 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/17fe9bd5-5973-47bd-945f-55bb8caa9d65-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k9twj\" (UID: \"17fe9bd5-5973-47bd-945f-55bb8caa9d65\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.045727 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f915acae-6293-415a-a2a4-266340eee2e9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fws6m\" (UID: \"f915acae-6293-415a-a2a4-266340eee2e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fws6m" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.050438 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkrch\" (UniqueName: \"kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-kube-api-access-vkrch\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.054409 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qj4lx" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.055215 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9b7r\" (UniqueName: \"kubernetes.io/projected/ff56ef31-a073-49ac-b00a-e9611e2e0025-kube-api-access-c9b7r\") pod \"cluster-image-registry-operator-dc59b4c8b-7pzd7\" (UID: \"ff56ef31-a073-49ac-b00a-e9611e2e0025\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.069136 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.088256 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.096648 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ff56ef31-a073-49ac-b00a-e9611e2e0025-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-7pzd7\" (UID: \"ff56ef31-a073-49ac-b00a-e9611e2e0025\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.103352 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd"] Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.107473 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:27 crc kubenswrapper[4997]: E1205 06:57:27.108071 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:27.608053996 +0000 UTC m=+148.136961257 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.112440 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5b8mm"] Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.116525 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz8xx\" (UniqueName: \"kubernetes.io/projected/560926cd-d0fe-4823-b217-803ba02d2702-kube-api-access-bz8xx\") pod \"router-default-5444994796-v4mxn\" (UID: \"560926cd-d0fe-4823-b217-803ba02d2702\") " pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.136254 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wpl2k"] Dec 05 06:57:27 crc kubenswrapper[4997]: W1205 06:57:27.146869 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7face06_9e66_4375_8b67_974b8936289f.slice/crio-5bffc598a91716c44b506f0c1346e69e50df865edcb890322e77a9b9abd96045 WatchSource:0}: Error finding container 5bffc598a91716c44b506f0c1346e69e50df865edcb890322e77a9b9abd96045: Status 404 returned error can't find the container with id 5bffc598a91716c44b506f0c1346e69e50df865edcb890322e77a9b9abd96045 Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.148261 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg89t\" (UniqueName: \"kubernetes.io/projected/6008a937-1ce8-48ba-b1eb-bf915c452468-kube-api-access-fg89t\") pod \"machine-config-server-swdhl\" (UID: \"6008a937-1ce8-48ba-b1eb-bf915c452468\") " pod="openshift-machine-config-operator/machine-config-server-swdhl" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.155255 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh"] Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.164504 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmrmp\" (UniqueName: \"kubernetes.io/projected/17fe9bd5-5973-47bd-945f-55bb8caa9d65-kube-api-access-jmrmp\") pod \"marketplace-operator-79b997595-k9twj\" (UID: \"17fe9bd5-5973-47bd-945f-55bb8caa9d65\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.169316 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9k5xc"] Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.180959 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xkcb\" (UniqueName: \"kubernetes.io/projected/eba9ab59-079c-4512-b47c-b6c0e2a2886a-kube-api-access-9xkcb\") pod \"ingress-canary-hbt6g\" (UID: \"eba9ab59-079c-4512-b47c-b6c0e2a2886a\") " pod="openshift-ingress-canary/ingress-canary-hbt6g" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.196436 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/487642f9-0987-4953-b369-f092392d0209-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wl94z\" (UID: \"487642f9-0987-4953-b369-f092392d0209\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wl94z" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.209057 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:27 crc kubenswrapper[4997]: E1205 06:57:27.209482 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:27.709450591 +0000 UTC m=+148.238357852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.209812 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:27 crc kubenswrapper[4997]: E1205 06:57:27.210091 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:27.710084867 +0000 UTC m=+148.238992128 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.219182 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hbt6g" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.220411 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwbbp\" (UniqueName: \"kubernetes.io/projected/ff590dc9-b474-4ae3-a66d-ea8f6e6797d0-kube-api-access-jwbbp\") pod \"csi-hostpathplugin-pjknx\" (UID: \"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0\") " pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.224988 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zm2tt"] Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.252423 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-f8k76\" (UID: \"33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f8k76" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.256317 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-swdhl" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.263946 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f915acae-6293-415a-a2a4-266340eee2e9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fws6m\" (UID: \"f915acae-6293-415a-a2a4-266340eee2e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fws6m" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.278533 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68bfb\" (UniqueName: \"kubernetes.io/projected/a4b45170-1957-4ea8-b27e-e08dd68a56cd-kube-api-access-68bfb\") pod \"migrator-59844c95c7-brhqt\" (UID: \"a4b45170-1957-4ea8-b27e-e08dd68a56cd\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-brhqt" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.284198 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-pjknx" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.299587 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8v9t\" (UniqueName: \"kubernetes.io/projected/baf64b43-1f21-4b84-ac70-3e0d879244a5-kube-api-access-v8v9t\") pod \"multus-admission-controller-857f4d67dd-7snv7\" (UID: \"baf64b43-1f21-4b84-ac70-3e0d879244a5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7snv7" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.310854 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:27 crc kubenswrapper[4997]: E1205 06:57:27.311382 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:27.811350968 +0000 UTC m=+148.340258229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.311513 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:27 crc kubenswrapper[4997]: E1205 06:57:27.311930 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:27.811920753 +0000 UTC m=+148.340828014 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.324887 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-jktft"] Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.331539 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbng5\" (UniqueName: \"kubernetes.io/projected/f692422b-da9f-4775-825e-ee198e5b61b8-kube-api-access-kbng5\") pod \"collect-profiles-29415285-28q92\" (UID: \"f692422b-da9f-4775-825e-ee198e5b61b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.342800 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s4gz\" (UniqueName: \"kubernetes.io/projected/e0d60887-d7ee-4e14-96f1-81d1283ffa2e-kube-api-access-2s4gz\") pod \"package-server-manager-789f6589d5-s8cx6\" (UID: \"e0d60887-d7ee-4e14-96f1-81d1283ffa2e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-s8cx6" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.357553 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz6q4\" (UniqueName: \"kubernetes.io/projected/80dca7ef-76c4-4616-9d8a-81ff066f6464-kube-api-access-dz6q4\") pod \"machine-config-operator-74547568cd-2kfps\" (UID: \"80dca7ef-76c4-4616-9d8a-81ff066f6464\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.364215 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.373147 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4h54\" (UniqueName: \"kubernetes.io/projected/1db3923d-a8bf-4ed6-b30e-4bdf15471401-kube-api-access-q4h54\") pod \"machine-config-controller-84d6567774-wsvhz\" (UID: \"1db3923d-a8bf-4ed6-b30e-4bdf15471401\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.386815 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.391429 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5pjp\" (UniqueName: \"kubernetes.io/projected/1ea90520-ebc6-41e2-9854-54caf800fda7-kube-api-access-c5pjp\") pod \"packageserver-d55dfcdfc-rdtmj\" (UID: \"1ea90520-ebc6-41e2-9854-54caf800fda7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.393056 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.407860 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f8k76" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.414467 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wl94z" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.414508 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:27 crc kubenswrapper[4997]: E1205 06:57:27.414860 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:27.914843957 +0000 UTC m=+148.443751218 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.418061 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2z8ws\" (UniqueName: \"kubernetes.io/projected/ab4d334f-b4a4-4d04-bf90-d2178f991f5f-kube-api-access-2z8ws\") pod \"control-plane-machine-set-operator-78cbb6b69f-zs9km\" (UID: \"ab4d334f-b4a4-4d04-bf90-d2178f991f5f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs9km" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.421485 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-67z92"] Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.421916 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-brhqt" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.429235 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qd6rv"] Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.433657 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jnjd\" (UniqueName: \"kubernetes.io/projected/e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6-kube-api-access-9jnjd\") pod \"olm-operator-6b444d44fb-8vkgq\" (UID: \"e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.437662 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fws6m" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.444892 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.452472 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.461353 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-s8cx6" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.466167 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28rbk\" (UniqueName: \"kubernetes.io/projected/49b1c6db-0638-4367-8264-eec97250ad0a-kube-api-access-28rbk\") pod \"dns-default-zcnpm\" (UID: \"49b1c6db-0638-4367-8264-eec97250ad0a\") " pod="openshift-dns/dns-default-zcnpm" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.469124 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7snv7" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.472208 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgknv\" (UniqueName: \"kubernetes.io/projected/3c394687-7ae3-4500-bda2-a5de1cb7b8b1-kube-api-access-fgknv\") pod \"kube-storage-version-migrator-operator-b67b599dd-4fgv6\" (UID: \"3c394687-7ae3-4500-bda2-a5de1cb7b8b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4fgv6" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.474671 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.492583 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq96d\" (UniqueName: \"kubernetes.io/projected/2be0da6d-6565-4336-89e0-590353e31685-kube-api-access-gq96d\") pod \"service-ca-operator-777779d784-7jtjl\" (UID: \"2be0da6d-6565-4336-89e0-590353e31685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7jtjl" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.499397 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7jtjl" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.507244 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs9km" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.516041 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:27 crc kubenswrapper[4997]: E1205 06:57:27.516534 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:28.016510719 +0000 UTC m=+148.545418190 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.522113 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2t2h\" (UniqueName: \"kubernetes.io/projected/1247bce5-d4bf-4a8b-a438-3b150109ef5f-kube-api-access-q2t2h\") pod \"service-ca-9c57cc56f-vbxst\" (UID: \"1247bce5-d4bf-4a8b-a438-3b150109ef5f\") " pod="openshift-service-ca/service-ca-9c57cc56f-vbxst" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.526959 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.536082 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-vbxst" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.539929 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkfz8\" (UniqueName: \"kubernetes.io/projected/cf990560-e430-4c2b-8965-c5ea18f8b695-kube-api-access-bkfz8\") pod \"catalog-operator-68c6474976-d28rt\" (UID: \"cf990560-e430-4c2b-8965-c5ea18f8b695\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.548408 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zcnpm" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.587350 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qj4lx"] Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.587403 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm"] Dec 05 06:57:27 crc kubenswrapper[4997]: W1205 06:57:27.592028 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6008a937_1ce8_48ba_b1eb_bf915c452468.slice/crio-de586b383e294ccf4e56e203f0d967b4eb87628e05513b7da010dbb411f1a7db WatchSource:0}: Error finding container de586b383e294ccf4e56e203f0d967b4eb87628e05513b7da010dbb411f1a7db: Status 404 returned error can't find the container with id de586b383e294ccf4e56e203f0d967b4eb87628e05513b7da010dbb411f1a7db Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.613803 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" event={"ID":"8c6d7e0f-f064-48a0-b05b-287b09867b45","Type":"ContainerStarted","Data":"b85d119d65b7e939f9f8474bd9d74580f4a5dbeb0804fd631225971d0373e0e8"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.615744 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" event={"ID":"8c6d7e0f-f064-48a0-b05b-287b09867b45","Type":"ContainerStarted","Data":"4ef87092f95f9c56dd6a11c4435aa3b230a6be4ca438bf5a6e192de12dba437b"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.618034 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:27 crc kubenswrapper[4997]: E1205 06:57:27.621171 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:28.121146338 +0000 UTC m=+148.650053599 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.637951 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zxx6b" event={"ID":"18378604-4db4-4da8-94c4-c52452ee6805","Type":"ContainerStarted","Data":"37426104dfec6408eab6676822be99e343fec4b04aadc564f2927f92e9dbd074"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.638005 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zxx6b" event={"ID":"18378604-4db4-4da8-94c4-c52452ee6805","Type":"ContainerStarted","Data":"dddacfc348f841e93aba07b171a73a79ebc535268c54cd6a6d75ed0cadfc00fb"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.641594 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qz7ht" event={"ID":"f4c37dd5-377e-4ed4-99f5-b179110153af","Type":"ContainerStarted","Data":"2a1efbc8561d728bf70dd212306c2c8f4fd4b67e72900e9a89bcffa6a069e42e"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.641660 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qz7ht" event={"ID":"f4c37dd5-377e-4ed4-99f5-b179110153af","Type":"ContainerStarted","Data":"f0fb61acd59beff6203d6dbc96930fc2d7bb756cf72637a1be58a15a775719f0"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.642944 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-qz7ht" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.667520 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" event={"ID":"0a38ab13-190d-40c5-9a69-389e2fae6884","Type":"ContainerStarted","Data":"5ebbff2803e4c29ed107d52e30a61afd3238b1e455a3b09537e671087018c3f3"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.682685 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" event={"ID":"e6a372ba-59d9-4fcc-ba21-7b9de5316239","Type":"ContainerStarted","Data":"e185ef630ce8f8eb0f0eafee24a427eab0bff85e1965dff9023bab13a95480f3"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.682729 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.682740 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" event={"ID":"e6a372ba-59d9-4fcc-ba21-7b9de5316239","Type":"ContainerStarted","Data":"e6ef3c4bbe66efeeea728677ebda5fc005e6b7cb8728228bcc91270d9605c7f2"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.687895 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-67z92" event={"ID":"0836455d-d750-4a39-a7cc-6e95fe3e0abd","Type":"ContainerStarted","Data":"100a523d2eafc410d7532d8900dc68b1a09009b8d0e12b984c91d8423c574982"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.695534 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" event={"ID":"821da4d3-7e2a-480e-8290-a7080b916b08","Type":"ContainerStarted","Data":"5887873529b36f29ac1bc24e95c8fa39c77de4760f0bca7677e171271e7f7452"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.697855 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.700139 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4fgv6" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.701810 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9k5xc" event={"ID":"800c31da-5f11-4029-a43f-ad96ac33b062","Type":"ContainerStarted","Data":"1d4079a9f24e10294dad71fbf6accc77975d8b7c515812ee8b97bf6e34f888ff"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.723496 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5msxl"] Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.724123 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:27 crc kubenswrapper[4997]: E1205 06:57:27.724405 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:28.224393261 +0000 UTC m=+148.753300522 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.728108 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.739360 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qd6rv" event={"ID":"4c78711c-d17c-4b1c-b00e-f8ac8972f1d9","Type":"ContainerStarted","Data":"0b5d9504b3bea0ccdcdec64bc6990aad58f5986d6a73c37fad6819c000bda566"} Dec 05 06:57:27 crc kubenswrapper[4997]: W1205 06:57:27.739981 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0dc484d1_5411_48aa_81d8_c208d0c5e2a6.slice/crio-4819ace87e99826ca6807fe2db4e9ffad4d2664af420f4276f5a30a08e054b14 WatchSource:0}: Error finding container 4819ace87e99826ca6807fe2db4e9ffad4d2664af420f4276f5a30a08e054b14: Status 404 returned error can't find the container with id 4819ace87e99826ca6807fe2db4e9ffad4d2664af420f4276f5a30a08e054b14 Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.784740 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.798386 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" event={"ID":"c7face06-9e66-4375-8b67-974b8936289f","Type":"ContainerStarted","Data":"5bffc598a91716c44b506f0c1346e69e50df865edcb890322e77a9b9abd96045"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.808987 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-lskhh" event={"ID":"0160154e-35ae-471c-b848-214b9f33ec93","Type":"ContainerStarted","Data":"beb6b2ddc048d904d5272115b463f6cb94839a1053b7f034b77aaa415d342db0"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.809019 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-lskhh" event={"ID":"0160154e-35ae-471c-b848-214b9f33ec93","Type":"ContainerStarted","Data":"dca35c09a300c42e2a141902bd5ba976fceec3765d300b85974ab61bc38f1783"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.811923 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" event={"ID":"e9370d84-0bfb-4e61-a5b7-bcb43e903593","Type":"ContainerStarted","Data":"74480444f454366411c6240f077a7a7625f9a9e01e972dc8c892b19eb86f3e30"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.817049 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jktft" event={"ID":"bdc72754-1951-4312-9dd0-3458357f2caa","Type":"ContainerStarted","Data":"cd2ad9fba48f3f40aa2f270e5dde9ada72d0b18b41d56887fe886aa74dad3ea3"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.818411 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" event={"ID":"f3993123-571c-4be4-a525-d8e0481b1aa5","Type":"ContainerStarted","Data":"c79e0a00c6184219eda9b9b8295323e2df26c62c0c033c6405f7de24c89b6012"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.818435 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" event={"ID":"f3993123-571c-4be4-a525-d8e0481b1aa5","Type":"ContainerStarted","Data":"4f4c4d6b928c633bdbd5126afc64a5029f240f3c790bc34241e79b3338a13bef"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.826706 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.828909 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" event={"ID":"a412ad65-d072-433a-8d6f-7e5390195f89","Type":"ContainerStarted","Data":"75211d00b7ee261b58a1f03075a25f2f030695d8d9afbd33fdafb70a239de409"} Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.828953 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:27 crc kubenswrapper[4997]: E1205 06:57:27.829019 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:28.32900305 +0000 UTC m=+148.857910311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.830257 4997 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-wxs8l container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" start-of-body= Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.830291 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" podUID="a412ad65-d072-433a-8d6f-7e5390195f89" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.935490 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:27 crc kubenswrapper[4997]: E1205 06:57:27.936503 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:28.436487004 +0000 UTC m=+148.965394265 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:27 crc kubenswrapper[4997]: I1205 06:57:27.961842 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7"] Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.010468 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hbt6g"] Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.043200 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:28 crc kubenswrapper[4997]: E1205 06:57:28.043576 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:28.543555337 +0000 UTC m=+149.072462608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.067537 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-qz7ht" Dec 05 06:57:28 crc kubenswrapper[4997]: W1205 06:57:28.067650 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod560926cd_d0fe_4823_b217_803ba02d2702.slice/crio-9f601a7e567982d748b3b94e84342a436a41b62799359bb5297d7f9211ca0776 WatchSource:0}: Error finding container 9f601a7e567982d748b3b94e84342a436a41b62799359bb5297d7f9211ca0776: Status 404 returned error can't find the container with id 9f601a7e567982d748b3b94e84342a436a41b62799359bb5297d7f9211ca0776 Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.145282 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:28 crc kubenswrapper[4997]: E1205 06:57:28.145768 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:28.645753032 +0000 UTC m=+149.174660293 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.234815 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj"] Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.247397 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:28 crc kubenswrapper[4997]: E1205 06:57:28.247675 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:28.74765878 +0000 UTC m=+149.276566031 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.351308 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:28 crc kubenswrapper[4997]: E1205 06:57:28.351670 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:28.851656142 +0000 UTC m=+149.380563403 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.452953 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:28 crc kubenswrapper[4997]: E1205 06:57:28.453499 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:28.953478288 +0000 UTC m=+149.482385549 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.554293 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:28 crc kubenswrapper[4997]: E1205 06:57:28.554751 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:29.054734868 +0000 UTC m=+149.583642129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.655495 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:28 crc kubenswrapper[4997]: E1205 06:57:28.655820 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:29.155804234 +0000 UTC m=+149.684711485 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.661457 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wl94z"] Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.691882 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pjknx"] Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.765600 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:28 crc kubenswrapper[4997]: E1205 06:57:28.765958 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:29.265945538 +0000 UTC m=+149.794852799 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.830785 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7" event={"ID":"ff56ef31-a073-49ac-b00a-e9611e2e0025","Type":"ContainerStarted","Data":"3b87d85f8d64b056972ed53d83ef834f788d2d4b2a84bcaefe3d713385fbab8f"} Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.834947 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-v4mxn" event={"ID":"560926cd-d0fe-4823-b217-803ba02d2702","Type":"ContainerStarted","Data":"9f601a7e567982d748b3b94e84342a436a41b62799359bb5297d7f9211ca0776"} Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.841695 4997 generic.go:334] "Generic (PLEG): container finished" podID="821da4d3-7e2a-480e-8290-a7080b916b08" containerID="ab99d7914a258ee8df95a602c437a6b5bd25404d371fc627f19773fd14b72e40" exitCode=0 Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.841741 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" event={"ID":"821da4d3-7e2a-480e-8290-a7080b916b08","Type":"ContainerDied","Data":"ab99d7914a258ee8df95a602c437a6b5bd25404d371fc627f19773fd14b72e40"} Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.867875 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" event={"ID":"f3993123-571c-4be4-a525-d8e0481b1aa5","Type":"ContainerStarted","Data":"7a5825b984564b381848696d22bab010bd318e24d16d976ce44bcf55ec260010"} Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.868243 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:28 crc kubenswrapper[4997]: E1205 06:57:28.868379 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:29.368350508 +0000 UTC m=+149.897257769 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.868577 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:28 crc kubenswrapper[4997]: E1205 06:57:28.874042 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:29.374020647 +0000 UTC m=+149.902927908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.882976 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" event={"ID":"0dc484d1-5411-48aa-81d8-c208d0c5e2a6","Type":"ContainerStarted","Data":"4819ace87e99826ca6807fe2db4e9ffad4d2664af420f4276f5a30a08e054b14"} Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.890871 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" event={"ID":"1ea90520-ebc6-41e2-9854-54caf800fda7","Type":"ContainerStarted","Data":"7e8c418f43b707952b09680c4c0fba1d848cff79a771d527cfe9d5acf4a33352"} Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.915914 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qd6rv" event={"ID":"4c78711c-d17c-4b1c-b00e-f8ac8972f1d9","Type":"ContainerStarted","Data":"7e8be08e04df66c107e9b592ad39d1f310c11816986252102dcadf4562d2dee5"} Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.927689 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k9twj"] Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.970542 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" podStartSLOduration=130.970520053 podStartE2EDuration="2m10.970520053s" podCreationTimestamp="2025-12-05 06:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:28.938897543 +0000 UTC m=+149.467804814" watchObservedRunningTime="2025-12-05 06:57:28.970520053 +0000 UTC m=+149.499427314" Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.971081 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-qz7ht" podStartSLOduration=130.971076368 podStartE2EDuration="2m10.971076368s" podCreationTimestamp="2025-12-05 06:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:28.970656286 +0000 UTC m=+149.499563557" watchObservedRunningTime="2025-12-05 06:57:28.971076368 +0000 UTC m=+149.499983629" Dec 05 06:57:28 crc kubenswrapper[4997]: I1205 06:57:28.971472 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:28 crc kubenswrapper[4997]: E1205 06:57:28.972515 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:29.472499105 +0000 UTC m=+150.001406366 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.002546 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9k5xc" event={"ID":"800c31da-5f11-4029-a43f-ad96ac33b062","Type":"ContainerStarted","Data":"fe8c5d596a0a1c6aaaa17f3cde4a19b65d8850e680ea11d530c06ab8ec77d7e3"} Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.054994 4997 generic.go:334] "Generic (PLEG): container finished" podID="c7face06-9e66-4375-8b67-974b8936289f" containerID="320334c9f51c76140ba2c272242c5839b49bb43276fbfc78581143e63cd5992c" exitCode=0 Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.055126 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" event={"ID":"c7face06-9e66-4375-8b67-974b8936289f","Type":"ContainerDied","Data":"320334c9f51c76140ba2c272242c5839b49bb43276fbfc78581143e63cd5992c"} Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.058760 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-swdhl" event={"ID":"6008a937-1ce8-48ba-b1eb-bf915c452468","Type":"ContainerStarted","Data":"de586b383e294ccf4e56e203f0d967b4eb87628e05513b7da010dbb411f1a7db"} Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.064308 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" event={"ID":"0a38ab13-190d-40c5-9a69-389e2fae6884","Type":"ContainerStarted","Data":"a5fe8cbc600bdc6d50b247ce9eeb3580052821f375efb95eb0d296e7169ffde4"} Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.065761 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.072838 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:29 crc kubenswrapper[4997]: E1205 06:57:29.075312 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:29.575291486 +0000 UTC m=+150.104198747 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.075413 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" event={"ID":"65e6a96e-df08-432d-8319-93e3e483f126","Type":"ContainerStarted","Data":"a3b9faf939b5fb0a19f1591005a5c4b9b45c544b77f2ba965c1d886592aa336e"} Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.082394 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" podStartSLOduration=130.082379522 podStartE2EDuration="2m10.082379522s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:29.080098073 +0000 UTC m=+149.609005344" watchObservedRunningTime="2025-12-05 06:57:29.082379522 +0000 UTC m=+149.611286783" Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.108720 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-s8cx6"] Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.114819 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7snv7"] Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.118118 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hbt6g" event={"ID":"eba9ab59-079c-4512-b47c-b6c0e2a2886a","Type":"ContainerStarted","Data":"5a265c6518fcb62c4d2023edfe5938fb475f473d40446172631e090849c72f28"} Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.122818 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f8k76"] Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.123640 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-lskhh" podStartSLOduration=131.123605366 podStartE2EDuration="2m11.123605366s" podCreationTimestamp="2025-12-05 06:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:29.120040682 +0000 UTC m=+149.648947943" watchObservedRunningTime="2025-12-05 06:57:29.123605366 +0000 UTC m=+149.652512647" Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.126940 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" event={"ID":"a412ad65-d072-433a-8d6f-7e5390195f89","Type":"ContainerStarted","Data":"d894fdd8ca69d1463d83abe8e556551cbff7daa975da4022fb133542d1312cd2"} Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.131065 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wl94z" event={"ID":"487642f9-0987-4953-b369-f092392d0209","Type":"ContainerStarted","Data":"6d92440c05825e5fbe6ff03d64d29066176db48a3cc36b121e310b5735004b3b"} Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.154284 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qj4lx" event={"ID":"df171b06-9a20-42ed-8caa-c9f8aebd611c","Type":"ContainerStarted","Data":"be3ee804bbfcfb6f76bdc70566f3f9084c7320137f87bf41eb38ec52f3b26748"} Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.154336 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qj4lx" event={"ID":"df171b06-9a20-42ed-8caa-c9f8aebd611c","Type":"ContainerStarted","Data":"89f54af258cf68bb5a35cd0dbd0830dc2b9b3e247940056e4303acccf9744113"} Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.174057 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:29 crc kubenswrapper[4997]: E1205 06:57:29.174402 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:29.67438477 +0000 UTC m=+150.203292031 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.275746 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.280570 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" podStartSLOduration=130.28055058 podStartE2EDuration="2m10.28055058s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:29.279596664 +0000 UTC m=+149.808503925" watchObservedRunningTime="2025-12-05 06:57:29.28055058 +0000 UTC m=+149.809457841" Dec 05 06:57:29 crc kubenswrapper[4997]: E1205 06:57:29.285580 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:29.785557601 +0000 UTC m=+150.314464862 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.318725 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" podStartSLOduration=130.318699432 podStartE2EDuration="2m10.318699432s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:29.318007444 +0000 UTC m=+149.846914725" watchObservedRunningTime="2025-12-05 06:57:29.318699432 +0000 UTC m=+149.847606683" Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.357946 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-wpl2k" podStartSLOduration=130.357924532 podStartE2EDuration="2m10.357924532s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:29.357060679 +0000 UTC m=+149.885967940" watchObservedRunningTime="2025-12-05 06:57:29.357924532 +0000 UTC m=+149.886831793" Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.369630 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.377319 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:29 crc kubenswrapper[4997]: E1205 06:57:29.377708 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:29.877686071 +0000 UTC m=+150.406593332 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.377851 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.417822 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qj4lx" podStartSLOduration=130.417798116 podStartE2EDuration="2m10.417798116s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:29.407078614 +0000 UTC m=+149.935985875" watchObservedRunningTime="2025-12-05 06:57:29.417798116 +0000 UTC m=+149.946705377" Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.479572 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:29 crc kubenswrapper[4997]: E1205 06:57:29.480007 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:29.979991439 +0000 UTC m=+150.508898700 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.562930 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qd6rv" podStartSLOduration=131.562908928 podStartE2EDuration="2m11.562908928s" podCreationTimestamp="2025-12-05 06:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:29.55537496 +0000 UTC m=+150.084282221" watchObservedRunningTime="2025-12-05 06:57:29.562908928 +0000 UTC m=+150.091816189" Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.582733 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps"] Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.583470 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:29 crc kubenswrapper[4997]: E1205 06:57:29.583954 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:30.08393297 +0000 UTC m=+150.612840231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.593212 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-brhqt"] Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.686222 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:29 crc kubenswrapper[4997]: E1205 06:57:29.687970 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:30.186603719 +0000 UTC m=+150.715510990 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.787679 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:29 crc kubenswrapper[4997]: E1205 06:57:29.788347 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:30.288314501 +0000 UTC m=+150.817221792 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:29 crc kubenswrapper[4997]: I1205 06:57:29.897912 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:29 crc kubenswrapper[4997]: E1205 06:57:29.898443 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:30.398414374 +0000 UTC m=+150.927321635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.018357 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:30 crc kubenswrapper[4997]: E1205 06:57:30.018548 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:30.51852072 +0000 UTC m=+151.047427981 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.019103 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:30 crc kubenswrapper[4997]: E1205 06:57:30.019431 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:30.519419164 +0000 UTC m=+151.048326425 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.037903 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz"] Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.105393 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vbxst"] Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.120838 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:30 crc kubenswrapper[4997]: E1205 06:57:30.121297 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:30.621276869 +0000 UTC m=+151.150184130 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.139870 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92"] Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.194239 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs9km"] Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.210074 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" event={"ID":"0dc484d1-5411-48aa-81d8-c208d0c5e2a6","Type":"ContainerStarted","Data":"19dfad0e5e4f45c297096c3577c1b86b479f4536a99aa10336921b8a7ce85603"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.218205 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7jtjl"] Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.224964 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:30 crc kubenswrapper[4997]: E1205 06:57:30.225673 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:30.725660683 +0000 UTC m=+151.254567944 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.266441 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zcnpm"] Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.267866 4997 generic.go:334] "Generic (PLEG): container finished" podID="bdc72754-1951-4312-9dd0-3458357f2caa" containerID="07e11596e6c18b6e9683b679ffd2e7745c5eea46872a3c561b8d979a8b630d71" exitCode=0 Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.267946 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jktft" event={"ID":"bdc72754-1951-4312-9dd0-3458357f2caa","Type":"ContainerDied","Data":"07e11596e6c18b6e9683b679ffd2e7745c5eea46872a3c561b8d979a8b630d71"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.267973 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jktft" event={"ID":"bdc72754-1951-4312-9dd0-3458357f2caa","Type":"ContainerStarted","Data":"6e069c4186e0c51e064cb4eba85ac1e5cd10f0297fa391102d4019ddb8ffa195"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.268552 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jktft" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.276534 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" event={"ID":"65e6a96e-df08-432d-8319-93e3e483f126","Type":"ContainerStarted","Data":"5a437a25d49350a38c7dd55238e1ef262e0b1a36be7e1828c7a7daaecc5041f5"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.303177 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" event={"ID":"1ea90520-ebc6-41e2-9854-54caf800fda7","Type":"ContainerStarted","Data":"458cb0372a71a40e81fdb7bbcacd2181e6904664f75e8625c880c8a3cc4f38c2"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.304990 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.310508 4997 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-rdtmj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:5443/healthz\": dial tcp 10.217.0.30:5443: connect: connection refused" start-of-body= Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.310548 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" podUID="1ea90520-ebc6-41e2-9854-54caf800fda7" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.30:5443/healthz\": dial tcp 10.217.0.30:5443: connect: connection refused" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.323174 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt"] Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.326495 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.326996 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hbt6g" event={"ID":"eba9ab59-079c-4512-b47c-b6c0e2a2886a","Type":"ContainerStarted","Data":"5be0fec6a1dc2b7c58cbe631fe8183dabf4fa4d95bee0b4adc60507f4a15632f"} Dec 05 06:57:30 crc kubenswrapper[4997]: E1205 06:57:30.327017 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:30.826987065 +0000 UTC m=+151.355894346 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.349215 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" podStartSLOduration=131.349185798 podStartE2EDuration="2m11.349185798s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:30.348313705 +0000 UTC m=+150.877220966" watchObservedRunningTime="2025-12-05 06:57:30.349185798 +0000 UTC m=+150.878093049" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.360315 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" event={"ID":"821da4d3-7e2a-480e-8290-a7080b916b08","Type":"ContainerStarted","Data":"115ea27ffbe8383392230c429671afa9f74468ecf9f9cde444830304c7b05910"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.379930 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" event={"ID":"8c6d7e0f-f064-48a0-b05b-287b09867b45","Type":"ContainerStarted","Data":"596f5f1136665454dd5249652d12b2b3d02dd262a95183e43f068a1ee5ce0178"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.402167 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-swdhl" event={"ID":"6008a937-1ce8-48ba-b1eb-bf915c452468","Type":"ContainerStarted","Data":"d6fb89ef129c9cd412618ae6cfe0ab9031b685cb834eaff43c15b44342ffbfec"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.414215 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jktft" podStartSLOduration=132.414189286 podStartE2EDuration="2m12.414189286s" podCreationTimestamp="2025-12-05 06:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:30.412282696 +0000 UTC m=+150.941189957" watchObservedRunningTime="2025-12-05 06:57:30.414189286 +0000 UTC m=+150.943096547" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.417500 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq"] Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.426735 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz" event={"ID":"1db3923d-a8bf-4ed6-b30e-4bdf15471401","Type":"ContainerStarted","Data":"93f42258b1c71d7c1066cc57dd1bdfcf415295314467ff704bc513ca7b51077b"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.426868 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7snv7" event={"ID":"baf64b43-1f21-4b84-ac70-3e0d879244a5","Type":"ContainerStarted","Data":"b763f49ec20e9dda8622bcfcc6924a5393a2791edecf64560393fd4483fb5f1f"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.426941 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f8k76" event={"ID":"33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3","Type":"ContainerStarted","Data":"1d0f509d2ec346f65c378d0ef5a35891efa25f14e9840e9e4929674e0c412cc8"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.427915 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:30 crc kubenswrapper[4997]: E1205 06:57:30.432802 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:30.932782305 +0000 UTC m=+151.461689566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.454286 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pjknx" event={"ID":"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0","Type":"ContainerStarted","Data":"83420e27cfbce0addc003a15226baa173cd074d3b699a455014a69938ee1866d"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.456652 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-5msxl" podStartSLOduration=131.456605121 podStartE2EDuration="2m11.456605121s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:30.454629258 +0000 UTC m=+150.983536529" watchObservedRunningTime="2025-12-05 06:57:30.456605121 +0000 UTC m=+150.985512382" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.483270 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-zm2tt" event={"ID":"e9370d84-0bfb-4e61-a5b7-bcb43e903593","Type":"ContainerStarted","Data":"5d19b952f664c684226e36f2b936fac78b042e633c21b4daefb46cdb46e64b31"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.493345 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fws6m"] Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.507310 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9k5xc" event={"ID":"800c31da-5f11-4029-a43f-ad96ac33b062","Type":"ContainerStarted","Data":"76a73a7cc90d65df9dfc9e071a192ab058ecddb7da26db8db03e876253b79be0"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.531943 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:30 crc kubenswrapper[4997]: E1205 06:57:30.532373 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:31.032343561 +0000 UTC m=+151.561250822 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.532668 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:30 crc kubenswrapper[4997]: E1205 06:57:30.534915 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:31.034522727 +0000 UTC m=+151.563429988 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.549740 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-s8cx6" event={"ID":"e0d60887-d7ee-4e14-96f1-81d1283ffa2e","Type":"ContainerStarted","Data":"029c0627f8aa9d89497f33acf871d2412b201e015f20e9009819b904c41c932c"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.621409 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" podStartSLOduration=131.62137719 podStartE2EDuration="2m11.62137719s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:30.531254892 +0000 UTC m=+151.060162163" watchObservedRunningTime="2025-12-05 06:57:30.62137719 +0000 UTC m=+151.150284451" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.623929 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-hbt6g" podStartSLOduration=6.623915446 podStartE2EDuration="6.623915446s" podCreationTimestamp="2025-12-05 06:57:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:30.611302285 +0000 UTC m=+151.140209546" watchObservedRunningTime="2025-12-05 06:57:30.623915446 +0000 UTC m=+151.152822697" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.637009 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.641649 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7" event={"ID":"ff56ef31-a073-49ac-b00a-e9611e2e0025","Type":"ContainerStarted","Data":"c58768bc9d52779b8255e08f083bc38be60aa90cb70fbb8641a05cb2d399131a"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.641729 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4fgv6"] Dec 05 06:57:30 crc kubenswrapper[4997]: E1205 06:57:30.648275 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:31.148206715 +0000 UTC m=+151.677113986 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.648709 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:30 crc kubenswrapper[4997]: E1205 06:57:30.657801 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:31.157780206 +0000 UTC m=+151.686687467 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.663863 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-649kx" podStartSLOduration=132.663839066 podStartE2EDuration="2m12.663839066s" podCreationTimestamp="2025-12-05 06:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:30.662803088 +0000 UTC m=+151.191710359" watchObservedRunningTime="2025-12-05 06:57:30.663839066 +0000 UTC m=+151.192746327" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.674858 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" event={"ID":"17fe9bd5-5973-47bd-945f-55bb8caa9d65","Type":"ContainerStarted","Data":"fd4cffdda303ba5d75ec82b29977cb9148b2ea1268e3de5b24b55dc798848fae"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.676120 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.679810 4997 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-k9twj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.685176 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" podUID="17fe9bd5-5973-47bd-945f-55bb8caa9d65" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.698255 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-swdhl" podStartSLOduration=6.698219809 podStartE2EDuration="6.698219809s" podCreationTimestamp="2025-12-05 06:57:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:30.696810942 +0000 UTC m=+151.225718213" watchObservedRunningTime="2025-12-05 06:57:30.698219809 +0000 UTC m=+151.227127070" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.730093 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zxx6b" event={"ID":"18378604-4db4-4da8-94c4-c52452ee6805","Type":"ContainerStarted","Data":"ffd9685b20f80ca1b1ab7437ff52e2a5c9fbe5d7e8017c531af546f12603283b"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.763127 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-67z92" event={"ID":"0836455d-d750-4a39-a7cc-6e95fe3e0abd","Type":"ContainerStarted","Data":"d2b5a8423b64ca71353b35d7326197a3684fcce72587ef86f4592fa306a28fff"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.763504 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-67z92" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.766263 4997 patch_prober.go:28] interesting pod/downloads-7954f5f757-67z92 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.766440 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-67z92" podUID="0836455d-d750-4a39-a7cc-6e95fe3e0abd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.769429 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:30 crc kubenswrapper[4997]: E1205 06:57:30.773921 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:31.273894027 +0000 UTC m=+151.802801328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.777542 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9k5xc" podStartSLOduration=132.777522583 podStartE2EDuration="2m12.777522583s" podCreationTimestamp="2025-12-05 06:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:30.743598542 +0000 UTC m=+151.272505803" watchObservedRunningTime="2025-12-05 06:57:30.777522583 +0000 UTC m=+151.306429844" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.781539 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7pzd7" podStartSLOduration=131.781524698 podStartE2EDuration="2m11.781524698s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:30.775776227 +0000 UTC m=+151.304683508" watchObservedRunningTime="2025-12-05 06:57:30.781524698 +0000 UTC m=+151.310431959" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.811878 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" podStartSLOduration=131.811858914 podStartE2EDuration="2m11.811858914s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:30.802645883 +0000 UTC m=+151.331553144" watchObservedRunningTime="2025-12-05 06:57:30.811858914 +0000 UTC m=+151.340766185" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.812550 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-brhqt" event={"ID":"a4b45170-1957-4ea8-b27e-e08dd68a56cd","Type":"ContainerStarted","Data":"996c09ee20b941502bcd80e3125c4c2927e3c04a606ace1cd50f3f74bbee471d"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.848478 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-zxx6b" podStartSLOduration=132.848461027 podStartE2EDuration="2m12.848461027s" podCreationTimestamp="2025-12-05 06:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:30.843539077 +0000 UTC m=+151.372446338" watchObservedRunningTime="2025-12-05 06:57:30.848461027 +0000 UTC m=+151.377368288" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.851272 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" event={"ID":"c7face06-9e66-4375-8b67-974b8936289f","Type":"ContainerStarted","Data":"fae5d15fc21990f52a6d2988565a9e08ee4600dff45b259ce31a28b8b5bb20ba"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.879821 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:30 crc kubenswrapper[4997]: E1205 06:57:30.880184 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:31.38017168 +0000 UTC m=+151.909078941 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.886899 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" event={"ID":"80dca7ef-76c4-4616-9d8a-81ff066f6464","Type":"ContainerStarted","Data":"8a4f89df0f2d55f3bef9847540f89974a3fcd572b968406e69aa9f511f474017"} Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.909545 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-67z92" podStartSLOduration=132.909525171 podStartE2EDuration="2m12.909525171s" podCreationTimestamp="2025-12-05 06:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:30.906195094 +0000 UTC m=+151.435102375" watchObservedRunningTime="2025-12-05 06:57:30.909525171 +0000 UTC m=+151.438432432" Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.981248 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:30 crc kubenswrapper[4997]: E1205 06:57:30.981927 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:31.481907723 +0000 UTC m=+152.010814984 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:30 crc kubenswrapper[4997]: I1205 06:57:30.996464 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wl94z" podStartSLOduration=131.996448525 podStartE2EDuration="2m11.996448525s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:30.996052035 +0000 UTC m=+151.524959296" watchObservedRunningTime="2025-12-05 06:57:30.996448525 +0000 UTC m=+151.525355786" Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.032721 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-v4mxn" event={"ID":"560926cd-d0fe-4823-b217-803ba02d2702","Type":"ContainerStarted","Data":"b1f93cc446f7e19b2205ef3427d222750e6e0ede1bda9b6377aab35a0bda7082"} Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.070192 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-v4mxn" podStartSLOduration=132.070170192 podStartE2EDuration="2m12.070170192s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:31.068674583 +0000 UTC m=+151.597581844" watchObservedRunningTime="2025-12-05 06:57:31.070170192 +0000 UTC m=+151.599077453" Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.082861 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:31 crc kubenswrapper[4997]: E1205 06:57:31.083932 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:31.583916243 +0000 UTC m=+152.112823584 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.183723 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:31 crc kubenswrapper[4997]: E1205 06:57:31.183977 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:31.683945981 +0000 UTC m=+152.212853242 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.184820 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:31 crc kubenswrapper[4997]: E1205 06:57:31.187313 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:31.68730136 +0000 UTC m=+152.216208621 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.286357 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:31 crc kubenswrapper[4997]: E1205 06:57:31.287025 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:31.78700911 +0000 UTC m=+152.315916371 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.391140 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.391637 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:31 crc kubenswrapper[4997]: E1205 06:57:31.391930 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:31.891918037 +0000 UTC m=+152.420825298 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.398437 4997 patch_prober.go:28] interesting pod/router-default-5444994796-v4mxn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 06:57:31 crc kubenswrapper[4997]: [-]has-synced failed: reason withheld Dec 05 06:57:31 crc kubenswrapper[4997]: [+]process-running ok Dec 05 06:57:31 crc kubenswrapper[4997]: healthz check failed Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.399236 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v4mxn" podUID="560926cd-d0fe-4823-b217-803ba02d2702" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.493384 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:31 crc kubenswrapper[4997]: E1205 06:57:31.493503 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:31.993478684 +0000 UTC m=+152.522385945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.493660 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:31 crc kubenswrapper[4997]: E1205 06:57:31.494015 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:31.994001419 +0000 UTC m=+152.522908680 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.561702 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.562582 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.585250 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.595577 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:31 crc kubenswrapper[4997]: E1205 06:57:31.596153 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:32.096121802 +0000 UTC m=+152.625029073 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.700377 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:31 crc kubenswrapper[4997]: E1205 06:57:31.700853 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:32.200838923 +0000 UTC m=+152.729746194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.802192 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:31 crc kubenswrapper[4997]: E1205 06:57:31.802561 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:32.302546635 +0000 UTC m=+152.831453896 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:31 crc kubenswrapper[4997]: I1205 06:57:31.903475 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:31 crc kubenswrapper[4997]: E1205 06:57:31.903885 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:32.403871867 +0000 UTC m=+152.932779128 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.004799 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:32 crc kubenswrapper[4997]: E1205 06:57:32.005171 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:32.505151419 +0000 UTC m=+153.034058680 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.054359 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7snv7" event={"ID":"baf64b43-1f21-4b84-ac70-3e0d879244a5","Type":"ContainerStarted","Data":"52f45653d8a3e60e4fff8b8913d261606b7e8def21ed1a38637af8692bed7518"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.081955 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq" event={"ID":"e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6","Type":"ContainerStarted","Data":"228a874fdc32d0e9340346a3712cb11b986395339037f62b844ba40a53f7eafa"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.082011 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq" event={"ID":"e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6","Type":"ContainerStarted","Data":"28efa5cb89cea15b6e2c9f8776d81c0856d6a4975f1167230291b6ada2994845"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.083325 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.086707 4997 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-8vkgq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.086745 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq" podUID="e85d6a1d-cc7f-47d4-8d6a-ce61c1d94eb6" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.091210 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-vbxst" event={"ID":"1247bce5-d4bf-4a8b-a438-3b150109ef5f","Type":"ContainerStarted","Data":"9117636b85f3e46ece819091f4770e9cdf77ebf41985a83b6aded98aeb82c8c4"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.091242 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-vbxst" event={"ID":"1247bce5-d4bf-4a8b-a438-3b150109ef5f","Type":"ContainerStarted","Data":"4fa8ae6d326f33020747dfac4ec27e6f58fc8d627487793a0ebc70f348f14963"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.104833 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f8k76" event={"ID":"33a48f7a-01bf-4dc4-8b1d-589d72e3e1e3","Type":"ContainerStarted","Data":"a9ea002d2a83e7242ab58d696529c6e1d65ee71c85f2673cb148957638b892b5"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.106780 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:32 crc kubenswrapper[4997]: E1205 06:57:32.107110 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:32.607095428 +0000 UTC m=+153.136002689 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.113243 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq" podStartSLOduration=133.113224649 podStartE2EDuration="2m13.113224649s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:32.113139387 +0000 UTC m=+152.642046648" watchObservedRunningTime="2025-12-05 06:57:32.113224649 +0000 UTC m=+152.642131910" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.122801 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wl94z" event={"ID":"487642f9-0987-4953-b369-f092392d0209","Type":"ContainerStarted","Data":"4f8359e37aacf5d088c45ce31423015822a72e3bb0e41addbd5ae6227f2e9da5"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.130033 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" event={"ID":"f692422b-da9f-4775-825e-ee198e5b61b8","Type":"ContainerStarted","Data":"69e70f6cd56b05b725830aa87f03ff2323fde92fedc1b17a8fc26c0917b692da"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.130098 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" event={"ID":"f692422b-da9f-4775-825e-ee198e5b61b8","Type":"ContainerStarted","Data":"ce7145efdab5eb6e4044397dcb768c91cff7fa1a145c0a6f307d0651b57b90df"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.138315 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f8k76" podStartSLOduration=133.138296468 podStartE2EDuration="2m13.138296468s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:32.13726734 +0000 UTC m=+152.666174601" watchObservedRunningTime="2025-12-05 06:57:32.138296468 +0000 UTC m=+152.667203729" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.159077 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" event={"ID":"c7face06-9e66-4375-8b67-974b8936289f","Type":"ContainerStarted","Data":"bc7eca0ccad48d064c273811fc6ab3175ce6160274990f848664a3147ca53784"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.174289 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zcnpm" event={"ID":"49b1c6db-0638-4367-8264-eec97250ad0a","Type":"ContainerStarted","Data":"6b2120aad0a25c92ffcc727a011a843a03a3f55b7398a1ebc5d53632ba8b5715"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.174333 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zcnpm" event={"ID":"49b1c6db-0638-4367-8264-eec97250ad0a","Type":"ContainerStarted","Data":"fae9dd362059584ae2a8aba087b52834f202273a9c266f4f3312b68cade630ad"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.208879 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:32 crc kubenswrapper[4997]: E1205 06:57:32.210162 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:32.710140985 +0000 UTC m=+153.239048246 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.214866 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-vbxst" podStartSLOduration=133.214830959 podStartE2EDuration="2m13.214830959s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:32.163097979 +0000 UTC m=+152.692005250" watchObservedRunningTime="2025-12-05 06:57:32.214830959 +0000 UTC m=+152.743738220" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.220154 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4fgv6" event={"ID":"3c394687-7ae3-4500-bda2-a5de1cb7b8b1","Type":"ContainerStarted","Data":"d08a858d81123c1f5f8cf5b3288c0ec56b095717f4ff5a8e12b7576874280038"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.220210 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4fgv6" event={"ID":"3c394687-7ae3-4500-bda2-a5de1cb7b8b1","Type":"ContainerStarted","Data":"189d49e73b915304b53c5f26f0f51c825c6408ca1cabd2e0ec9482b3cb8405c6"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.258390 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt" event={"ID":"cf990560-e430-4c2b-8965-c5ea18f8b695","Type":"ContainerStarted","Data":"3da9ae3acf9f429b9171155518ea1f5ab0633cc6ad3762f6b7cb492b9fe02bbc"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.258447 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt" event={"ID":"cf990560-e430-4c2b-8965-c5ea18f8b695","Type":"ContainerStarted","Data":"402ce6aa3ad2965df788ae7ee0f6d4c4e53c7984671fa990ea4bbc9f3122a38b"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.259644 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.259805 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" podStartSLOduration=134.25978904 podStartE2EDuration="2m14.25978904s" podCreationTimestamp="2025-12-05 06:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:32.217411986 +0000 UTC m=+152.746319247" watchObservedRunningTime="2025-12-05 06:57:32.25978904 +0000 UTC m=+152.788696301" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.261175 4997 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-d28rt container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.261270 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt" podUID="cf990560-e430-4c2b-8965-c5ea18f8b695" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.262947 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" event={"ID":"17fe9bd5-5973-47bd-945f-55bb8caa9d65","Type":"ContainerStarted","Data":"feb7b9a4853bb63a72df3919b0a2cb6b319deb2104e5684bc736c65bf79d9844"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.264549 4997 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-k9twj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.264646 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" podUID="17fe9bd5-5973-47bd-945f-55bb8caa9d65" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.286903 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fws6m" event={"ID":"f915acae-6293-415a-a2a4-266340eee2e9","Type":"ContainerStarted","Data":"9e718e0c5f8e05e4b2a7af8d24201e93cb9c46bc678c3b1919b4107b7ee56d9a"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.296483 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt" podStartSLOduration=133.296465023 podStartE2EDuration="2m13.296465023s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:32.291487662 +0000 UTC m=+152.820394933" watchObservedRunningTime="2025-12-05 06:57:32.296465023 +0000 UTC m=+152.825372284" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.298829 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" podStartSLOduration=133.298822446 podStartE2EDuration="2m13.298822446s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:32.259929424 +0000 UTC m=+152.788836685" watchObservedRunningTime="2025-12-05 06:57:32.298822446 +0000 UTC m=+152.827729707" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.308879 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs9km" event={"ID":"ab4d334f-b4a4-4d04-bf90-d2178f991f5f","Type":"ContainerStarted","Data":"a424ac8575ddca461da01b856d43c28fd552ffea444a42fba6db00d7d3d154a5"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.308933 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs9km" event={"ID":"ab4d334f-b4a4-4d04-bf90-d2178f991f5f","Type":"ContainerStarted","Data":"f9a70bc6489fa23b81c0ae2045dc5f8db06d736fe1d961f6e513d4a1bd279ee9"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.310144 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:32 crc kubenswrapper[4997]: E1205 06:57:32.313425 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:32.813412749 +0000 UTC m=+153.342320010 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.334098 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4fgv6" podStartSLOduration=133.334077812 podStartE2EDuration="2m13.334077812s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:32.332657995 +0000 UTC m=+152.861565276" watchObservedRunningTime="2025-12-05 06:57:32.334077812 +0000 UTC m=+152.862985073" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.334390 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" event={"ID":"0dc484d1-5411-48aa-81d8-c208d0c5e2a6","Type":"ContainerStarted","Data":"dd4e83a6c90b4949388cdc88993154fcfb956b0f3dd2b0f58f8be5163cb1e8a8"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.353450 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-brhqt" event={"ID":"a4b45170-1957-4ea8-b27e-e08dd68a56cd","Type":"ContainerStarted","Data":"f9f4de205e2743496c3b0b06e826b256a612249fd14505e8e9f1d2b1ae8e25b2"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.353539 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-brhqt" event={"ID":"a4b45170-1957-4ea8-b27e-e08dd68a56cd","Type":"ContainerStarted","Data":"b1a39070c8c86753ce237b08d7b4c896190cd31770bdf8a8a6058809f92de3f0"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.376290 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fws6m" podStartSLOduration=133.37627348 podStartE2EDuration="2m13.37627348s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:32.374539475 +0000 UTC m=+152.903446736" watchObservedRunningTime="2025-12-05 06:57:32.37627348 +0000 UTC m=+152.905180741" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.379394 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz" event={"ID":"1db3923d-a8bf-4ed6-b30e-4bdf15471401","Type":"ContainerStarted","Data":"5ec53ec13ac570f88f61b9dd9dea27d7b1354e47cb4760cb4f156f29bdc829e8"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.379435 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz" event={"ID":"1db3923d-a8bf-4ed6-b30e-4bdf15471401","Type":"ContainerStarted","Data":"2883ffbb354b017bbc5db76c535fd0553070b09d40ba89d6c1f9fd3f07bb39d5"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.381170 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" event={"ID":"80dca7ef-76c4-4616-9d8a-81ff066f6464","Type":"ContainerStarted","Data":"5ba6ed9f96604a615a16edd781267cb5c6bd35a1714dfbe975d744e8767b513b"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.381190 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" event={"ID":"80dca7ef-76c4-4616-9d8a-81ff066f6464","Type":"ContainerStarted","Data":"380ddf58b579ed80e32cfad5756357ecb583f77f41cef6cb851d8817229f7a9e"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.393002 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7jtjl" event={"ID":"2be0da6d-6565-4336-89e0-590353e31685","Type":"ContainerStarted","Data":"1014ebb59d5bb3f85dcbb6a652c40b1ae73095eccfcd5623f52ad025a0718f32"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.393053 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7jtjl" event={"ID":"2be0da6d-6565-4336-89e0-590353e31685","Type":"ContainerStarted","Data":"b25fdf22d266ada1ef0220f27d8d3df4d97b1cd7b4c37f460ac1f73861e2ec70"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.397888 4997 patch_prober.go:28] interesting pod/router-default-5444994796-v4mxn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 06:57:32 crc kubenswrapper[4997]: [-]has-synced failed: reason withheld Dec 05 06:57:32 crc kubenswrapper[4997]: [+]process-running ok Dec 05 06:57:32 crc kubenswrapper[4997]: healthz check failed Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.397961 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v4mxn" podUID="560926cd-d0fe-4823-b217-803ba02d2702" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.412695 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-s8cx6" event={"ID":"e0d60887-d7ee-4e14-96f1-81d1283ffa2e","Type":"ContainerStarted","Data":"bb45bc0cadda6aecc2362d767a0ede19013a05bd85594761f215f557ad1a8e90"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.412743 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-s8cx6" event={"ID":"e0d60887-d7ee-4e14-96f1-81d1283ffa2e","Type":"ContainerStarted","Data":"b37826a55ef99c5bcd12f8a3faec37a8cb6eb77b681f214f40a6a8ab5139e882"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.413468 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-s8cx6" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.414067 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:32 crc kubenswrapper[4997]: E1205 06:57:32.415182 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:32.915159572 +0000 UTC m=+153.444066843 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.418293 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs9km" podStartSLOduration=133.418280304 podStartE2EDuration="2m13.418280304s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:32.414248028 +0000 UTC m=+152.943155299" watchObservedRunningTime="2025-12-05 06:57:32.418280304 +0000 UTC m=+152.947187565" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.434290 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pjknx" event={"ID":"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0","Type":"ContainerStarted","Data":"91f5a66c48d1d4a43ebd87bdde90bd559c5d86ab7460242701ad761ef72715c5"} Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.439292 4997 patch_prober.go:28] interesting pod/downloads-7954f5f757-67z92 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.439432 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-67z92" podUID="0836455d-d750-4a39-a7cc-6e95fe3e0abd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.459144 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rdtmj" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.461877 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4kdd" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.471304 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jktft" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.474122 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7jtjl" podStartSLOduration=133.474110941 podStartE2EDuration="2m13.474110941s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:32.471461332 +0000 UTC m=+153.000368593" watchObservedRunningTime="2025-12-05 06:57:32.474110941 +0000 UTC m=+153.003018192" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.495462 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-brhqt" podStartSLOduration=133.495442932 podStartE2EDuration="2m13.495442932s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:32.492568096 +0000 UTC m=+153.021475367" watchObservedRunningTime="2025-12-05 06:57:32.495442932 +0000 UTC m=+153.024350193" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.520882 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.520910 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2kfps" podStartSLOduration=133.52089416 podStartE2EDuration="2m13.52089416s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:32.519924824 +0000 UTC m=+153.048832095" watchObservedRunningTime="2025-12-05 06:57:32.52089416 +0000 UTC m=+153.049801411" Dec 05 06:57:32 crc kubenswrapper[4997]: E1205 06:57:32.524170 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:33.024155626 +0000 UTC m=+153.553062877 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.570633 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wsvhz" podStartSLOduration=133.570596017 podStartE2EDuration="2m13.570596017s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:32.553582689 +0000 UTC m=+153.082489960" watchObservedRunningTime="2025-12-05 06:57:32.570596017 +0000 UTC m=+153.099503278" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.597676 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n5wgm" podStartSLOduration=133.597649277 podStartE2EDuration="2m13.597649277s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:32.58596119 +0000 UTC m=+153.114868471" watchObservedRunningTime="2025-12-05 06:57:32.597649277 +0000 UTC m=+153.126556548" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.623160 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:32 crc kubenswrapper[4997]: E1205 06:57:32.623409 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:33.123364502 +0000 UTC m=+153.652271763 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.623594 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:32 crc kubenswrapper[4997]: E1205 06:57:32.624320 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:33.124299697 +0000 UTC m=+153.653206948 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.642035 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-s8cx6" podStartSLOduration=133.642018533 podStartE2EDuration="2m13.642018533s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:32.639303842 +0000 UTC m=+153.168211113" watchObservedRunningTime="2025-12-05 06:57:32.642018533 +0000 UTC m=+153.170925784" Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.724823 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:32 crc kubenswrapper[4997]: E1205 06:57:32.725378 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:33.225357572 +0000 UTC m=+153.754264833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.826630 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:32 crc kubenswrapper[4997]: E1205 06:57:32.826978 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:33.326959492 +0000 UTC m=+153.855866753 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.930650 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:32 crc kubenswrapper[4997]: E1205 06:57:32.930886 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:33.430849702 +0000 UTC m=+153.959757033 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:32 crc kubenswrapper[4997]: I1205 06:57:32.931344 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:32 crc kubenswrapper[4997]: E1205 06:57:32.931714 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:33.431705985 +0000 UTC m=+153.960613246 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.032806 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:33 crc kubenswrapper[4997]: E1205 06:57:33.033029 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:33.532997246 +0000 UTC m=+154.061904497 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.033314 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:33 crc kubenswrapper[4997]: E1205 06:57:33.033671 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:33.533663053 +0000 UTC m=+154.062570304 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.134790 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:33 crc kubenswrapper[4997]: E1205 06:57:33.134958 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:33.634930515 +0000 UTC m=+154.163837766 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.135034 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:33 crc kubenswrapper[4997]: E1205 06:57:33.135320 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:33.635312454 +0000 UTC m=+154.164219715 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.236562 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:33 crc kubenswrapper[4997]: E1205 06:57:33.236868 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:33.736815161 +0000 UTC m=+154.265722432 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.237286 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:33 crc kubenswrapper[4997]: E1205 06:57:33.237755 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:33.737732915 +0000 UTC m=+154.266640176 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.339104 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:33 crc kubenswrapper[4997]: E1205 06:57:33.339676 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:33.839649264 +0000 UTC m=+154.368556525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.340225 4997 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.392140 4997 patch_prober.go:28] interesting pod/router-default-5444994796-v4mxn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 06:57:33 crc kubenswrapper[4997]: [-]has-synced failed: reason withheld Dec 05 06:57:33 crc kubenswrapper[4997]: [+]process-running ok Dec 05 06:57:33 crc kubenswrapper[4997]: healthz check failed Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.392256 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v4mxn" podUID="560926cd-d0fe-4823-b217-803ba02d2702" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.440481 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:33 crc kubenswrapper[4997]: E1205 06:57:33.441100 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:33.941078448 +0000 UTC m=+154.469985769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.482277 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zcnpm" event={"ID":"49b1c6db-0638-4367-8264-eec97250ad0a","Type":"ContainerStarted","Data":"7d31d04a2d7926ad43811ae587632af427405696d869880a837b516aeef79407"} Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.483012 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-zcnpm" Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.488303 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7snv7" event={"ID":"baf64b43-1f21-4b84-ac70-3e0d879244a5","Type":"ContainerStarted","Data":"25f7153b666cf89741e9a8a0fcc5fe09bb8e92f595633f7f879473ca513b1f37"} Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.500913 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fws6m" event={"ID":"f915acae-6293-415a-a2a4-266340eee2e9","Type":"ContainerStarted","Data":"11e16591ea094eeeafcb996cc2bf09be51165dc48b1a55cee66b2077d1cbd5b7"} Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.526693 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pjknx" event={"ID":"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0","Type":"ContainerStarted","Data":"3a6d89daa7e03dc42b6958ad7981187af5f455df8acfc5b27b2b6f4956d76e3b"} Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.526756 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pjknx" event={"ID":"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0","Type":"ContainerStarted","Data":"ed56b36c0db2d5c9e8822af23243fd14a6d268afddc241e65a2709b4d06e6b0f"} Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.528493 4997 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-k9twj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.528636 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" podUID="17fe9bd5-5973-47bd-945f-55bb8caa9d65" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.535033 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-d28rt" Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.541752 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:33 crc kubenswrapper[4997]: E1205 06:57:33.542134 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:34.042094832 +0000 UTC m=+154.571002173 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.552559 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-zcnpm" podStartSLOduration=9.552541487 podStartE2EDuration="9.552541487s" podCreationTimestamp="2025-12-05 06:57:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:33.513380108 +0000 UTC m=+154.042287399" watchObservedRunningTime="2025-12-05 06:57:33.552541487 +0000 UTC m=+154.081448748" Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.571504 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-7snv7" podStartSLOduration=134.571482524 podStartE2EDuration="2m14.571482524s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:33.55113925 +0000 UTC m=+154.080046521" watchObservedRunningTime="2025-12-05 06:57:33.571482524 +0000 UTC m=+154.100389795" Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.619230 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8vkgq" Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.649355 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:33 crc kubenswrapper[4997]: E1205 06:57:33.653211 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:34.153185272 +0000 UTC m=+154.682092713 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.752384 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:33 crc kubenswrapper[4997]: E1205 06:57:33.752577 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:34.252539762 +0000 UTC m=+154.781447023 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.752817 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:33 crc kubenswrapper[4997]: E1205 06:57:33.753221 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:34.25321343 +0000 UTC m=+154.782120691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.853952 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:33 crc kubenswrapper[4997]: E1205 06:57:33.855408 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:34.355384444 +0000 UTC m=+154.884291705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.956205 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:33 crc kubenswrapper[4997]: E1205 06:57:33.956698 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 06:57:34.456672616 +0000 UTC m=+154.985579877 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xbd9" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:33 crc kubenswrapper[4997]: I1205 06:57:33.964201 4997 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-05T06:57:33.340250949Z","Handler":null,"Name":""} Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.057455 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:34 crc kubenswrapper[4997]: E1205 06:57:34.057983 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 06:57:34.557949827 +0000 UTC m=+155.086857088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.091760 4997 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.091842 4997 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.159256 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.200931 4997 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.200988 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.244982 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xbd9\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.264222 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.273811 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.371704 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2pls6"] Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.372862 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2pls6" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.375945 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.384797 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2pls6"] Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.392822 4997 patch_prober.go:28] interesting pod/router-default-5444994796-v4mxn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 06:57:34 crc kubenswrapper[4997]: [-]has-synced failed: reason withheld Dec 05 06:57:34 crc kubenswrapper[4997]: [+]process-running ok Dec 05 06:57:34 crc kubenswrapper[4997]: healthz check failed Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.392873 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v4mxn" podUID="560926cd-d0fe-4823-b217-803ba02d2702" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.470016 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2s94\" (UniqueName: \"kubernetes.io/projected/3ec5dfd3-3221-4141-a2d4-aea02733a985-kube-api-access-r2s94\") pod \"community-operators-2pls6\" (UID: \"3ec5dfd3-3221-4141-a2d4-aea02733a985\") " pod="openshift-marketplace/community-operators-2pls6" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.470097 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ec5dfd3-3221-4141-a2d4-aea02733a985-utilities\") pod \"community-operators-2pls6\" (UID: \"3ec5dfd3-3221-4141-a2d4-aea02733a985\") " pod="openshift-marketplace/community-operators-2pls6" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.470168 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ec5dfd3-3221-4141-a2d4-aea02733a985-catalog-content\") pod \"community-operators-2pls6\" (UID: \"3ec5dfd3-3221-4141-a2d4-aea02733a985\") " pod="openshift-marketplace/community-operators-2pls6" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.547881 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.575057 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ec5dfd3-3221-4141-a2d4-aea02733a985-catalog-content\") pod \"community-operators-2pls6\" (UID: \"3ec5dfd3-3221-4141-a2d4-aea02733a985\") " pod="openshift-marketplace/community-operators-2pls6" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.575402 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2s94\" (UniqueName: \"kubernetes.io/projected/3ec5dfd3-3221-4141-a2d4-aea02733a985-kube-api-access-r2s94\") pod \"community-operators-2pls6\" (UID: \"3ec5dfd3-3221-4141-a2d4-aea02733a985\") " pod="openshift-marketplace/community-operators-2pls6" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.575496 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ec5dfd3-3221-4141-a2d4-aea02733a985-utilities\") pod \"community-operators-2pls6\" (UID: \"3ec5dfd3-3221-4141-a2d4-aea02733a985\") " pod="openshift-marketplace/community-operators-2pls6" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.575624 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ec5dfd3-3221-4141-a2d4-aea02733a985-catalog-content\") pod \"community-operators-2pls6\" (UID: \"3ec5dfd3-3221-4141-a2d4-aea02733a985\") " pod="openshift-marketplace/community-operators-2pls6" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.575973 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ec5dfd3-3221-4141-a2d4-aea02733a985-utilities\") pod \"community-operators-2pls6\" (UID: \"3ec5dfd3-3221-4141-a2d4-aea02733a985\") " pod="openshift-marketplace/community-operators-2pls6" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.577816 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pjknx" event={"ID":"ff590dc9-b474-4ae3-a66d-ea8f6e6797d0","Type":"ContainerStarted","Data":"adb8b4e6c43218ae0a3f8a290adc8150fd57076c650a3c18a0133b41f5f29a66"} Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.582327 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n6tmw"] Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.583191 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6tmw" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.584764 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.601665 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2s94\" (UniqueName: \"kubernetes.io/projected/3ec5dfd3-3221-4141-a2d4-aea02733a985-kube-api-access-r2s94\") pod \"community-operators-2pls6\" (UID: \"3ec5dfd3-3221-4141-a2d4-aea02733a985\") " pod="openshift-marketplace/community-operators-2pls6" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.607411 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-pjknx" podStartSLOduration=10.607398774 podStartE2EDuration="10.607398774s" podCreationTimestamp="2025-12-05 06:57:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:34.605385561 +0000 UTC m=+155.134292832" watchObservedRunningTime="2025-12-05 06:57:34.607398774 +0000 UTC m=+155.136306035" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.611165 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n6tmw"] Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.676507 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9vmz\" (UniqueName: \"kubernetes.io/projected/c0ca09da-0456-4ab9-8260-03c64893a6dd-kube-api-access-t9vmz\") pod \"certified-operators-n6tmw\" (UID: \"c0ca09da-0456-4ab9-8260-03c64893a6dd\") " pod="openshift-marketplace/certified-operators-n6tmw" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.676628 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0ca09da-0456-4ab9-8260-03c64893a6dd-utilities\") pod \"certified-operators-n6tmw\" (UID: \"c0ca09da-0456-4ab9-8260-03c64893a6dd\") " pod="openshift-marketplace/certified-operators-n6tmw" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.676752 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0ca09da-0456-4ab9-8260-03c64893a6dd-catalog-content\") pod \"certified-operators-n6tmw\" (UID: \"c0ca09da-0456-4ab9-8260-03c64893a6dd\") " pod="openshift-marketplace/certified-operators-n6tmw" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.690363 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2pls6" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.763624 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cssd2"] Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.765107 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cssd2" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.779508 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cssd2"] Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.779759 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9vmz\" (UniqueName: \"kubernetes.io/projected/c0ca09da-0456-4ab9-8260-03c64893a6dd-kube-api-access-t9vmz\") pod \"certified-operators-n6tmw\" (UID: \"c0ca09da-0456-4ab9-8260-03c64893a6dd\") " pod="openshift-marketplace/certified-operators-n6tmw" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.779811 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0ca09da-0456-4ab9-8260-03c64893a6dd-utilities\") pod \"certified-operators-n6tmw\" (UID: \"c0ca09da-0456-4ab9-8260-03c64893a6dd\") " pod="openshift-marketplace/certified-operators-n6tmw" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.779860 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0ca09da-0456-4ab9-8260-03c64893a6dd-catalog-content\") pod \"certified-operators-n6tmw\" (UID: \"c0ca09da-0456-4ab9-8260-03c64893a6dd\") " pod="openshift-marketplace/certified-operators-n6tmw" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.780946 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0ca09da-0456-4ab9-8260-03c64893a6dd-catalog-content\") pod \"certified-operators-n6tmw\" (UID: \"c0ca09da-0456-4ab9-8260-03c64893a6dd\") " pod="openshift-marketplace/certified-operators-n6tmw" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.781097 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0ca09da-0456-4ab9-8260-03c64893a6dd-utilities\") pod \"certified-operators-n6tmw\" (UID: \"c0ca09da-0456-4ab9-8260-03c64893a6dd\") " pod="openshift-marketplace/certified-operators-n6tmw" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.807381 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9vmz\" (UniqueName: \"kubernetes.io/projected/c0ca09da-0456-4ab9-8260-03c64893a6dd-kube-api-access-t9vmz\") pod \"certified-operators-n6tmw\" (UID: \"c0ca09da-0456-4ab9-8260-03c64893a6dd\") " pod="openshift-marketplace/certified-operators-n6tmw" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.881050 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd5485d-8c41-484e-9373-4624f05a0f56-catalog-content\") pod \"community-operators-cssd2\" (UID: \"7fd5485d-8c41-484e-9373-4624f05a0f56\") " pod="openshift-marketplace/community-operators-cssd2" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.881140 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd5485d-8c41-484e-9373-4624f05a0f56-utilities\") pod \"community-operators-cssd2\" (UID: \"7fd5485d-8c41-484e-9373-4624f05a0f56\") " pod="openshift-marketplace/community-operators-cssd2" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.881178 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thxwv\" (UniqueName: \"kubernetes.io/projected/7fd5485d-8c41-484e-9373-4624f05a0f56-kube-api-access-thxwv\") pod \"community-operators-cssd2\" (UID: \"7fd5485d-8c41-484e-9373-4624f05a0f56\") " pod="openshift-marketplace/community-operators-cssd2" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.936740 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5xbd9"] Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.951088 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6tmw" Dec 05 06:57:34 crc kubenswrapper[4997]: W1205 06:57:34.959016 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd09450e_798d_41d5_b925_90ab15df8dac.slice/crio-dd16f29590ca291a136e4ff7ad0143adfb64361ee01979c09a146a027afcd726 WatchSource:0}: Error finding container dd16f29590ca291a136e4ff7ad0143adfb64361ee01979c09a146a027afcd726: Status 404 returned error can't find the container with id dd16f29590ca291a136e4ff7ad0143adfb64361ee01979c09a146a027afcd726 Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.963491 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5hfgl"] Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.965022 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5hfgl" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.973042 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5hfgl"] Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.982662 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd5485d-8c41-484e-9373-4624f05a0f56-catalog-content\") pod \"community-operators-cssd2\" (UID: \"7fd5485d-8c41-484e-9373-4624f05a0f56\") " pod="openshift-marketplace/community-operators-cssd2" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.982744 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd5485d-8c41-484e-9373-4624f05a0f56-utilities\") pod \"community-operators-cssd2\" (UID: \"7fd5485d-8c41-484e-9373-4624f05a0f56\") " pod="openshift-marketplace/community-operators-cssd2" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.982799 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thxwv\" (UniqueName: \"kubernetes.io/projected/7fd5485d-8c41-484e-9373-4624f05a0f56-kube-api-access-thxwv\") pod \"community-operators-cssd2\" (UID: \"7fd5485d-8c41-484e-9373-4624f05a0f56\") " pod="openshift-marketplace/community-operators-cssd2" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.983604 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd5485d-8c41-484e-9373-4624f05a0f56-catalog-content\") pod \"community-operators-cssd2\" (UID: \"7fd5485d-8c41-484e-9373-4624f05a0f56\") " pod="openshift-marketplace/community-operators-cssd2" Dec 05 06:57:34 crc kubenswrapper[4997]: I1205 06:57:34.983834 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd5485d-8c41-484e-9373-4624f05a0f56-utilities\") pod \"community-operators-cssd2\" (UID: \"7fd5485d-8c41-484e-9373-4624f05a0f56\") " pod="openshift-marketplace/community-operators-cssd2" Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.013950 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thxwv\" (UniqueName: \"kubernetes.io/projected/7fd5485d-8c41-484e-9373-4624f05a0f56-kube-api-access-thxwv\") pod \"community-operators-cssd2\" (UID: \"7fd5485d-8c41-484e-9373-4624f05a0f56\") " pod="openshift-marketplace/community-operators-cssd2" Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.084360 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa0cd037-9961-467b-a659-83922f8ed830-utilities\") pod \"certified-operators-5hfgl\" (UID: \"aa0cd037-9961-467b-a659-83922f8ed830\") " pod="openshift-marketplace/certified-operators-5hfgl" Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.084420 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa0cd037-9961-467b-a659-83922f8ed830-catalog-content\") pod \"certified-operators-5hfgl\" (UID: \"aa0cd037-9961-467b-a659-83922f8ed830\") " pod="openshift-marketplace/certified-operators-5hfgl" Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.084673 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5px8g\" (UniqueName: \"kubernetes.io/projected/aa0cd037-9961-467b-a659-83922f8ed830-kube-api-access-5px8g\") pod \"certified-operators-5hfgl\" (UID: \"aa0cd037-9961-467b-a659-83922f8ed830\") " pod="openshift-marketplace/certified-operators-5hfgl" Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.098356 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cssd2" Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.116360 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2pls6"] Dec 05 06:57:35 crc kubenswrapper[4997]: W1205 06:57:35.140925 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ec5dfd3_3221_4141_a2d4_aea02733a985.slice/crio-b857b36770f2e4fa834bc88a9e2f2cff0bc6813ceaba068b977d90eb0e5d6c7e WatchSource:0}: Error finding container b857b36770f2e4fa834bc88a9e2f2cff0bc6813ceaba068b977d90eb0e5d6c7e: Status 404 returned error can't find the container with id b857b36770f2e4fa834bc88a9e2f2cff0bc6813ceaba068b977d90eb0e5d6c7e Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.186321 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5px8g\" (UniqueName: \"kubernetes.io/projected/aa0cd037-9961-467b-a659-83922f8ed830-kube-api-access-5px8g\") pod \"certified-operators-5hfgl\" (UID: \"aa0cd037-9961-467b-a659-83922f8ed830\") " pod="openshift-marketplace/certified-operators-5hfgl" Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.186385 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa0cd037-9961-467b-a659-83922f8ed830-utilities\") pod \"certified-operators-5hfgl\" (UID: \"aa0cd037-9961-467b-a659-83922f8ed830\") " pod="openshift-marketplace/certified-operators-5hfgl" Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.187050 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa0cd037-9961-467b-a659-83922f8ed830-utilities\") pod \"certified-operators-5hfgl\" (UID: \"aa0cd037-9961-467b-a659-83922f8ed830\") " pod="openshift-marketplace/certified-operators-5hfgl" Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.188281 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa0cd037-9961-467b-a659-83922f8ed830-catalog-content\") pod \"certified-operators-5hfgl\" (UID: \"aa0cd037-9961-467b-a659-83922f8ed830\") " pod="openshift-marketplace/certified-operators-5hfgl" Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.190118 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa0cd037-9961-467b-a659-83922f8ed830-catalog-content\") pod \"certified-operators-5hfgl\" (UID: \"aa0cd037-9961-467b-a659-83922f8ed830\") " pod="openshift-marketplace/certified-operators-5hfgl" Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.209015 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5px8g\" (UniqueName: \"kubernetes.io/projected/aa0cd037-9961-467b-a659-83922f8ed830-kube-api-access-5px8g\") pod \"certified-operators-5hfgl\" (UID: \"aa0cd037-9961-467b-a659-83922f8ed830\") " pod="openshift-marketplace/certified-operators-5hfgl" Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.295968 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5hfgl" Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.303651 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n6tmw"] Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.393588 4997 patch_prober.go:28] interesting pod/router-default-5444994796-v4mxn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 06:57:35 crc kubenswrapper[4997]: [-]has-synced failed: reason withheld Dec 05 06:57:35 crc kubenswrapper[4997]: [+]process-running ok Dec 05 06:57:35 crc kubenswrapper[4997]: healthz check failed Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.393697 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v4mxn" podUID="560926cd-d0fe-4823-b217-803ba02d2702" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.589785 4997 generic.go:334] "Generic (PLEG): container finished" podID="3ec5dfd3-3221-4141-a2d4-aea02733a985" containerID="08691ef05fca2b5aa6f6734da33f74f246a7c44cc89dd5b7afa16ec3c67cb724" exitCode=0 Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.590119 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2pls6" event={"ID":"3ec5dfd3-3221-4141-a2d4-aea02733a985","Type":"ContainerDied","Data":"08691ef05fca2b5aa6f6734da33f74f246a7c44cc89dd5b7afa16ec3c67cb724"} Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.591206 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2pls6" event={"ID":"3ec5dfd3-3221-4141-a2d4-aea02733a985","Type":"ContainerStarted","Data":"b857b36770f2e4fa834bc88a9e2f2cff0bc6813ceaba068b977d90eb0e5d6c7e"} Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.593278 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.599230 4997 generic.go:334] "Generic (PLEG): container finished" podID="f692422b-da9f-4775-825e-ee198e5b61b8" containerID="69e70f6cd56b05b725830aa87f03ff2323fde92fedc1b17a8fc26c0917b692da" exitCode=0 Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.599324 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" event={"ID":"f692422b-da9f-4775-825e-ee198e5b61b8","Type":"ContainerDied","Data":"69e70f6cd56b05b725830aa87f03ff2323fde92fedc1b17a8fc26c0917b692da"} Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.612286 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" event={"ID":"cd09450e-798d-41d5-b925-90ab15df8dac","Type":"ContainerStarted","Data":"d489a9092e2cdddebf8931a4ffd9ead2066e0f6d2c9185c1800bd007548a2afb"} Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.612339 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" event={"ID":"cd09450e-798d-41d5-b925-90ab15df8dac","Type":"ContainerStarted","Data":"dd16f29590ca291a136e4ff7ad0143adfb64361ee01979c09a146a027afcd726"} Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.613049 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.623794 4997 generic.go:334] "Generic (PLEG): container finished" podID="c0ca09da-0456-4ab9-8260-03c64893a6dd" containerID="88139d49c80dee570c01f582f9d8414dd3cdde56cd4dbdbc2e2eab5539c440ee" exitCode=0 Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.624762 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6tmw" event={"ID":"c0ca09da-0456-4ab9-8260-03c64893a6dd","Type":"ContainerDied","Data":"88139d49c80dee570c01f582f9d8414dd3cdde56cd4dbdbc2e2eab5539c440ee"} Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.624785 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6tmw" event={"ID":"c0ca09da-0456-4ab9-8260-03c64893a6dd","Type":"ContainerStarted","Data":"fe654c25c33b74a9a9df34f0abf0ce44035a96ac33fb0b5b2ec6def7b4510fac"} Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.677849 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cssd2"] Dec 05 06:57:35 crc kubenswrapper[4997]: W1205 06:57:35.692905 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fd5485d_8c41_484e_9373_4624f05a0f56.slice/crio-6e27ce24291bc9b6844f40ee3423a13df8c2d0f28048ead46b869b99a1ffefc1 WatchSource:0}: Error finding container 6e27ce24291bc9b6844f40ee3423a13df8c2d0f28048ead46b869b99a1ffefc1: Status 404 returned error can't find the container with id 6e27ce24291bc9b6844f40ee3423a13df8c2d0f28048ead46b869b99a1ffefc1 Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.712834 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" podStartSLOduration=136.71282025 podStartE2EDuration="2m16.71282025s" podCreationTimestamp="2025-12-05 06:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:35.687779492 +0000 UTC m=+156.216686763" watchObservedRunningTime="2025-12-05 06:57:35.71282025 +0000 UTC m=+156.241727511" Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.763984 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 05 06:57:35 crc kubenswrapper[4997]: I1205 06:57:35.773464 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5hfgl"] Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.338409 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.339002 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.350514 4997 patch_prober.go:28] interesting pod/console-f9d7485db-lskhh container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.350688 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-lskhh" podUID="0160154e-35ae-471c-b848-214b9f33ec93" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.396968 4997 patch_prober.go:28] interesting pod/router-default-5444994796-v4mxn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 06:57:36 crc kubenswrapper[4997]: [-]has-synced failed: reason withheld Dec 05 06:57:36 crc kubenswrapper[4997]: [+]process-running ok Dec 05 06:57:36 crc kubenswrapper[4997]: healthz check failed Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.397026 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v4mxn" podUID="560926cd-d0fe-4823-b217-803ba02d2702" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.470220 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.470318 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.540375 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.569835 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6zt6h"] Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.576126 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6zt6h" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.579902 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6zt6h"] Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.583125 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.636174 4997 generic.go:334] "Generic (PLEG): container finished" podID="aa0cd037-9961-467b-a659-83922f8ed830" containerID="3174af97edd06b4ca2391bfa21badaf2c56d284cdba5792990bb4259b9a5a3d0" exitCode=0 Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.636259 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hfgl" event={"ID":"aa0cd037-9961-467b-a659-83922f8ed830","Type":"ContainerDied","Data":"3174af97edd06b4ca2391bfa21badaf2c56d284cdba5792990bb4259b9a5a3d0"} Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.636513 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hfgl" event={"ID":"aa0cd037-9961-467b-a659-83922f8ed830","Type":"ContainerStarted","Data":"650b42b244b24a569ecf470122e2ebb60bfa943dac234597e9c9b6eb358a23a9"} Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.642490 4997 generic.go:334] "Generic (PLEG): container finished" podID="7fd5485d-8c41-484e-9373-4624f05a0f56" containerID="9d924c2e9a88691bbeef9d9010647afbaf92684dccb8443f3964823cac62a347" exitCode=0 Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.643669 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cssd2" event={"ID":"7fd5485d-8c41-484e-9373-4624f05a0f56","Type":"ContainerDied","Data":"9d924c2e9a88691bbeef9d9010647afbaf92684dccb8443f3964823cac62a347"} Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.643746 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cssd2" event={"ID":"7fd5485d-8c41-484e-9373-4624f05a0f56","Type":"ContainerStarted","Data":"6e27ce24291bc9b6844f40ee3423a13df8c2d0f28048ead46b869b99a1ffefc1"} Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.647996 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-5b8mm" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.669400 4997 patch_prober.go:28] interesting pod/downloads-7954f5f757-67z92 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.669723 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-67z92" podUID="0836455d-d750-4a39-a7cc-6e95fe3e0abd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.670332 4997 patch_prober.go:28] interesting pod/downloads-7954f5f757-67z92 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.670403 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-67z92" podUID="0836455d-d750-4a39-a7cc-6e95fe3e0abd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.715887 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmfbl\" (UniqueName: \"kubernetes.io/projected/19ea766c-8ab7-4499-a51b-cd6968a62a76-kube-api-access-qmfbl\") pod \"redhat-marketplace-6zt6h\" (UID: \"19ea766c-8ab7-4499-a51b-cd6968a62a76\") " pod="openshift-marketplace/redhat-marketplace-6zt6h" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.715963 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19ea766c-8ab7-4499-a51b-cd6968a62a76-catalog-content\") pod \"redhat-marketplace-6zt6h\" (UID: \"19ea766c-8ab7-4499-a51b-cd6968a62a76\") " pod="openshift-marketplace/redhat-marketplace-6zt6h" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.716050 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19ea766c-8ab7-4499-a51b-cd6968a62a76-utilities\") pod \"redhat-marketplace-6zt6h\" (UID: \"19ea766c-8ab7-4499-a51b-cd6968a62a76\") " pod="openshift-marketplace/redhat-marketplace-6zt6h" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.824470 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmfbl\" (UniqueName: \"kubernetes.io/projected/19ea766c-8ab7-4499-a51b-cd6968a62a76-kube-api-access-qmfbl\") pod \"redhat-marketplace-6zt6h\" (UID: \"19ea766c-8ab7-4499-a51b-cd6968a62a76\") " pod="openshift-marketplace/redhat-marketplace-6zt6h" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.824538 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19ea766c-8ab7-4499-a51b-cd6968a62a76-catalog-content\") pod \"redhat-marketplace-6zt6h\" (UID: \"19ea766c-8ab7-4499-a51b-cd6968a62a76\") " pod="openshift-marketplace/redhat-marketplace-6zt6h" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.824661 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19ea766c-8ab7-4499-a51b-cd6968a62a76-utilities\") pod \"redhat-marketplace-6zt6h\" (UID: \"19ea766c-8ab7-4499-a51b-cd6968a62a76\") " pod="openshift-marketplace/redhat-marketplace-6zt6h" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.825399 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19ea766c-8ab7-4499-a51b-cd6968a62a76-utilities\") pod \"redhat-marketplace-6zt6h\" (UID: \"19ea766c-8ab7-4499-a51b-cd6968a62a76\") " pod="openshift-marketplace/redhat-marketplace-6zt6h" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.825531 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19ea766c-8ab7-4499-a51b-cd6968a62a76-catalog-content\") pod \"redhat-marketplace-6zt6h\" (UID: \"19ea766c-8ab7-4499-a51b-cd6968a62a76\") " pod="openshift-marketplace/redhat-marketplace-6zt6h" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.888632 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmfbl\" (UniqueName: \"kubernetes.io/projected/19ea766c-8ab7-4499-a51b-cd6968a62a76-kube-api-access-qmfbl\") pod \"redhat-marketplace-6zt6h\" (UID: \"19ea766c-8ab7-4499-a51b-cd6968a62a76\") " pod="openshift-marketplace/redhat-marketplace-6zt6h" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.907220 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6zt6h" Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.995899 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jmdrt"] Dec 05 06:57:36 crc kubenswrapper[4997]: I1205 06:57:36.996928 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jmdrt" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.030332 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4cb14f4-531c-4540-836d-3cd359b68be9-utilities\") pod \"redhat-marketplace-jmdrt\" (UID: \"d4cb14f4-531c-4540-836d-3cd359b68be9\") " pod="openshift-marketplace/redhat-marketplace-jmdrt" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.030376 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4cb14f4-531c-4540-836d-3cd359b68be9-catalog-content\") pod \"redhat-marketplace-jmdrt\" (UID: \"d4cb14f4-531c-4540-836d-3cd359b68be9\") " pod="openshift-marketplace/redhat-marketplace-jmdrt" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.030408 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rdkv\" (UniqueName: \"kubernetes.io/projected/d4cb14f4-531c-4540-836d-3cd359b68be9-kube-api-access-8rdkv\") pod \"redhat-marketplace-jmdrt\" (UID: \"d4cb14f4-531c-4540-836d-3cd359b68be9\") " pod="openshift-marketplace/redhat-marketplace-jmdrt" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.052374 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jmdrt"] Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.132016 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rdkv\" (UniqueName: \"kubernetes.io/projected/d4cb14f4-531c-4540-836d-3cd359b68be9-kube-api-access-8rdkv\") pod \"redhat-marketplace-jmdrt\" (UID: \"d4cb14f4-531c-4540-836d-3cd359b68be9\") " pod="openshift-marketplace/redhat-marketplace-jmdrt" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.132128 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4cb14f4-531c-4540-836d-3cd359b68be9-utilities\") pod \"redhat-marketplace-jmdrt\" (UID: \"d4cb14f4-531c-4540-836d-3cd359b68be9\") " pod="openshift-marketplace/redhat-marketplace-jmdrt" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.132147 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4cb14f4-531c-4540-836d-3cd359b68be9-catalog-content\") pod \"redhat-marketplace-jmdrt\" (UID: \"d4cb14f4-531c-4540-836d-3cd359b68be9\") " pod="openshift-marketplace/redhat-marketplace-jmdrt" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.132561 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4cb14f4-531c-4540-836d-3cd359b68be9-catalog-content\") pod \"redhat-marketplace-jmdrt\" (UID: \"d4cb14f4-531c-4540-836d-3cd359b68be9\") " pod="openshift-marketplace/redhat-marketplace-jmdrt" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.133470 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4cb14f4-531c-4540-836d-3cd359b68be9-utilities\") pod \"redhat-marketplace-jmdrt\" (UID: \"d4cb14f4-531c-4540-836d-3cd359b68be9\") " pod="openshift-marketplace/redhat-marketplace-jmdrt" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.152234 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rdkv\" (UniqueName: \"kubernetes.io/projected/d4cb14f4-531c-4540-836d-3cd359b68be9-kube-api-access-8rdkv\") pod \"redhat-marketplace-jmdrt\" (UID: \"d4cb14f4-531c-4540-836d-3cd359b68be9\") " pod="openshift-marketplace/redhat-marketplace-jmdrt" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.160046 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.234435 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbng5\" (UniqueName: \"kubernetes.io/projected/f692422b-da9f-4775-825e-ee198e5b61b8-kube-api-access-kbng5\") pod \"f692422b-da9f-4775-825e-ee198e5b61b8\" (UID: \"f692422b-da9f-4775-825e-ee198e5b61b8\") " Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.234578 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f692422b-da9f-4775-825e-ee198e5b61b8-config-volume\") pod \"f692422b-da9f-4775-825e-ee198e5b61b8\" (UID: \"f692422b-da9f-4775-825e-ee198e5b61b8\") " Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.234623 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f692422b-da9f-4775-825e-ee198e5b61b8-secret-volume\") pod \"f692422b-da9f-4775-825e-ee198e5b61b8\" (UID: \"f692422b-da9f-4775-825e-ee198e5b61b8\") " Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.235602 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f692422b-da9f-4775-825e-ee198e5b61b8-config-volume" (OuterVolumeSpecName: "config-volume") pod "f692422b-da9f-4775-825e-ee198e5b61b8" (UID: "f692422b-da9f-4775-825e-ee198e5b61b8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.242893 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f692422b-da9f-4775-825e-ee198e5b61b8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f692422b-da9f-4775-825e-ee198e5b61b8" (UID: "f692422b-da9f-4775-825e-ee198e5b61b8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.252441 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f692422b-da9f-4775-825e-ee198e5b61b8-kube-api-access-kbng5" (OuterVolumeSpecName: "kube-api-access-kbng5") pod "f692422b-da9f-4775-825e-ee198e5b61b8" (UID: "f692422b-da9f-4775-825e-ee198e5b61b8"). InnerVolumeSpecName "kube-api-access-kbng5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.336660 4997 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f692422b-da9f-4775-825e-ee198e5b61b8-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.337039 4997 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f692422b-da9f-4775-825e-ee198e5b61b8-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.337051 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbng5\" (UniqueName: \"kubernetes.io/projected/f692422b-da9f-4775-825e-ee198e5b61b8-kube-api-access-kbng5\") on node \"crc\" DevicePath \"\"" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.375675 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jmdrt" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.389060 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.393667 4997 patch_prober.go:28] interesting pod/router-default-5444994796-v4mxn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 06:57:37 crc kubenswrapper[4997]: [-]has-synced failed: reason withheld Dec 05 06:57:37 crc kubenswrapper[4997]: [+]process-running ok Dec 05 06:57:37 crc kubenswrapper[4997]: healthz check failed Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.393709 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v4mxn" podUID="560926cd-d0fe-4823-b217-803ba02d2702" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.458593 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.500871 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6zt6h"] Dec 05 06:57:37 crc kubenswrapper[4997]: W1205 06:57:37.542291 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19ea766c_8ab7_4499_a51b_cd6968a62a76.slice/crio-d9e165880b88eb79dad36357ea95ee4cfbd2ec5e4d25341631943616d98e3e28 WatchSource:0}: Error finding container d9e165880b88eb79dad36357ea95ee4cfbd2ec5e4d25341631943616d98e3e28: Status 404 returned error can't find the container with id d9e165880b88eb79dad36357ea95ee4cfbd2ec5e4d25341631943616d98e3e28 Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.560693 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sqj49"] Dec 05 06:57:37 crc kubenswrapper[4997]: E1205 06:57:37.561037 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f692422b-da9f-4775-825e-ee198e5b61b8" containerName="collect-profiles" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.561053 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f692422b-da9f-4775-825e-ee198e5b61b8" containerName="collect-profiles" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.561211 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f692422b-da9f-4775-825e-ee198e5b61b8" containerName="collect-profiles" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.562337 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sqj49" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.564995 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.572281 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sqj49"] Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.644476 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f67be9f3-d0bd-439c-ab66-72a6a29b3104-utilities\") pod \"redhat-operators-sqj49\" (UID: \"f67be9f3-d0bd-439c-ab66-72a6a29b3104\") " pod="openshift-marketplace/redhat-operators-sqj49" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.644555 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdh4n\" (UniqueName: \"kubernetes.io/projected/f67be9f3-d0bd-439c-ab66-72a6a29b3104-kube-api-access-qdh4n\") pod \"redhat-operators-sqj49\" (UID: \"f67be9f3-d0bd-439c-ab66-72a6a29b3104\") " pod="openshift-marketplace/redhat-operators-sqj49" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.644637 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f67be9f3-d0bd-439c-ab66-72a6a29b3104-catalog-content\") pod \"redhat-operators-sqj49\" (UID: \"f67be9f3-d0bd-439c-ab66-72a6a29b3104\") " pod="openshift-marketplace/redhat-operators-sqj49" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.659650 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" event={"ID":"f692422b-da9f-4775-825e-ee198e5b61b8","Type":"ContainerDied","Data":"ce7145efdab5eb6e4044397dcb768c91cff7fa1a145c0a6f307d0651b57b90df"} Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.659691 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce7145efdab5eb6e4044397dcb768c91cff7fa1a145c0a6f307d0651b57b90df" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.659749 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.665322 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zt6h" event={"ID":"19ea766c-8ab7-4499-a51b-cd6968a62a76","Type":"ContainerStarted","Data":"d9e165880b88eb79dad36357ea95ee4cfbd2ec5e4d25341631943616d98e3e28"} Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.753392 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f67be9f3-d0bd-439c-ab66-72a6a29b3104-utilities\") pod \"redhat-operators-sqj49\" (UID: \"f67be9f3-d0bd-439c-ab66-72a6a29b3104\") " pod="openshift-marketplace/redhat-operators-sqj49" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.754192 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdh4n\" (UniqueName: \"kubernetes.io/projected/f67be9f3-d0bd-439c-ab66-72a6a29b3104-kube-api-access-qdh4n\") pod \"redhat-operators-sqj49\" (UID: \"f67be9f3-d0bd-439c-ab66-72a6a29b3104\") " pod="openshift-marketplace/redhat-operators-sqj49" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.754562 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f67be9f3-d0bd-439c-ab66-72a6a29b3104-catalog-content\") pod \"redhat-operators-sqj49\" (UID: \"f67be9f3-d0bd-439c-ab66-72a6a29b3104\") " pod="openshift-marketplace/redhat-operators-sqj49" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.754119 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f67be9f3-d0bd-439c-ab66-72a6a29b3104-utilities\") pod \"redhat-operators-sqj49\" (UID: \"f67be9f3-d0bd-439c-ab66-72a6a29b3104\") " pod="openshift-marketplace/redhat-operators-sqj49" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.760536 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f67be9f3-d0bd-439c-ab66-72a6a29b3104-catalog-content\") pod \"redhat-operators-sqj49\" (UID: \"f67be9f3-d0bd-439c-ab66-72a6a29b3104\") " pod="openshift-marketplace/redhat-operators-sqj49" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.777690 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdh4n\" (UniqueName: \"kubernetes.io/projected/f67be9f3-d0bd-439c-ab66-72a6a29b3104-kube-api-access-qdh4n\") pod \"redhat-operators-sqj49\" (UID: \"f67be9f3-d0bd-439c-ab66-72a6a29b3104\") " pod="openshift-marketplace/redhat-operators-sqj49" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.778474 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jmdrt"] Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.886258 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sqj49" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.963624 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dx8th"] Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.965112 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dx8th" Dec 05 06:57:37 crc kubenswrapper[4997]: I1205 06:57:37.977447 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dx8th"] Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.060887 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-utilities\") pod \"redhat-operators-dx8th\" (UID: \"dc3cd3bb-c155-44a7-a99f-f69d095b36e9\") " pod="openshift-marketplace/redhat-operators-dx8th" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.060934 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-catalog-content\") pod \"redhat-operators-dx8th\" (UID: \"dc3cd3bb-c155-44a7-a99f-f69d095b36e9\") " pod="openshift-marketplace/redhat-operators-dx8th" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.061011 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb6tc\" (UniqueName: \"kubernetes.io/projected/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-kube-api-access-tb6tc\") pod \"redhat-operators-dx8th\" (UID: \"dc3cd3bb-c155-44a7-a99f-f69d095b36e9\") " pod="openshift-marketplace/redhat-operators-dx8th" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.163183 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-utilities\") pod \"redhat-operators-dx8th\" (UID: \"dc3cd3bb-c155-44a7-a99f-f69d095b36e9\") " pod="openshift-marketplace/redhat-operators-dx8th" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.163244 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-catalog-content\") pod \"redhat-operators-dx8th\" (UID: \"dc3cd3bb-c155-44a7-a99f-f69d095b36e9\") " pod="openshift-marketplace/redhat-operators-dx8th" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.163397 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb6tc\" (UniqueName: \"kubernetes.io/projected/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-kube-api-access-tb6tc\") pod \"redhat-operators-dx8th\" (UID: \"dc3cd3bb-c155-44a7-a99f-f69d095b36e9\") " pod="openshift-marketplace/redhat-operators-dx8th" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.164915 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-utilities\") pod \"redhat-operators-dx8th\" (UID: \"dc3cd3bb-c155-44a7-a99f-f69d095b36e9\") " pod="openshift-marketplace/redhat-operators-dx8th" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.167215 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-catalog-content\") pod \"redhat-operators-dx8th\" (UID: \"dc3cd3bb-c155-44a7-a99f-f69d095b36e9\") " pod="openshift-marketplace/redhat-operators-dx8th" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.183581 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb6tc\" (UniqueName: \"kubernetes.io/projected/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-kube-api-access-tb6tc\") pod \"redhat-operators-dx8th\" (UID: \"dc3cd3bb-c155-44a7-a99f-f69d095b36e9\") " pod="openshift-marketplace/redhat-operators-dx8th" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.332740 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.346274 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.346404 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.353865 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.354157 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.369224 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ad0fd5ce-f3e6-491e-9ea6-64c19409327c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ad0fd5ce-f3e6-491e-9ea6-64c19409327c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.369286 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ad0fd5ce-f3e6-491e-9ea6-64c19409327c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ad0fd5ce-f3e6-491e-9ea6-64c19409327c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.393455 4997 patch_prober.go:28] interesting pod/router-default-5444994796-v4mxn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 06:57:38 crc kubenswrapper[4997]: [-]has-synced failed: reason withheld Dec 05 06:57:38 crc kubenswrapper[4997]: [+]process-running ok Dec 05 06:57:38 crc kubenswrapper[4997]: healthz check failed Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.393504 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v4mxn" podUID="560926cd-d0fe-4823-b217-803ba02d2702" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.472661 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ad0fd5ce-f3e6-491e-9ea6-64c19409327c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ad0fd5ce-f3e6-491e-9ea6-64c19409327c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.472708 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ad0fd5ce-f3e6-491e-9ea6-64c19409327c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ad0fd5ce-f3e6-491e-9ea6-64c19409327c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.473100 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ad0fd5ce-f3e6-491e-9ea6-64c19409327c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ad0fd5ce-f3e6-491e-9ea6-64c19409327c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.473440 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dx8th" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.501159 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ad0fd5ce-f3e6-491e-9ea6-64c19409327c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ad0fd5ce-f3e6-491e-9ea6-64c19409327c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.561714 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sqj49"] Dec 05 06:57:38 crc kubenswrapper[4997]: W1205 06:57:38.628534 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf67be9f3_d0bd_439c_ab66_72a6a29b3104.slice/crio-dcc1955a3a7954ed1e2a9c92876736cd8c77e3ece2c376570f0b0cca1bc294b7 WatchSource:0}: Error finding container dcc1955a3a7954ed1e2a9c92876736cd8c77e3ece2c376570f0b0cca1bc294b7: Status 404 returned error can't find the container with id dcc1955a3a7954ed1e2a9c92876736cd8c77e3ece2c376570f0b0cca1bc294b7 Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.679078 4997 generic.go:334] "Generic (PLEG): container finished" podID="19ea766c-8ab7-4499-a51b-cd6968a62a76" containerID="8aa4e42c3fc0a15620790818f529b5aeb899bb0c726805e692b09bb9f940209a" exitCode=0 Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.679162 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zt6h" event={"ID":"19ea766c-8ab7-4499-a51b-cd6968a62a76","Type":"ContainerDied","Data":"8aa4e42c3fc0a15620790818f529b5aeb899bb0c726805e692b09bb9f940209a"} Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.680748 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqj49" event={"ID":"f67be9f3-d0bd-439c-ab66-72a6a29b3104","Type":"ContainerStarted","Data":"dcc1955a3a7954ed1e2a9c92876736cd8c77e3ece2c376570f0b0cca1bc294b7"} Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.692835 4997 generic.go:334] "Generic (PLEG): container finished" podID="d4cb14f4-531c-4540-836d-3cd359b68be9" containerID="46139ad4d7fad459e8f2cc00b9dd0f5214db555de63812d56911adb28224c2a7" exitCode=0 Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.693332 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jmdrt" event={"ID":"d4cb14f4-531c-4540-836d-3cd359b68be9","Type":"ContainerDied","Data":"46139ad4d7fad459e8f2cc00b9dd0f5214db555de63812d56911adb28224c2a7"} Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.693360 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jmdrt" event={"ID":"d4cb14f4-531c-4540-836d-3cd359b68be9","Type":"ContainerStarted","Data":"71613afd69ac6c9967265c8d0680d568ad00d51a42fb435be0fd984cf50a55b5"} Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.700722 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 06:57:38 crc kubenswrapper[4997]: I1205 06:57:38.910830 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dx8th"] Dec 05 06:57:38 crc kubenswrapper[4997]: W1205 06:57:38.943034 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc3cd3bb_c155_44a7_a99f_f69d095b36e9.slice/crio-b7ce4b802cc9fe738b0edc996a212cc40bedd3636e446892e21d4a365bee831b WatchSource:0}: Error finding container b7ce4b802cc9fe738b0edc996a212cc40bedd3636e446892e21d4a365bee831b: Status 404 returned error can't find the container with id b7ce4b802cc9fe738b0edc996a212cc40bedd3636e446892e21d4a365bee831b Dec 05 06:57:39 crc kubenswrapper[4997]: I1205 06:57:39.171656 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 05 06:57:39 crc kubenswrapper[4997]: I1205 06:57:39.392604 4997 patch_prober.go:28] interesting pod/router-default-5444994796-v4mxn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 06:57:39 crc kubenswrapper[4997]: [+]has-synced ok Dec 05 06:57:39 crc kubenswrapper[4997]: [+]process-running ok Dec 05 06:57:39 crc kubenswrapper[4997]: healthz check failed Dec 05 06:57:39 crc kubenswrapper[4997]: I1205 06:57:39.392797 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v4mxn" podUID="560926cd-d0fe-4823-b217-803ba02d2702" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 06:57:39 crc kubenswrapper[4997]: I1205 06:57:39.702554 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ad0fd5ce-f3e6-491e-9ea6-64c19409327c","Type":"ContainerStarted","Data":"d3d0e865eaa5985933e8d7209f55f7a1c54212caa74c0b160189cc274a6f7852"} Dec 05 06:57:39 crc kubenswrapper[4997]: I1205 06:57:39.705056 4997 generic.go:334] "Generic (PLEG): container finished" podID="f67be9f3-d0bd-439c-ab66-72a6a29b3104" containerID="e0d79dda4df3f7129af8e06d8e2d17503ae56ed340b4833f0e360903e10a69bc" exitCode=0 Dec 05 06:57:39 crc kubenswrapper[4997]: I1205 06:57:39.705237 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqj49" event={"ID":"f67be9f3-d0bd-439c-ab66-72a6a29b3104","Type":"ContainerDied","Data":"e0d79dda4df3f7129af8e06d8e2d17503ae56ed340b4833f0e360903e10a69bc"} Dec 05 06:57:39 crc kubenswrapper[4997]: I1205 06:57:39.746173 4997 generic.go:334] "Generic (PLEG): container finished" podID="dc3cd3bb-c155-44a7-a99f-f69d095b36e9" containerID="b8f93ce726e7b888c80953194a54ceb5acca65ff8395559196beb12b30c2f8ef" exitCode=0 Dec 05 06:57:39 crc kubenswrapper[4997]: I1205 06:57:39.746216 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dx8th" event={"ID":"dc3cd3bb-c155-44a7-a99f-f69d095b36e9","Type":"ContainerDied","Data":"b8f93ce726e7b888c80953194a54ceb5acca65ff8395559196beb12b30c2f8ef"} Dec 05 06:57:39 crc kubenswrapper[4997]: I1205 06:57:39.746241 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dx8th" event={"ID":"dc3cd3bb-c155-44a7-a99f-f69d095b36e9","Type":"ContainerStarted","Data":"b7ce4b802cc9fe738b0edc996a212cc40bedd3636e446892e21d4a365bee831b"} Dec 05 06:57:39 crc kubenswrapper[4997]: I1205 06:57:39.940743 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 05 06:57:39 crc kubenswrapper[4997]: I1205 06:57:39.944865 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 06:57:39 crc kubenswrapper[4997]: I1205 06:57:39.950603 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 05 06:57:39 crc kubenswrapper[4997]: I1205 06:57:39.951427 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 05 06:57:39 crc kubenswrapper[4997]: I1205 06:57:39.968434 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 05 06:57:40 crc kubenswrapper[4997]: I1205 06:57:40.023127 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/28a563a7-928d-4835-bf50-26b75b935f93-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"28a563a7-928d-4835-bf50-26b75b935f93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 06:57:40 crc kubenswrapper[4997]: I1205 06:57:40.023389 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28a563a7-928d-4835-bf50-26b75b935f93-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"28a563a7-928d-4835-bf50-26b75b935f93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 06:57:40 crc kubenswrapper[4997]: I1205 06:57:40.124653 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28a563a7-928d-4835-bf50-26b75b935f93-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"28a563a7-928d-4835-bf50-26b75b935f93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 06:57:40 crc kubenswrapper[4997]: I1205 06:57:40.124733 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/28a563a7-928d-4835-bf50-26b75b935f93-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"28a563a7-928d-4835-bf50-26b75b935f93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 06:57:40 crc kubenswrapper[4997]: I1205 06:57:40.125146 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/28a563a7-928d-4835-bf50-26b75b935f93-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"28a563a7-928d-4835-bf50-26b75b935f93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 06:57:40 crc kubenswrapper[4997]: I1205 06:57:40.170515 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28a563a7-928d-4835-bf50-26b75b935f93-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"28a563a7-928d-4835-bf50-26b75b935f93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 06:57:40 crc kubenswrapper[4997]: I1205 06:57:40.250303 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs\") pod \"network-metrics-daemon-b2vlg\" (UID: \"81f6790f-312b-42ce-9bd4-58bac09ce615\") " pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:57:40 crc kubenswrapper[4997]: I1205 06:57:40.259141 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f6790f-312b-42ce-9bd4-58bac09ce615-metrics-certs\") pod \"network-metrics-daemon-b2vlg\" (UID: \"81f6790f-312b-42ce-9bd4-58bac09ce615\") " pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:57:40 crc kubenswrapper[4997]: I1205 06:57:40.291238 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 06:57:40 crc kubenswrapper[4997]: I1205 06:57:40.373665 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b2vlg" Dec 05 06:57:40 crc kubenswrapper[4997]: I1205 06:57:40.409819 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:40 crc kubenswrapper[4997]: I1205 06:57:40.434493 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-v4mxn" Dec 05 06:57:40 crc kubenswrapper[4997]: I1205 06:57:40.763318 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ad0fd5ce-f3e6-491e-9ea6-64c19409327c","Type":"ContainerStarted","Data":"755dfe40ac1df21e62b31f054976816e26e0e177f96037c38bef85f4b9523ec1"} Dec 05 06:57:40 crc kubenswrapper[4997]: I1205 06:57:40.801200 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.801173127 podStartE2EDuration="2.801173127s" podCreationTimestamp="2025-12-05 06:57:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:40.795262462 +0000 UTC m=+161.324169723" watchObservedRunningTime="2025-12-05 06:57:40.801173127 +0000 UTC m=+161.330080388" Dec 05 06:57:40 crc kubenswrapper[4997]: I1205 06:57:40.842993 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-b2vlg"] Dec 05 06:57:41 crc kubenswrapper[4997]: I1205 06:57:41.028875 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 05 06:57:41 crc kubenswrapper[4997]: I1205 06:57:41.927278 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" event={"ID":"81f6790f-312b-42ce-9bd4-58bac09ce615","Type":"ContainerStarted","Data":"57cddc016bc68a92caa485d3ee0803e8d675fa4065344032259dcaabe4f11ffe"} Dec 05 06:57:41 crc kubenswrapper[4997]: I1205 06:57:41.941789 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ad0fd5ce-f3e6-491e-9ea6-64c19409327c","Type":"ContainerDied","Data":"755dfe40ac1df21e62b31f054976816e26e0e177f96037c38bef85f4b9523ec1"} Dec 05 06:57:41 crc kubenswrapper[4997]: I1205 06:57:41.941906 4997 generic.go:334] "Generic (PLEG): container finished" podID="ad0fd5ce-f3e6-491e-9ea6-64c19409327c" containerID="755dfe40ac1df21e62b31f054976816e26e0e177f96037c38bef85f4b9523ec1" exitCode=0 Dec 05 06:57:41 crc kubenswrapper[4997]: I1205 06:57:41.960578 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"28a563a7-928d-4835-bf50-26b75b935f93","Type":"ContainerStarted","Data":"292e3e5e7299a43b585b7edaaa1848f5d4e57827c8a3ca45fa58684bfc77243b"} Dec 05 06:57:42 crc kubenswrapper[4997]: I1205 06:57:42.553222 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-zcnpm" Dec 05 06:57:43 crc kubenswrapper[4997]: I1205 06:57:43.006687 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" event={"ID":"81f6790f-312b-42ce-9bd4-58bac09ce615","Type":"ContainerStarted","Data":"e76b1ebad4259811344a15623a525729a1e9019fea8acfd72a33305e4cdd40a0"} Dec 05 06:57:43 crc kubenswrapper[4997]: I1205 06:57:43.017844 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"28a563a7-928d-4835-bf50-26b75b935f93","Type":"ContainerStarted","Data":"c9980b289cea0d2f989bb83924de37aec1ed877b9ecf8a4da5ed4369f962740f"} Dec 05 06:57:43 crc kubenswrapper[4997]: I1205 06:57:43.055088 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.05506962 podStartE2EDuration="4.05506962s" podCreationTimestamp="2025-12-05 06:57:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:43.049510773 +0000 UTC m=+163.578418054" watchObservedRunningTime="2025-12-05 06:57:43.05506962 +0000 UTC m=+163.583976881" Dec 05 06:57:43 crc kubenswrapper[4997]: I1205 06:57:43.548821 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 06:57:43 crc kubenswrapper[4997]: I1205 06:57:43.629404 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ad0fd5ce-f3e6-491e-9ea6-64c19409327c-kube-api-access\") pod \"ad0fd5ce-f3e6-491e-9ea6-64c19409327c\" (UID: \"ad0fd5ce-f3e6-491e-9ea6-64c19409327c\") " Dec 05 06:57:43 crc kubenswrapper[4997]: I1205 06:57:43.629470 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ad0fd5ce-f3e6-491e-9ea6-64c19409327c-kubelet-dir\") pod \"ad0fd5ce-f3e6-491e-9ea6-64c19409327c\" (UID: \"ad0fd5ce-f3e6-491e-9ea6-64c19409327c\") " Dec 05 06:57:43 crc kubenswrapper[4997]: I1205 06:57:43.629711 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ad0fd5ce-f3e6-491e-9ea6-64c19409327c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ad0fd5ce-f3e6-491e-9ea6-64c19409327c" (UID: "ad0fd5ce-f3e6-491e-9ea6-64c19409327c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 06:57:43 crc kubenswrapper[4997]: I1205 06:57:43.650794 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad0fd5ce-f3e6-491e-9ea6-64c19409327c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ad0fd5ce-f3e6-491e-9ea6-64c19409327c" (UID: "ad0fd5ce-f3e6-491e-9ea6-64c19409327c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:57:43 crc kubenswrapper[4997]: I1205 06:57:43.735889 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ad0fd5ce-f3e6-491e-9ea6-64c19409327c-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 06:57:43 crc kubenswrapper[4997]: I1205 06:57:43.735941 4997 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ad0fd5ce-f3e6-491e-9ea6-64c19409327c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 06:57:44 crc kubenswrapper[4997]: I1205 06:57:44.035097 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b2vlg" event={"ID":"81f6790f-312b-42ce-9bd4-58bac09ce615","Type":"ContainerStarted","Data":"2ce54db4a561fbee5097a22e4c2d527c0aaea2305eb01d12b66ec93301a57c1a"} Dec 05 06:57:44 crc kubenswrapper[4997]: I1205 06:57:44.037470 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ad0fd5ce-f3e6-491e-9ea6-64c19409327c","Type":"ContainerDied","Data":"d3d0e865eaa5985933e8d7209f55f7a1c54212caa74c0b160189cc274a6f7852"} Dec 05 06:57:44 crc kubenswrapper[4997]: I1205 06:57:44.037500 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3d0e865eaa5985933e8d7209f55f7a1c54212caa74c0b160189cc274a6f7852" Dec 05 06:57:44 crc kubenswrapper[4997]: I1205 06:57:44.037571 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 06:57:44 crc kubenswrapper[4997]: I1205 06:57:44.049309 4997 generic.go:334] "Generic (PLEG): container finished" podID="28a563a7-928d-4835-bf50-26b75b935f93" containerID="c9980b289cea0d2f989bb83924de37aec1ed877b9ecf8a4da5ed4369f962740f" exitCode=0 Dec 05 06:57:44 crc kubenswrapper[4997]: I1205 06:57:44.049357 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"28a563a7-928d-4835-bf50-26b75b935f93","Type":"ContainerDied","Data":"c9980b289cea0d2f989bb83924de37aec1ed877b9ecf8a4da5ed4369f962740f"} Dec 05 06:57:45 crc kubenswrapper[4997]: I1205 06:57:45.091959 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-b2vlg" podStartSLOduration=147.091931079 podStartE2EDuration="2m27.091931079s" podCreationTimestamp="2025-12-05 06:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:57:45.087208534 +0000 UTC m=+165.616115815" watchObservedRunningTime="2025-12-05 06:57:45.091931079 +0000 UTC m=+165.620838350" Dec 05 06:57:46 crc kubenswrapper[4997]: I1205 06:57:46.333851 4997 patch_prober.go:28] interesting pod/console-f9d7485db-lskhh container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 05 06:57:46 crc kubenswrapper[4997]: I1205 06:57:46.333949 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-lskhh" podUID="0160154e-35ae-471c-b848-214b9f33ec93" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 05 06:57:46 crc kubenswrapper[4997]: I1205 06:57:46.682397 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-67z92" Dec 05 06:57:49 crc kubenswrapper[4997]: I1205 06:57:49.769957 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 06:57:49 crc kubenswrapper[4997]: I1205 06:57:49.770446 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 06:57:54 crc kubenswrapper[4997]: I1205 06:57:54.554861 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 06:57:54 crc kubenswrapper[4997]: I1205 06:57:54.947640 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 06:57:55 crc kubenswrapper[4997]: I1205 06:57:55.126297 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/28a563a7-928d-4835-bf50-26b75b935f93-kubelet-dir\") pod \"28a563a7-928d-4835-bf50-26b75b935f93\" (UID: \"28a563a7-928d-4835-bf50-26b75b935f93\") " Dec 05 06:57:55 crc kubenswrapper[4997]: I1205 06:57:55.126407 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28a563a7-928d-4835-bf50-26b75b935f93-kube-api-access\") pod \"28a563a7-928d-4835-bf50-26b75b935f93\" (UID: \"28a563a7-928d-4835-bf50-26b75b935f93\") " Dec 05 06:57:55 crc kubenswrapper[4997]: I1205 06:57:55.127592 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28a563a7-928d-4835-bf50-26b75b935f93-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "28a563a7-928d-4835-bf50-26b75b935f93" (UID: "28a563a7-928d-4835-bf50-26b75b935f93"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 06:57:55 crc kubenswrapper[4997]: I1205 06:57:55.131724 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28a563a7-928d-4835-bf50-26b75b935f93-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "28a563a7-928d-4835-bf50-26b75b935f93" (UID: "28a563a7-928d-4835-bf50-26b75b935f93"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:57:55 crc kubenswrapper[4997]: I1205 06:57:55.184716 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"28a563a7-928d-4835-bf50-26b75b935f93","Type":"ContainerDied","Data":"292e3e5e7299a43b585b7edaaa1848f5d4e57827c8a3ca45fa58684bfc77243b"} Dec 05 06:57:55 crc kubenswrapper[4997]: I1205 06:57:55.184759 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="292e3e5e7299a43b585b7edaaa1848f5d4e57827c8a3ca45fa58684bfc77243b" Dec 05 06:57:55 crc kubenswrapper[4997]: I1205 06:57:55.184856 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 06:57:55 crc kubenswrapper[4997]: I1205 06:57:55.227653 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28a563a7-928d-4835-bf50-26b75b935f93-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 06:57:55 crc kubenswrapper[4997]: I1205 06:57:55.227679 4997 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/28a563a7-928d-4835-bf50-26b75b935f93-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 06:57:56 crc kubenswrapper[4997]: I1205 06:57:56.337330 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:57:56 crc kubenswrapper[4997]: I1205 06:57:56.345977 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-lskhh" Dec 05 06:58:07 crc kubenswrapper[4997]: I1205 06:58:07.287459 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 06:58:07 crc kubenswrapper[4997]: I1205 06:58:07.465365 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-s8cx6" Dec 05 06:58:09 crc kubenswrapper[4997]: E1205 06:58:09.307393 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 05 06:58:09 crc kubenswrapper[4997]: E1205 06:58:09.307929 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qmfbl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-6zt6h_openshift-marketplace(19ea766c-8ab7-4499-a51b-cd6968a62a76): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 06:58:09 crc kubenswrapper[4997]: E1205 06:58:09.309179 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-6zt6h" podUID="19ea766c-8ab7-4499-a51b-cd6968a62a76" Dec 05 06:58:13 crc kubenswrapper[4997]: E1205 06:58:13.906691 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-6zt6h" podUID="19ea766c-8ab7-4499-a51b-cd6968a62a76" Dec 05 06:58:14 crc kubenswrapper[4997]: E1205 06:58:14.127501 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 05 06:58:14 crc kubenswrapper[4997]: E1205 06:58:14.127691 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5px8g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5hfgl_openshift-marketplace(aa0cd037-9961-467b-a659-83922f8ed830): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 06:58:14 crc kubenswrapper[4997]: E1205 06:58:14.129529 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-5hfgl" podUID="aa0cd037-9961-467b-a659-83922f8ed830" Dec 05 06:58:14 crc kubenswrapper[4997]: I1205 06:58:14.147325 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 05 06:58:14 crc kubenswrapper[4997]: E1205 06:58:14.149101 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28a563a7-928d-4835-bf50-26b75b935f93" containerName="pruner" Dec 05 06:58:14 crc kubenswrapper[4997]: I1205 06:58:14.149242 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a563a7-928d-4835-bf50-26b75b935f93" containerName="pruner" Dec 05 06:58:14 crc kubenswrapper[4997]: E1205 06:58:14.149410 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad0fd5ce-f3e6-491e-9ea6-64c19409327c" containerName="pruner" Dec 05 06:58:14 crc kubenswrapper[4997]: I1205 06:58:14.149515 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad0fd5ce-f3e6-491e-9ea6-64c19409327c" containerName="pruner" Dec 05 06:58:14 crc kubenswrapper[4997]: I1205 06:58:14.149945 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="28a563a7-928d-4835-bf50-26b75b935f93" containerName="pruner" Dec 05 06:58:14 crc kubenswrapper[4997]: I1205 06:58:14.150076 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad0fd5ce-f3e6-491e-9ea6-64c19409327c" containerName="pruner" Dec 05 06:58:14 crc kubenswrapper[4997]: I1205 06:58:14.150798 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 06:58:14 crc kubenswrapper[4997]: I1205 06:58:14.153698 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 05 06:58:14 crc kubenswrapper[4997]: I1205 06:58:14.155030 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 05 06:58:14 crc kubenswrapper[4997]: I1205 06:58:14.155753 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 05 06:58:14 crc kubenswrapper[4997]: I1205 06:58:14.322272 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c93e8cb8-7eea-411f-92a5-71d25785226a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c93e8cb8-7eea-411f-92a5-71d25785226a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 06:58:14 crc kubenswrapper[4997]: I1205 06:58:14.322358 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c93e8cb8-7eea-411f-92a5-71d25785226a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c93e8cb8-7eea-411f-92a5-71d25785226a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 06:58:14 crc kubenswrapper[4997]: I1205 06:58:14.423878 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c93e8cb8-7eea-411f-92a5-71d25785226a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c93e8cb8-7eea-411f-92a5-71d25785226a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 06:58:14 crc kubenswrapper[4997]: I1205 06:58:14.423998 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c93e8cb8-7eea-411f-92a5-71d25785226a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c93e8cb8-7eea-411f-92a5-71d25785226a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 06:58:14 crc kubenswrapper[4997]: I1205 06:58:14.424105 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c93e8cb8-7eea-411f-92a5-71d25785226a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c93e8cb8-7eea-411f-92a5-71d25785226a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 06:58:14 crc kubenswrapper[4997]: I1205 06:58:14.459419 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c93e8cb8-7eea-411f-92a5-71d25785226a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c93e8cb8-7eea-411f-92a5-71d25785226a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 06:58:14 crc kubenswrapper[4997]: I1205 06:58:14.466537 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 06:58:14 crc kubenswrapper[4997]: E1205 06:58:14.785480 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 05 06:58:14 crc kubenswrapper[4997]: E1205 06:58:14.785911 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8rdkv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-jmdrt_openshift-marketplace(d4cb14f4-531c-4540-836d-3cd359b68be9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 06:58:14 crc kubenswrapper[4997]: E1205 06:58:14.787262 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-jmdrt" podUID="d4cb14f4-531c-4540-836d-3cd359b68be9" Dec 05 06:58:18 crc kubenswrapper[4997]: I1205 06:58:18.526775 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 05 06:58:18 crc kubenswrapper[4997]: I1205 06:58:18.527882 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 06:58:18 crc kubenswrapper[4997]: I1205 06:58:18.538747 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 05 06:58:18 crc kubenswrapper[4997]: I1205 06:58:18.681874 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6d67d224-c501-4501-89a3-120c9f883811-var-lock\") pod \"installer-9-crc\" (UID: \"6d67d224-c501-4501-89a3-120c9f883811\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 06:58:18 crc kubenswrapper[4997]: I1205 06:58:18.681945 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6d67d224-c501-4501-89a3-120c9f883811-kube-api-access\") pod \"installer-9-crc\" (UID: \"6d67d224-c501-4501-89a3-120c9f883811\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 06:58:18 crc kubenswrapper[4997]: I1205 06:58:18.681978 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6d67d224-c501-4501-89a3-120c9f883811-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6d67d224-c501-4501-89a3-120c9f883811\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 06:58:18 crc kubenswrapper[4997]: I1205 06:58:18.782878 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6d67d224-c501-4501-89a3-120c9f883811-var-lock\") pod \"installer-9-crc\" (UID: \"6d67d224-c501-4501-89a3-120c9f883811\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 06:58:18 crc kubenswrapper[4997]: I1205 06:58:18.782917 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6d67d224-c501-4501-89a3-120c9f883811-kube-api-access\") pod \"installer-9-crc\" (UID: \"6d67d224-c501-4501-89a3-120c9f883811\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 06:58:18 crc kubenswrapper[4997]: I1205 06:58:18.782949 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6d67d224-c501-4501-89a3-120c9f883811-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6d67d224-c501-4501-89a3-120c9f883811\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 06:58:18 crc kubenswrapper[4997]: I1205 06:58:18.783047 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6d67d224-c501-4501-89a3-120c9f883811-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6d67d224-c501-4501-89a3-120c9f883811\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 06:58:18 crc kubenswrapper[4997]: I1205 06:58:18.783082 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6d67d224-c501-4501-89a3-120c9f883811-var-lock\") pod \"installer-9-crc\" (UID: \"6d67d224-c501-4501-89a3-120c9f883811\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 06:58:18 crc kubenswrapper[4997]: I1205 06:58:18.803470 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6d67d224-c501-4501-89a3-120c9f883811-kube-api-access\") pod \"installer-9-crc\" (UID: \"6d67d224-c501-4501-89a3-120c9f883811\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 06:58:18 crc kubenswrapper[4997]: I1205 06:58:18.856832 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 06:58:19 crc kubenswrapper[4997]: I1205 06:58:19.770190 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 06:58:19 crc kubenswrapper[4997]: I1205 06:58:19.770253 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 06:58:19 crc kubenswrapper[4997]: I1205 06:58:19.770327 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 06:58:19 crc kubenswrapper[4997]: I1205 06:58:19.771033 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 06:58:19 crc kubenswrapper[4997]: I1205 06:58:19.771154 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a" gracePeriod=600 Dec 05 06:58:20 crc kubenswrapper[4997]: E1205 06:58:20.774330 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-jmdrt" podUID="d4cb14f4-531c-4540-836d-3cd359b68be9" Dec 05 06:58:20 crc kubenswrapper[4997]: E1205 06:58:20.845222 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 05 06:58:20 crc kubenswrapper[4997]: E1205 06:58:20.845677 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-thxwv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cssd2_openshift-marketplace(7fd5485d-8c41-484e-9373-4624f05a0f56): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 06:58:20 crc kubenswrapper[4997]: E1205 06:58:20.846862 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-cssd2" podUID="7fd5485d-8c41-484e-9373-4624f05a0f56" Dec 05 06:58:21 crc kubenswrapper[4997]: I1205 06:58:21.329864 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a" exitCode=0 Dec 05 06:58:21 crc kubenswrapper[4997]: I1205 06:58:21.329947 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a"} Dec 05 06:58:23 crc kubenswrapper[4997]: E1205 06:58:23.820469 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cssd2" podUID="7fd5485d-8c41-484e-9373-4624f05a0f56" Dec 05 06:58:23 crc kubenswrapper[4997]: E1205 06:58:23.820544 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5hfgl" podUID="aa0cd037-9961-467b-a659-83922f8ed830" Dec 05 06:58:23 crc kubenswrapper[4997]: E1205 06:58:23.852674 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 05 06:58:23 crc kubenswrapper[4997]: E1205 06:58:23.853036 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tb6tc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-dx8th_openshift-marketplace(dc3cd3bb-c155-44a7-a99f-f69d095b36e9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 06:58:23 crc kubenswrapper[4997]: E1205 06:58:23.854418 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-dx8th" podUID="dc3cd3bb-c155-44a7-a99f-f69d095b36e9" Dec 05 06:58:23 crc kubenswrapper[4997]: E1205 06:58:23.910200 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 05 06:58:23 crc kubenswrapper[4997]: E1205 06:58:23.910675 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qdh4n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-sqj49_openshift-marketplace(f67be9f3-d0bd-439c-ab66-72a6a29b3104): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 06:58:23 crc kubenswrapper[4997]: E1205 06:58:23.911872 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-sqj49" podUID="f67be9f3-d0bd-439c-ab66-72a6a29b3104" Dec 05 06:58:23 crc kubenswrapper[4997]: E1205 06:58:23.913147 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 05 06:58:23 crc kubenswrapper[4997]: E1205 06:58:23.913370 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r2s94,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-2pls6_openshift-marketplace(3ec5dfd3-3221-4141-a2d4-aea02733a985): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 06:58:23 crc kubenswrapper[4997]: E1205 06:58:23.914637 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-2pls6" podUID="3ec5dfd3-3221-4141-a2d4-aea02733a985" Dec 05 06:58:23 crc kubenswrapper[4997]: E1205 06:58:23.949437 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 05 06:58:23 crc kubenswrapper[4997]: E1205 06:58:23.950407 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t9vmz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-n6tmw_openshift-marketplace(c0ca09da-0456-4ab9-8260-03c64893a6dd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 06:58:23 crc kubenswrapper[4997]: E1205 06:58:23.952400 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-n6tmw" podUID="c0ca09da-0456-4ab9-8260-03c64893a6dd" Dec 05 06:58:24 crc kubenswrapper[4997]: I1205 06:58:24.069356 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 05 06:58:24 crc kubenswrapper[4997]: I1205 06:58:24.140038 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 05 06:58:24 crc kubenswrapper[4997]: I1205 06:58:24.353554 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"2f6d912e8d30078a8f27c58e585fd50e33dbd4541657a6aef0990fa4ea86b3b6"} Dec 05 06:58:24 crc kubenswrapper[4997]: I1205 06:58:24.355079 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6d67d224-c501-4501-89a3-120c9f883811","Type":"ContainerStarted","Data":"5485311f58b27194e15e61bbd64fc8e7d6063385ad804b0d760accae1d96e9d0"} Dec 05 06:58:24 crc kubenswrapper[4997]: I1205 06:58:24.357106 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c93e8cb8-7eea-411f-92a5-71d25785226a","Type":"ContainerStarted","Data":"f3c95eb9b7b87081c0be9fd24bfef9e1822819d4e2f4e1e0e813b86c6c169007"} Dec 05 06:58:24 crc kubenswrapper[4997]: E1205 06:58:24.358906 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-2pls6" podUID="3ec5dfd3-3221-4141-a2d4-aea02733a985" Dec 05 06:58:24 crc kubenswrapper[4997]: E1205 06:58:24.359215 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-n6tmw" podUID="c0ca09da-0456-4ab9-8260-03c64893a6dd" Dec 05 06:58:24 crc kubenswrapper[4997]: E1205 06:58:24.360448 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-sqj49" podUID="f67be9f3-d0bd-439c-ab66-72a6a29b3104" Dec 05 06:58:24 crc kubenswrapper[4997]: E1205 06:58:24.365553 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-dx8th" podUID="dc3cd3bb-c155-44a7-a99f-f69d095b36e9" Dec 05 06:58:25 crc kubenswrapper[4997]: I1205 06:58:25.371932 4997 generic.go:334] "Generic (PLEG): container finished" podID="c93e8cb8-7eea-411f-92a5-71d25785226a" containerID="b4146b4d232b8abb1c7c1d28062b260a4349b924f3f6da74245283029a2b52f9" exitCode=0 Dec 05 06:58:25 crc kubenswrapper[4997]: I1205 06:58:25.372035 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c93e8cb8-7eea-411f-92a5-71d25785226a","Type":"ContainerDied","Data":"b4146b4d232b8abb1c7c1d28062b260a4349b924f3f6da74245283029a2b52f9"} Dec 05 06:58:25 crc kubenswrapper[4997]: I1205 06:58:25.373841 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6d67d224-c501-4501-89a3-120c9f883811","Type":"ContainerStarted","Data":"0d6b37b511617b0bd20413663db8ea5f3ab1a96ab96b4467cc6d858ebefb1bec"} Dec 05 06:58:25 crc kubenswrapper[4997]: I1205 06:58:25.408358 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=7.408338059 podStartE2EDuration="7.408338059s" podCreationTimestamp="2025-12-05 06:58:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:58:25.407592038 +0000 UTC m=+205.936499309" watchObservedRunningTime="2025-12-05 06:58:25.408338059 +0000 UTC m=+205.937245320" Dec 05 06:58:26 crc kubenswrapper[4997]: I1205 06:58:26.382997 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zt6h" event={"ID":"19ea766c-8ab7-4499-a51b-cd6968a62a76","Type":"ContainerStarted","Data":"072d16b6f5b85a1dc575341b25ee9b439d5cbd8391d9c089da4d9be91d991414"} Dec 05 06:58:26 crc kubenswrapper[4997]: I1205 06:58:26.602434 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 06:58:26 crc kubenswrapper[4997]: I1205 06:58:26.685017 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c93e8cb8-7eea-411f-92a5-71d25785226a-kube-api-access\") pod \"c93e8cb8-7eea-411f-92a5-71d25785226a\" (UID: \"c93e8cb8-7eea-411f-92a5-71d25785226a\") " Dec 05 06:58:26 crc kubenswrapper[4997]: I1205 06:58:26.685062 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c93e8cb8-7eea-411f-92a5-71d25785226a-kubelet-dir\") pod \"c93e8cb8-7eea-411f-92a5-71d25785226a\" (UID: \"c93e8cb8-7eea-411f-92a5-71d25785226a\") " Dec 05 06:58:26 crc kubenswrapper[4997]: I1205 06:58:26.685257 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c93e8cb8-7eea-411f-92a5-71d25785226a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c93e8cb8-7eea-411f-92a5-71d25785226a" (UID: "c93e8cb8-7eea-411f-92a5-71d25785226a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 06:58:26 crc kubenswrapper[4997]: I1205 06:58:26.690085 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c93e8cb8-7eea-411f-92a5-71d25785226a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c93e8cb8-7eea-411f-92a5-71d25785226a" (UID: "c93e8cb8-7eea-411f-92a5-71d25785226a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:58:26 crc kubenswrapper[4997]: I1205 06:58:26.785827 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c93e8cb8-7eea-411f-92a5-71d25785226a-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 06:58:26 crc kubenswrapper[4997]: I1205 06:58:26.786081 4997 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c93e8cb8-7eea-411f-92a5-71d25785226a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 06:58:27 crc kubenswrapper[4997]: I1205 06:58:27.389497 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 06:58:27 crc kubenswrapper[4997]: I1205 06:58:27.389512 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c93e8cb8-7eea-411f-92a5-71d25785226a","Type":"ContainerDied","Data":"f3c95eb9b7b87081c0be9fd24bfef9e1822819d4e2f4e1e0e813b86c6c169007"} Dec 05 06:58:27 crc kubenswrapper[4997]: I1205 06:58:27.390832 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3c95eb9b7b87081c0be9fd24bfef9e1822819d4e2f4e1e0e813b86c6c169007" Dec 05 06:58:27 crc kubenswrapper[4997]: I1205 06:58:27.391639 4997 generic.go:334] "Generic (PLEG): container finished" podID="19ea766c-8ab7-4499-a51b-cd6968a62a76" containerID="072d16b6f5b85a1dc575341b25ee9b439d5cbd8391d9c089da4d9be91d991414" exitCode=0 Dec 05 06:58:27 crc kubenswrapper[4997]: I1205 06:58:27.391671 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zt6h" event={"ID":"19ea766c-8ab7-4499-a51b-cd6968a62a76","Type":"ContainerDied","Data":"072d16b6f5b85a1dc575341b25ee9b439d5cbd8391d9c089da4d9be91d991414"} Dec 05 06:58:28 crc kubenswrapper[4997]: I1205 06:58:28.399797 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zt6h" event={"ID":"19ea766c-8ab7-4499-a51b-cd6968a62a76","Type":"ContainerStarted","Data":"485e08d238a48edaa6a607645ffb2b538173a81e733c867397c153bde926b98c"} Dec 05 06:58:28 crc kubenswrapper[4997]: I1205 06:58:28.420903 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6zt6h" podStartSLOduration=3.332782429 podStartE2EDuration="52.420885117s" podCreationTimestamp="2025-12-05 06:57:36 +0000 UTC" firstStartedPulling="2025-12-05 06:57:38.686626027 +0000 UTC m=+159.215533298" lastFinishedPulling="2025-12-05 06:58:27.774728725 +0000 UTC m=+208.303635986" observedRunningTime="2025-12-05 06:58:28.416723796 +0000 UTC m=+208.945631067" watchObservedRunningTime="2025-12-05 06:58:28.420885117 +0000 UTC m=+208.949792378" Dec 05 06:58:36 crc kubenswrapper[4997]: I1205 06:58:36.907825 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6zt6h" Dec 05 06:58:36 crc kubenswrapper[4997]: I1205 06:58:36.908490 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6zt6h" Dec 05 06:58:37 crc kubenswrapper[4997]: I1205 06:58:37.043544 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6zt6h" Dec 05 06:58:37 crc kubenswrapper[4997]: I1205 06:58:37.453764 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cssd2" event={"ID":"7fd5485d-8c41-484e-9373-4624f05a0f56","Type":"ContainerStarted","Data":"224ed527028f096b800d7afd1880a7b1d9ce6311a4142e67d59309d5d107f0fe"} Dec 05 06:58:37 crc kubenswrapper[4997]: I1205 06:58:37.457440 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jmdrt" event={"ID":"d4cb14f4-531c-4540-836d-3cd359b68be9","Type":"ContainerStarted","Data":"f7282a58d752ffd2cb74f18156a55142843d056f36ebc97f04273d76bb405d0d"} Dec 05 06:58:37 crc kubenswrapper[4997]: I1205 06:58:37.506371 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6zt6h" Dec 05 06:58:38 crc kubenswrapper[4997]: I1205 06:58:38.467345 4997 generic.go:334] "Generic (PLEG): container finished" podID="d4cb14f4-531c-4540-836d-3cd359b68be9" containerID="f7282a58d752ffd2cb74f18156a55142843d056f36ebc97f04273d76bb405d0d" exitCode=0 Dec 05 06:58:38 crc kubenswrapper[4997]: I1205 06:58:38.467424 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jmdrt" event={"ID":"d4cb14f4-531c-4540-836d-3cd359b68be9","Type":"ContainerDied","Data":"f7282a58d752ffd2cb74f18156a55142843d056f36ebc97f04273d76bb405d0d"} Dec 05 06:58:38 crc kubenswrapper[4997]: I1205 06:58:38.473576 4997 generic.go:334] "Generic (PLEG): container finished" podID="3ec5dfd3-3221-4141-a2d4-aea02733a985" containerID="e7f84a21bad53bb6f660d0d6f35db8bf64bdbe637269e05bf17bfac3a68c2ab5" exitCode=0 Dec 05 06:58:38 crc kubenswrapper[4997]: I1205 06:58:38.473646 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2pls6" event={"ID":"3ec5dfd3-3221-4141-a2d4-aea02733a985","Type":"ContainerDied","Data":"e7f84a21bad53bb6f660d0d6f35db8bf64bdbe637269e05bf17bfac3a68c2ab5"} Dec 05 06:58:38 crc kubenswrapper[4997]: I1205 06:58:38.476265 4997 generic.go:334] "Generic (PLEG): container finished" podID="aa0cd037-9961-467b-a659-83922f8ed830" containerID="e705484faae7f53402859cd79406db73334f888b7aa230310b26c16a3f2f773a" exitCode=0 Dec 05 06:58:38 crc kubenswrapper[4997]: I1205 06:58:38.476589 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hfgl" event={"ID":"aa0cd037-9961-467b-a659-83922f8ed830","Type":"ContainerDied","Data":"e705484faae7f53402859cd79406db73334f888b7aa230310b26c16a3f2f773a"} Dec 05 06:58:38 crc kubenswrapper[4997]: I1205 06:58:38.479244 4997 generic.go:334] "Generic (PLEG): container finished" podID="7fd5485d-8c41-484e-9373-4624f05a0f56" containerID="224ed527028f096b800d7afd1880a7b1d9ce6311a4142e67d59309d5d107f0fe" exitCode=0 Dec 05 06:58:38 crc kubenswrapper[4997]: I1205 06:58:38.479396 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cssd2" event={"ID":"7fd5485d-8c41-484e-9373-4624f05a0f56","Type":"ContainerDied","Data":"224ed527028f096b800d7afd1880a7b1d9ce6311a4142e67d59309d5d107f0fe"} Dec 05 06:58:39 crc kubenswrapper[4997]: I1205 06:58:39.488297 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2pls6" event={"ID":"3ec5dfd3-3221-4141-a2d4-aea02733a985","Type":"ContainerStarted","Data":"4e95c9a0122f9f97de73329b76fe8d535a38995655562840fb0aafe6fd7824a1"} Dec 05 06:58:39 crc kubenswrapper[4997]: I1205 06:58:39.491457 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hfgl" event={"ID":"aa0cd037-9961-467b-a659-83922f8ed830","Type":"ContainerStarted","Data":"963332eff503d54d0385710006325ace7449be29c170c20f68f31fc51e41a1bd"} Dec 05 06:58:39 crc kubenswrapper[4997]: I1205 06:58:39.493517 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cssd2" event={"ID":"7fd5485d-8c41-484e-9373-4624f05a0f56","Type":"ContainerStarted","Data":"0c4a0f0e3571591f4f2543d161e46e84758d11dec95d7c51bf2753b0f1f104dd"} Dec 05 06:58:39 crc kubenswrapper[4997]: I1205 06:58:39.497227 4997 generic.go:334] "Generic (PLEG): container finished" podID="c0ca09da-0456-4ab9-8260-03c64893a6dd" containerID="a9229007162d88dc7222c56646b799e696134f1ac031cc24746970e7065276c1" exitCode=0 Dec 05 06:58:39 crc kubenswrapper[4997]: I1205 06:58:39.497276 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6tmw" event={"ID":"c0ca09da-0456-4ab9-8260-03c64893a6dd","Type":"ContainerDied","Data":"a9229007162d88dc7222c56646b799e696134f1ac031cc24746970e7065276c1"} Dec 05 06:58:39 crc kubenswrapper[4997]: I1205 06:58:39.503248 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jmdrt" event={"ID":"d4cb14f4-531c-4540-836d-3cd359b68be9","Type":"ContainerStarted","Data":"2ecaff6c9bf24e4087c79ba6b8be502a603c5b4ee044f951f8ab00076053738a"} Dec 05 06:58:39 crc kubenswrapper[4997]: I1205 06:58:39.524213 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2pls6" podStartSLOduration=2.133009548 podStartE2EDuration="1m5.524185194s" podCreationTimestamp="2025-12-05 06:57:34 +0000 UTC" firstStartedPulling="2025-12-05 06:57:35.593035442 +0000 UTC m=+156.121942703" lastFinishedPulling="2025-12-05 06:58:38.984211088 +0000 UTC m=+219.513118349" observedRunningTime="2025-12-05 06:58:39.519097068 +0000 UTC m=+220.048004329" watchObservedRunningTime="2025-12-05 06:58:39.524185194 +0000 UTC m=+220.053092455" Dec 05 06:58:39 crc kubenswrapper[4997]: I1205 06:58:39.563025 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cssd2" podStartSLOduration=3.114767776 podStartE2EDuration="1m5.563005902s" podCreationTimestamp="2025-12-05 06:57:34 +0000 UTC" firstStartedPulling="2025-12-05 06:57:36.650887488 +0000 UTC m=+157.179794749" lastFinishedPulling="2025-12-05 06:58:39.099125614 +0000 UTC m=+219.628032875" observedRunningTime="2025-12-05 06:58:39.562639353 +0000 UTC m=+220.091546624" watchObservedRunningTime="2025-12-05 06:58:39.563005902 +0000 UTC m=+220.091913163" Dec 05 06:58:39 crc kubenswrapper[4997]: I1205 06:58:39.586732 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jmdrt" podStartSLOduration=3.393771177 podStartE2EDuration="1m3.586707261s" podCreationTimestamp="2025-12-05 06:57:36 +0000 UTC" firstStartedPulling="2025-12-05 06:57:38.721912794 +0000 UTC m=+159.250820055" lastFinishedPulling="2025-12-05 06:58:38.914848878 +0000 UTC m=+219.443756139" observedRunningTime="2025-12-05 06:58:39.583075155 +0000 UTC m=+220.111982416" watchObservedRunningTime="2025-12-05 06:58:39.586707261 +0000 UTC m=+220.115614522" Dec 05 06:58:39 crc kubenswrapper[4997]: I1205 06:58:39.606738 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5hfgl" podStartSLOduration=3.045624896 podStartE2EDuration="1m5.606718772s" podCreationTimestamp="2025-12-05 06:57:34 +0000 UTC" firstStartedPulling="2025-12-05 06:57:36.640465654 +0000 UTC m=+157.169372905" lastFinishedPulling="2025-12-05 06:58:39.20155951 +0000 UTC m=+219.730466781" observedRunningTime="2025-12-05 06:58:39.605395327 +0000 UTC m=+220.134302588" watchObservedRunningTime="2025-12-05 06:58:39.606718772 +0000 UTC m=+220.135626023" Dec 05 06:58:40 crc kubenswrapper[4997]: I1205 06:58:40.522061 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6tmw" event={"ID":"c0ca09da-0456-4ab9-8260-03c64893a6dd","Type":"ContainerStarted","Data":"ebb32219ca7297fc6f6e84a88a97517ec975997f7f3538a900818823da9b3a34"} Dec 05 06:58:40 crc kubenswrapper[4997]: I1205 06:58:40.527046 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dx8th" event={"ID":"dc3cd3bb-c155-44a7-a99f-f69d095b36e9","Type":"ContainerStarted","Data":"01fbda8e2ee2bb9f02960f8f4158920e698bc21e081efcd0d869b63667476685"} Dec 05 06:58:40 crc kubenswrapper[4997]: I1205 06:58:40.551468 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n6tmw" podStartSLOduration=2.300195248 podStartE2EDuration="1m6.551448437s" podCreationTimestamp="2025-12-05 06:57:34 +0000 UTC" firstStartedPulling="2025-12-05 06:57:35.641750342 +0000 UTC m=+156.170657603" lastFinishedPulling="2025-12-05 06:58:39.893003531 +0000 UTC m=+220.421910792" observedRunningTime="2025-12-05 06:58:40.548628663 +0000 UTC m=+221.077535924" watchObservedRunningTime="2025-12-05 06:58:40.551448437 +0000 UTC m=+221.080355698" Dec 05 06:58:41 crc kubenswrapper[4997]: I1205 06:58:41.536515 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqj49" event={"ID":"f67be9f3-d0bd-439c-ab66-72a6a29b3104","Type":"ContainerStarted","Data":"d145b208473d5074697d7999b4e8745198267a2d6c50f5818a14d646533803b8"} Dec 05 06:58:41 crc kubenswrapper[4997]: I1205 06:58:41.539707 4997 generic.go:334] "Generic (PLEG): container finished" podID="dc3cd3bb-c155-44a7-a99f-f69d095b36e9" containerID="01fbda8e2ee2bb9f02960f8f4158920e698bc21e081efcd0d869b63667476685" exitCode=0 Dec 05 06:58:41 crc kubenswrapper[4997]: I1205 06:58:41.539755 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dx8th" event={"ID":"dc3cd3bb-c155-44a7-a99f-f69d095b36e9","Type":"ContainerDied","Data":"01fbda8e2ee2bb9f02960f8f4158920e698bc21e081efcd0d869b63667476685"} Dec 05 06:58:43 crc kubenswrapper[4997]: I1205 06:58:43.555507 4997 generic.go:334] "Generic (PLEG): container finished" podID="f67be9f3-d0bd-439c-ab66-72a6a29b3104" containerID="d145b208473d5074697d7999b4e8745198267a2d6c50f5818a14d646533803b8" exitCode=0 Dec 05 06:58:43 crc kubenswrapper[4997]: I1205 06:58:43.555623 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqj49" event={"ID":"f67be9f3-d0bd-439c-ab66-72a6a29b3104","Type":"ContainerDied","Data":"d145b208473d5074697d7999b4e8745198267a2d6c50f5818a14d646533803b8"} Dec 05 06:58:43 crc kubenswrapper[4997]: I1205 06:58:43.560098 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dx8th" event={"ID":"dc3cd3bb-c155-44a7-a99f-f69d095b36e9","Type":"ContainerStarted","Data":"432d5d56b6809ff4f1ec1062ff08cd6e03078944759a06994347dbe78ac97528"} Dec 05 06:58:44 crc kubenswrapper[4997]: I1205 06:58:44.596384 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dx8th" podStartSLOduration=5.433352488 podStartE2EDuration="1m7.596362395s" podCreationTimestamp="2025-12-05 06:57:37 +0000 UTC" firstStartedPulling="2025-12-05 06:57:39.749484114 +0000 UTC m=+160.278391375" lastFinishedPulling="2025-12-05 06:58:41.912494031 +0000 UTC m=+222.441401282" observedRunningTime="2025-12-05 06:58:44.592836052 +0000 UTC m=+225.121743333" watchObservedRunningTime="2025-12-05 06:58:44.596362395 +0000 UTC m=+225.125269656" Dec 05 06:58:44 crc kubenswrapper[4997]: I1205 06:58:44.691740 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2pls6" Dec 05 06:58:44 crc kubenswrapper[4997]: I1205 06:58:44.691825 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2pls6" Dec 05 06:58:44 crc kubenswrapper[4997]: I1205 06:58:44.745668 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2pls6" Dec 05 06:58:44 crc kubenswrapper[4997]: I1205 06:58:44.952088 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n6tmw" Dec 05 06:58:44 crc kubenswrapper[4997]: I1205 06:58:44.952198 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n6tmw" Dec 05 06:58:45 crc kubenswrapper[4997]: I1205 06:58:45.001358 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n6tmw" Dec 05 06:58:45 crc kubenswrapper[4997]: I1205 06:58:45.099254 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cssd2" Dec 05 06:58:45 crc kubenswrapper[4997]: I1205 06:58:45.099492 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cssd2" Dec 05 06:58:45 crc kubenswrapper[4997]: I1205 06:58:45.140434 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cssd2" Dec 05 06:58:45 crc kubenswrapper[4997]: I1205 06:58:45.297317 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5hfgl" Dec 05 06:58:45 crc kubenswrapper[4997]: I1205 06:58:45.297383 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5hfgl" Dec 05 06:58:45 crc kubenswrapper[4997]: I1205 06:58:45.349934 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5hfgl" Dec 05 06:58:45 crc kubenswrapper[4997]: I1205 06:58:45.617903 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5hfgl" Dec 05 06:58:45 crc kubenswrapper[4997]: I1205 06:58:45.618668 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n6tmw" Dec 05 06:58:46 crc kubenswrapper[4997]: I1205 06:58:46.632660 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-2pls6" podUID="3ec5dfd3-3221-4141-a2d4-aea02733a985" containerName="registry-server" probeResult="failure" output=< Dec 05 06:58:46 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 06:58:46 crc kubenswrapper[4997]: > Dec 05 06:58:46 crc kubenswrapper[4997]: I1205 06:58:46.633499 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-cssd2" podUID="7fd5485d-8c41-484e-9373-4624f05a0f56" containerName="registry-server" probeResult="failure" output=< Dec 05 06:58:46 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 06:58:46 crc kubenswrapper[4997]: > Dec 05 06:58:46 crc kubenswrapper[4997]: I1205 06:58:46.687517 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cssd2" Dec 05 06:58:47 crc kubenswrapper[4997]: I1205 06:58:47.322670 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wxs8l"] Dec 05 06:58:47 crc kubenswrapper[4997]: I1205 06:58:47.375947 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jmdrt" Dec 05 06:58:47 crc kubenswrapper[4997]: I1205 06:58:47.375995 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jmdrt" Dec 05 06:58:47 crc kubenswrapper[4997]: I1205 06:58:47.453903 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jmdrt" Dec 05 06:58:47 crc kubenswrapper[4997]: I1205 06:58:47.662433 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jmdrt" Dec 05 06:58:48 crc kubenswrapper[4997]: I1205 06:58:48.474799 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dx8th" Dec 05 06:58:48 crc kubenswrapper[4997]: I1205 06:58:48.474847 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dx8th" Dec 05 06:58:48 crc kubenswrapper[4997]: I1205 06:58:48.523463 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dx8th" Dec 05 06:58:48 crc kubenswrapper[4997]: I1205 06:58:48.656887 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dx8th" Dec 05 06:58:48 crc kubenswrapper[4997]: I1205 06:58:48.983429 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5hfgl"] Dec 05 06:58:48 crc kubenswrapper[4997]: I1205 06:58:48.984086 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5hfgl" podUID="aa0cd037-9961-467b-a659-83922f8ed830" containerName="registry-server" containerID="cri-o://963332eff503d54d0385710006325ace7449be29c170c20f68f31fc51e41a1bd" gracePeriod=2 Dec 05 06:58:49 crc kubenswrapper[4997]: I1205 06:58:49.187266 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cssd2"] Dec 05 06:58:49 crc kubenswrapper[4997]: I1205 06:58:49.187777 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cssd2" podUID="7fd5485d-8c41-484e-9373-4624f05a0f56" containerName="registry-server" containerID="cri-o://0c4a0f0e3571591f4f2543d161e46e84758d11dec95d7c51bf2753b0f1f104dd" gracePeriod=2 Dec 05 06:58:51 crc kubenswrapper[4997]: I1205 06:58:51.382972 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jmdrt"] Dec 05 06:58:51 crc kubenswrapper[4997]: I1205 06:58:51.383491 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jmdrt" podUID="d4cb14f4-531c-4540-836d-3cd359b68be9" containerName="registry-server" containerID="cri-o://2ecaff6c9bf24e4087c79ba6b8be502a603c5b4ee044f951f8ab00076053738a" gracePeriod=2 Dec 05 06:58:51 crc kubenswrapper[4997]: I1205 06:58:51.582989 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dx8th"] Dec 05 06:58:51 crc kubenswrapper[4997]: I1205 06:58:51.583226 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dx8th" podUID="dc3cd3bb-c155-44a7-a99f-f69d095b36e9" containerName="registry-server" containerID="cri-o://432d5d56b6809ff4f1ec1062ff08cd6e03078944759a06994347dbe78ac97528" gracePeriod=2 Dec 05 06:58:51 crc kubenswrapper[4997]: I1205 06:58:51.633580 4997 generic.go:334] "Generic (PLEG): container finished" podID="7fd5485d-8c41-484e-9373-4624f05a0f56" containerID="0c4a0f0e3571591f4f2543d161e46e84758d11dec95d7c51bf2753b0f1f104dd" exitCode=0 Dec 05 06:58:51 crc kubenswrapper[4997]: I1205 06:58:51.633652 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cssd2" event={"ID":"7fd5485d-8c41-484e-9373-4624f05a0f56","Type":"ContainerDied","Data":"0c4a0f0e3571591f4f2543d161e46e84758d11dec95d7c51bf2753b0f1f104dd"} Dec 05 06:58:51 crc kubenswrapper[4997]: I1205 06:58:51.637081 4997 generic.go:334] "Generic (PLEG): container finished" podID="aa0cd037-9961-467b-a659-83922f8ed830" containerID="963332eff503d54d0385710006325ace7449be29c170c20f68f31fc51e41a1bd" exitCode=0 Dec 05 06:58:51 crc kubenswrapper[4997]: I1205 06:58:51.637112 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hfgl" event={"ID":"aa0cd037-9961-467b-a659-83922f8ed830","Type":"ContainerDied","Data":"963332eff503d54d0385710006325ace7449be29c170c20f68f31fc51e41a1bd"} Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.375157 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5hfgl" Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.521338 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa0cd037-9961-467b-a659-83922f8ed830-utilities\") pod \"aa0cd037-9961-467b-a659-83922f8ed830\" (UID: \"aa0cd037-9961-467b-a659-83922f8ed830\") " Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.521395 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa0cd037-9961-467b-a659-83922f8ed830-catalog-content\") pod \"aa0cd037-9961-467b-a659-83922f8ed830\" (UID: \"aa0cd037-9961-467b-a659-83922f8ed830\") " Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.521452 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5px8g\" (UniqueName: \"kubernetes.io/projected/aa0cd037-9961-467b-a659-83922f8ed830-kube-api-access-5px8g\") pod \"aa0cd037-9961-467b-a659-83922f8ed830\" (UID: \"aa0cd037-9961-467b-a659-83922f8ed830\") " Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.522748 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa0cd037-9961-467b-a659-83922f8ed830-utilities" (OuterVolumeSpecName: "utilities") pod "aa0cd037-9961-467b-a659-83922f8ed830" (UID: "aa0cd037-9961-467b-a659-83922f8ed830"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.527794 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa0cd037-9961-467b-a659-83922f8ed830-kube-api-access-5px8g" (OuterVolumeSpecName: "kube-api-access-5px8g") pod "aa0cd037-9961-467b-a659-83922f8ed830" (UID: "aa0cd037-9961-467b-a659-83922f8ed830"). InnerVolumeSpecName "kube-api-access-5px8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.622545 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa0cd037-9961-467b-a659-83922f8ed830-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.622636 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5px8g\" (UniqueName: \"kubernetes.io/projected/aa0cd037-9961-467b-a659-83922f8ed830-kube-api-access-5px8g\") on node \"crc\" DevicePath \"\"" Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.645129 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hfgl" event={"ID":"aa0cd037-9961-467b-a659-83922f8ed830","Type":"ContainerDied","Data":"650b42b244b24a569ecf470122e2ebb60bfa943dac234597e9c9b6eb358a23a9"} Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.645182 4997 scope.go:117] "RemoveContainer" containerID="963332eff503d54d0385710006325ace7449be29c170c20f68f31fc51e41a1bd" Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.645276 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5hfgl" Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.647843 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqj49" event={"ID":"f67be9f3-d0bd-439c-ab66-72a6a29b3104","Type":"ContainerStarted","Data":"bb9a8fd6335f6cdc2749befa3b807f9f84860609bd15b8e49189cf40c00dc69f"} Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.661304 4997 scope.go:117] "RemoveContainer" containerID="e705484faae7f53402859cd79406db73334f888b7aa230310b26c16a3f2f773a" Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.663432 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa0cd037-9961-467b-a659-83922f8ed830-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa0cd037-9961-467b-a659-83922f8ed830" (UID: "aa0cd037-9961-467b-a659-83922f8ed830"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.680782 4997 scope.go:117] "RemoveContainer" containerID="3174af97edd06b4ca2391bfa21badaf2c56d284cdba5792990bb4259b9a5a3d0" Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.724354 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa0cd037-9961-467b-a659-83922f8ed830-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.976369 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sqj49" podStartSLOduration=6.249914425 podStartE2EDuration="1m15.976346182s" podCreationTimestamp="2025-12-05 06:57:37 +0000 UTC" firstStartedPulling="2025-12-05 06:57:39.706773181 +0000 UTC m=+160.235680442" lastFinishedPulling="2025-12-05 06:58:49.433204938 +0000 UTC m=+229.962112199" observedRunningTime="2025-12-05 06:58:52.673052901 +0000 UTC m=+233.201960172" watchObservedRunningTime="2025-12-05 06:58:52.976346182 +0000 UTC m=+233.505253443" Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.980506 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5hfgl"] Dec 05 06:58:52 crc kubenswrapper[4997]: I1205 06:58:52.983925 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5hfgl"] Dec 05 06:58:53 crc kubenswrapper[4997]: I1205 06:58:53.655949 4997 generic.go:334] "Generic (PLEG): container finished" podID="d4cb14f4-531c-4540-836d-3cd359b68be9" containerID="2ecaff6c9bf24e4087c79ba6b8be502a603c5b4ee044f951f8ab00076053738a" exitCode=0 Dec 05 06:58:53 crc kubenswrapper[4997]: I1205 06:58:53.656061 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jmdrt" event={"ID":"d4cb14f4-531c-4540-836d-3cd359b68be9","Type":"ContainerDied","Data":"2ecaff6c9bf24e4087c79ba6b8be502a603c5b4ee044f951f8ab00076053738a"} Dec 05 06:58:53 crc kubenswrapper[4997]: I1205 06:58:53.658109 4997 generic.go:334] "Generic (PLEG): container finished" podID="dc3cd3bb-c155-44a7-a99f-f69d095b36e9" containerID="432d5d56b6809ff4f1ec1062ff08cd6e03078944759a06994347dbe78ac97528" exitCode=0 Dec 05 06:58:53 crc kubenswrapper[4997]: I1205 06:58:53.658181 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dx8th" event={"ID":"dc3cd3bb-c155-44a7-a99f-f69d095b36e9","Type":"ContainerDied","Data":"432d5d56b6809ff4f1ec1062ff08cd6e03078944759a06994347dbe78ac97528"} Dec 05 06:58:53 crc kubenswrapper[4997]: I1205 06:58:53.756785 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa0cd037-9961-467b-a659-83922f8ed830" path="/var/lib/kubelet/pods/aa0cd037-9961-467b-a659-83922f8ed830/volumes" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.038728 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cssd2" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.143675 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd5485d-8c41-484e-9373-4624f05a0f56-utilities\") pod \"7fd5485d-8c41-484e-9373-4624f05a0f56\" (UID: \"7fd5485d-8c41-484e-9373-4624f05a0f56\") " Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.143728 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd5485d-8c41-484e-9373-4624f05a0f56-catalog-content\") pod \"7fd5485d-8c41-484e-9373-4624f05a0f56\" (UID: \"7fd5485d-8c41-484e-9373-4624f05a0f56\") " Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.143778 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thxwv\" (UniqueName: \"kubernetes.io/projected/7fd5485d-8c41-484e-9373-4624f05a0f56-kube-api-access-thxwv\") pod \"7fd5485d-8c41-484e-9373-4624f05a0f56\" (UID: \"7fd5485d-8c41-484e-9373-4624f05a0f56\") " Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.145104 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fd5485d-8c41-484e-9373-4624f05a0f56-utilities" (OuterVolumeSpecName: "utilities") pod "7fd5485d-8c41-484e-9373-4624f05a0f56" (UID: "7fd5485d-8c41-484e-9373-4624f05a0f56"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.151769 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fd5485d-8c41-484e-9373-4624f05a0f56-kube-api-access-thxwv" (OuterVolumeSpecName: "kube-api-access-thxwv") pod "7fd5485d-8c41-484e-9373-4624f05a0f56" (UID: "7fd5485d-8c41-484e-9373-4624f05a0f56"). InnerVolumeSpecName "kube-api-access-thxwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.206519 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fd5485d-8c41-484e-9373-4624f05a0f56-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fd5485d-8c41-484e-9373-4624f05a0f56" (UID: "7fd5485d-8c41-484e-9373-4624f05a0f56"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.245974 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd5485d-8c41-484e-9373-4624f05a0f56-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.246010 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thxwv\" (UniqueName: \"kubernetes.io/projected/7fd5485d-8c41-484e-9373-4624f05a0f56-kube-api-access-thxwv\") on node \"crc\" DevicePath \"\"" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.246024 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd5485d-8c41-484e-9373-4624f05a0f56-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.362135 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dx8th" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.386202 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jmdrt" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.548498 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-catalog-content\") pod \"dc3cd3bb-c155-44a7-a99f-f69d095b36e9\" (UID: \"dc3cd3bb-c155-44a7-a99f-f69d095b36e9\") " Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.548594 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4cb14f4-531c-4540-836d-3cd359b68be9-utilities\") pod \"d4cb14f4-531c-4540-836d-3cd359b68be9\" (UID: \"d4cb14f4-531c-4540-836d-3cd359b68be9\") " Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.548672 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4cb14f4-531c-4540-836d-3cd359b68be9-catalog-content\") pod \"d4cb14f4-531c-4540-836d-3cd359b68be9\" (UID: \"d4cb14f4-531c-4540-836d-3cd359b68be9\") " Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.548695 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tb6tc\" (UniqueName: \"kubernetes.io/projected/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-kube-api-access-tb6tc\") pod \"dc3cd3bb-c155-44a7-a99f-f69d095b36e9\" (UID: \"dc3cd3bb-c155-44a7-a99f-f69d095b36e9\") " Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.548715 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rdkv\" (UniqueName: \"kubernetes.io/projected/d4cb14f4-531c-4540-836d-3cd359b68be9-kube-api-access-8rdkv\") pod \"d4cb14f4-531c-4540-836d-3cd359b68be9\" (UID: \"d4cb14f4-531c-4540-836d-3cd359b68be9\") " Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.548734 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-utilities\") pod \"dc3cd3bb-c155-44a7-a99f-f69d095b36e9\" (UID: \"dc3cd3bb-c155-44a7-a99f-f69d095b36e9\") " Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.549315 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4cb14f4-531c-4540-836d-3cd359b68be9-utilities" (OuterVolumeSpecName: "utilities") pod "d4cb14f4-531c-4540-836d-3cd359b68be9" (UID: "d4cb14f4-531c-4540-836d-3cd359b68be9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.549871 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-utilities" (OuterVolumeSpecName: "utilities") pod "dc3cd3bb-c155-44a7-a99f-f69d095b36e9" (UID: "dc3cd3bb-c155-44a7-a99f-f69d095b36e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.552413 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-kube-api-access-tb6tc" (OuterVolumeSpecName: "kube-api-access-tb6tc") pod "dc3cd3bb-c155-44a7-a99f-f69d095b36e9" (UID: "dc3cd3bb-c155-44a7-a99f-f69d095b36e9"). InnerVolumeSpecName "kube-api-access-tb6tc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.555557 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4cb14f4-531c-4540-836d-3cd359b68be9-kube-api-access-8rdkv" (OuterVolumeSpecName: "kube-api-access-8rdkv") pod "d4cb14f4-531c-4540-836d-3cd359b68be9" (UID: "d4cb14f4-531c-4540-836d-3cd359b68be9"). InnerVolumeSpecName "kube-api-access-8rdkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.572397 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4cb14f4-531c-4540-836d-3cd359b68be9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4cb14f4-531c-4540-836d-3cd359b68be9" (UID: "d4cb14f4-531c-4540-836d-3cd359b68be9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.650437 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rdkv\" (UniqueName: \"kubernetes.io/projected/d4cb14f4-531c-4540-836d-3cd359b68be9-kube-api-access-8rdkv\") on node \"crc\" DevicePath \"\"" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.650471 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.650482 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4cb14f4-531c-4540-836d-3cd359b68be9-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.650491 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4cb14f4-531c-4540-836d-3cd359b68be9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.650500 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tb6tc\" (UniqueName: \"kubernetes.io/projected/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-kube-api-access-tb6tc\") on node \"crc\" DevicePath \"\"" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.666790 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cssd2" event={"ID":"7fd5485d-8c41-484e-9373-4624f05a0f56","Type":"ContainerDied","Data":"6e27ce24291bc9b6844f40ee3423a13df8c2d0f28048ead46b869b99a1ffefc1"} Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.666839 4997 scope.go:117] "RemoveContainer" containerID="0c4a0f0e3571591f4f2543d161e46e84758d11dec95d7c51bf2753b0f1f104dd" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.667286 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cssd2" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.669638 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jmdrt" event={"ID":"d4cb14f4-531c-4540-836d-3cd359b68be9","Type":"ContainerDied","Data":"71613afd69ac6c9967265c8d0680d568ad00d51a42fb435be0fd984cf50a55b5"} Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.669754 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jmdrt" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.672741 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dx8th" event={"ID":"dc3cd3bb-c155-44a7-a99f-f69d095b36e9","Type":"ContainerDied","Data":"b7ce4b802cc9fe738b0edc996a212cc40bedd3636e446892e21d4a365bee831b"} Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.672905 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dx8th" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.677558 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc3cd3bb-c155-44a7-a99f-f69d095b36e9" (UID: "dc3cd3bb-c155-44a7-a99f-f69d095b36e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.684485 4997 scope.go:117] "RemoveContainer" containerID="224ed527028f096b800d7afd1880a7b1d9ce6311a4142e67d59309d5d107f0fe" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.705957 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cssd2"] Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.708635 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cssd2"] Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.722736 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jmdrt"] Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.725066 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jmdrt"] Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.725737 4997 scope.go:117] "RemoveContainer" containerID="9d924c2e9a88691bbeef9d9010647afbaf92684dccb8443f3964823cac62a347" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.740238 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2pls6" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.740778 4997 scope.go:117] "RemoveContainer" containerID="2ecaff6c9bf24e4087c79ba6b8be502a603c5b4ee044f951f8ab00076053738a" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.752329 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc3cd3bb-c155-44a7-a99f-f69d095b36e9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.776715 4997 scope.go:117] "RemoveContainer" containerID="f7282a58d752ffd2cb74f18156a55142843d056f36ebc97f04273d76bb405d0d" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.797455 4997 scope.go:117] "RemoveContainer" containerID="46139ad4d7fad459e8f2cc00b9dd0f5214db555de63812d56911adb28224c2a7" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.818380 4997 scope.go:117] "RemoveContainer" containerID="432d5d56b6809ff4f1ec1062ff08cd6e03078944759a06994347dbe78ac97528" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.832912 4997 scope.go:117] "RemoveContainer" containerID="01fbda8e2ee2bb9f02960f8f4158920e698bc21e081efcd0d869b63667476685" Dec 05 06:58:54 crc kubenswrapper[4997]: I1205 06:58:54.846710 4997 scope.go:117] "RemoveContainer" containerID="b8f93ce726e7b888c80953194a54ceb5acca65ff8395559196beb12b30c2f8ef" Dec 05 06:58:55 crc kubenswrapper[4997]: I1205 06:58:55.015293 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dx8th"] Dec 05 06:58:55 crc kubenswrapper[4997]: I1205 06:58:55.022882 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dx8th"] Dec 05 06:58:55 crc kubenswrapper[4997]: I1205 06:58:55.755499 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fd5485d-8c41-484e-9373-4624f05a0f56" path="/var/lib/kubelet/pods/7fd5485d-8c41-484e-9373-4624f05a0f56/volumes" Dec 05 06:58:55 crc kubenswrapper[4997]: I1205 06:58:55.756595 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4cb14f4-531c-4540-836d-3cd359b68be9" path="/var/lib/kubelet/pods/d4cb14f4-531c-4540-836d-3cd359b68be9/volumes" Dec 05 06:58:55 crc kubenswrapper[4997]: I1205 06:58:55.757190 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc3cd3bb-c155-44a7-a99f-f69d095b36e9" path="/var/lib/kubelet/pods/dc3cd3bb-c155-44a7-a99f-f69d095b36e9/volumes" Dec 05 06:58:57 crc kubenswrapper[4997]: I1205 06:58:57.887797 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sqj49" Dec 05 06:58:57 crc kubenswrapper[4997]: I1205 06:58:57.887893 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sqj49" Dec 05 06:58:57 crc kubenswrapper[4997]: I1205 06:58:57.965514 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sqj49" Dec 05 06:58:58 crc kubenswrapper[4997]: I1205 06:58:58.787033 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sqj49" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.221183 4997 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.221807 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5" gracePeriod=15 Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.221864 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b" gracePeriod=15 Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.221933 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c" gracePeriod=15 Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.221991 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5" gracePeriod=15 Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.221986 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28" gracePeriod=15 Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.223359 4997 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.223712 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa0cd037-9961-467b-a659-83922f8ed830" containerName="registry-server" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.223735 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa0cd037-9961-467b-a659-83922f8ed830" containerName="registry-server" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.223754 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4cb14f4-531c-4540-836d-3cd359b68be9" containerName="extract-utilities" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.223767 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4cb14f4-531c-4540-836d-3cd359b68be9" containerName="extract-utilities" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.223782 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa0cd037-9961-467b-a659-83922f8ed830" containerName="extract-content" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.223795 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa0cd037-9961-467b-a659-83922f8ed830" containerName="extract-content" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.223811 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4cb14f4-531c-4540-836d-3cd359b68be9" containerName="registry-server" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.223856 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4cb14f4-531c-4540-836d-3cd359b68be9" containerName="registry-server" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.223875 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.223889 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.223914 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.223926 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.223946 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.223958 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.223979 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc3cd3bb-c155-44a7-a99f-f69d095b36e9" containerName="extract-utilities" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.223992 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc3cd3bb-c155-44a7-a99f-f69d095b36e9" containerName="extract-utilities" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.224011 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc3cd3bb-c155-44a7-a99f-f69d095b36e9" containerName="extract-content" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224025 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc3cd3bb-c155-44a7-a99f-f69d095b36e9" containerName="extract-content" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.224039 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224051 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.224066 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4cb14f4-531c-4540-836d-3cd359b68be9" containerName="extract-content" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224078 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4cb14f4-531c-4540-836d-3cd359b68be9" containerName="extract-content" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.224090 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224161 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.224185 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc3cd3bb-c155-44a7-a99f-f69d095b36e9" containerName="registry-server" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224200 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc3cd3bb-c155-44a7-a99f-f69d095b36e9" containerName="registry-server" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.224216 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224232 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.224254 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd5485d-8c41-484e-9373-4624f05a0f56" containerName="registry-server" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224271 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd5485d-8c41-484e-9373-4624f05a0f56" containerName="registry-server" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.224287 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd5485d-8c41-484e-9373-4624f05a0f56" containerName="extract-utilities" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224300 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd5485d-8c41-484e-9373-4624f05a0f56" containerName="extract-utilities" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.224320 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c93e8cb8-7eea-411f-92a5-71d25785226a" containerName="pruner" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224332 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c93e8cb8-7eea-411f-92a5-71d25785226a" containerName="pruner" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.224344 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224356 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.224371 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa0cd037-9961-467b-a659-83922f8ed830" containerName="extract-utilities" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224384 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa0cd037-9961-467b-a659-83922f8ed830" containerName="extract-utilities" Dec 05 06:59:02 crc kubenswrapper[4997]: E1205 06:59:02.224401 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd5485d-8c41-484e-9373-4624f05a0f56" containerName="extract-content" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224417 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd5485d-8c41-484e-9373-4624f05a0f56" containerName="extract-content" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224738 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224765 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fd5485d-8c41-484e-9373-4624f05a0f56" containerName="registry-server" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224788 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4cb14f4-531c-4540-836d-3cd359b68be9" containerName="registry-server" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224806 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224818 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224834 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa0cd037-9961-467b-a659-83922f8ed830" containerName="registry-server" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224850 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc3cd3bb-c155-44a7-a99f-f69d095b36e9" containerName="registry-server" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224867 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224887 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224904 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="c93e8cb8-7eea-411f-92a5-71d25785226a" containerName="pruner" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.224920 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.235607 4997 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.236119 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.240378 4997 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.251285 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.251374 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.251552 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.251747 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.252092 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.275721 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.353324 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.353395 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.353421 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.353472 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.353646 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.353681 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.353738 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.353755 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.353798 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.353769 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.353851 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.353862 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.353924 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.455641 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.455694 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.455779 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.455842 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.455883 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.455940 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.566749 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:02 crc kubenswrapper[4997]: W1205 06:59:02.586881 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-ffbdc8ae03f80ef0e1a3cd2ef2389c5c7af1a973a5236098d9ec91ed9c835062 WatchSource:0}: Error finding container ffbdc8ae03f80ef0e1a3cd2ef2389c5c7af1a973a5236098d9ec91ed9c835062: Status 404 returned error can't find the container with id ffbdc8ae03f80ef0e1a3cd2ef2389c5c7af1a973a5236098d9ec91ed9c835062 Dec 05 06:59:02 crc kubenswrapper[4997]: I1205 06:59:02.741229 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"ffbdc8ae03f80ef0e1a3cd2ef2389c5c7af1a973a5236098d9ec91ed9c835062"} Dec 05 06:59:04 crc kubenswrapper[4997]: I1205 06:59:04.762330 4997 generic.go:334] "Generic (PLEG): container finished" podID="6d67d224-c501-4501-89a3-120c9f883811" containerID="0d6b37b511617b0bd20413663db8ea5f3ab1a96ab96b4467cc6d858ebefb1bec" exitCode=0 Dec 05 06:59:04 crc kubenswrapper[4997]: I1205 06:59:04.762470 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6d67d224-c501-4501-89a3-120c9f883811","Type":"ContainerDied","Data":"0d6b37b511617b0bd20413663db8ea5f3ab1a96ab96b4467cc6d858ebefb1bec"} Dec 05 06:59:04 crc kubenswrapper[4997]: I1205 06:59:04.763965 4997 status_manager.go:851] "Failed to get status for pod" podUID="6d67d224-c501-4501-89a3-120c9f883811" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:04 crc kubenswrapper[4997]: I1205 06:59:04.764388 4997 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:04 crc kubenswrapper[4997]: I1205 06:59:04.765552 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"c684cfbd0f9840634c2f512dc4d929c06642049eeb4f1e9459c23dcee4bd9ce9"} Dec 05 06:59:04 crc kubenswrapper[4997]: I1205 06:59:04.766154 4997 status_manager.go:851] "Failed to get status for pod" podUID="6d67d224-c501-4501-89a3-120c9f883811" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:04 crc kubenswrapper[4997]: I1205 06:59:04.766802 4997 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:04 crc kubenswrapper[4997]: I1205 06:59:04.767877 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 05 06:59:04 crc kubenswrapper[4997]: I1205 06:59:04.769405 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 06:59:04 crc kubenswrapper[4997]: I1205 06:59:04.770306 4997 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b" exitCode=0 Dec 05 06:59:04 crc kubenswrapper[4997]: I1205 06:59:04.770333 4997 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5" exitCode=0 Dec 05 06:59:04 crc kubenswrapper[4997]: I1205 06:59:04.770343 4997 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28" exitCode=0 Dec 05 06:59:04 crc kubenswrapper[4997]: I1205 06:59:04.770350 4997 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c" exitCode=2 Dec 05 06:59:04 crc kubenswrapper[4997]: I1205 06:59:04.770395 4997 scope.go:117] "RemoveContainer" containerID="4e96ae1e8311ed32561a99a9db63870c9967f5f4d5c130ccb7f39819a8e97696" Dec 05 06:59:05 crc kubenswrapper[4997]: E1205 06:59:05.223824 4997 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:05 crc kubenswrapper[4997]: E1205 06:59:05.224308 4997 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:05 crc kubenswrapper[4997]: E1205 06:59:05.224775 4997 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:05 crc kubenswrapper[4997]: E1205 06:59:05.225352 4997 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:05 crc kubenswrapper[4997]: E1205 06:59:05.226080 4997 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:05 crc kubenswrapper[4997]: I1205 06:59:05.226180 4997 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 05 06:59:05 crc kubenswrapper[4997]: E1205 06:59:05.226451 4997 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="200ms" Dec 05 06:59:05 crc kubenswrapper[4997]: E1205 06:59:05.427400 4997 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="400ms" Dec 05 06:59:05 crc kubenswrapper[4997]: I1205 06:59:05.784382 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 06:59:05 crc kubenswrapper[4997]: E1205 06:59:05.841279 4997 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="800ms" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.139457 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.140842 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.141447 4997 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.141593 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.141899 4997 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.142133 4997 status_manager.go:851] "Failed to get status for pod" podUID="6d67d224-c501-4501-89a3-120c9f883811" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.142357 4997 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.142560 4997 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.142784 4997 status_manager.go:851] "Failed to get status for pod" podUID="6d67d224-c501-4501-89a3-120c9f883811" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:06 crc kubenswrapper[4997]: E1205 06:59:06.641873 4997 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="1.6s" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.792888 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6d67d224-c501-4501-89a3-120c9f883811","Type":"ContainerDied","Data":"5485311f58b27194e15e61bbd64fc8e7d6063385ad804b0d760accae1d96e9d0"} Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.792947 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5485311f58b27194e15e61bbd64fc8e7d6063385ad804b0d760accae1d96e9d0" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.793010 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.797335 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.799377 4997 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5" exitCode=0 Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.799444 4997 scope.go:117] "RemoveContainer" containerID="d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.799539 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.819217 4997 scope.go:117] "RemoveContainer" containerID="df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.837594 4997 scope.go:117] "RemoveContainer" containerID="df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.855104 4997 scope.go:117] "RemoveContainer" containerID="84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.875312 4997 scope.go:117] "RemoveContainer" containerID="1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.897885 4997 scope.go:117] "RemoveContainer" containerID="5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.929153 4997 scope.go:117] "RemoveContainer" containerID="d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b" Dec 05 06:59:06 crc kubenswrapper[4997]: E1205 06:59:06.929804 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\": container with ID starting with d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b not found: ID does not exist" containerID="d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.929867 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b"} err="failed to get container status \"d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\": rpc error: code = NotFound desc = could not find container \"d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b\": container with ID starting with d14db22bd0820b5cd9d4b12ae90a37a1501e86cafb5f7e8b9babb7e7f5af8a5b not found: ID does not exist" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.929992 4997 scope.go:117] "RemoveContainer" containerID="df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5" Dec 05 06:59:06 crc kubenswrapper[4997]: E1205 06:59:06.930666 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\": container with ID starting with df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5 not found: ID does not exist" containerID="df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.930718 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5"} err="failed to get container status \"df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\": rpc error: code = NotFound desc = could not find container \"df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5\": container with ID starting with df2380e4f7b1bd57279d8a0d0f09d6b452bd388f390d1b842120c0e01c0d92c5 not found: ID does not exist" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.930755 4997 scope.go:117] "RemoveContainer" containerID="df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28" Dec 05 06:59:06 crc kubenswrapper[4997]: E1205 06:59:06.931415 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\": container with ID starting with df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28 not found: ID does not exist" containerID="df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.931496 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28"} err="failed to get container status \"df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\": rpc error: code = NotFound desc = could not find container \"df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28\": container with ID starting with df23d8e7c7d28c7a4fbbf4946711763055f312179e96586f4b7ac2abcfce4e28 not found: ID does not exist" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.931587 4997 scope.go:117] "RemoveContainer" containerID="84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c" Dec 05 06:59:06 crc kubenswrapper[4997]: E1205 06:59:06.932102 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\": container with ID starting with 84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c not found: ID does not exist" containerID="84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.932158 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c"} err="failed to get container status \"84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\": rpc error: code = NotFound desc = could not find container \"84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c\": container with ID starting with 84a6abcd0d31dd52869cf856d902dab8475e3bdd5c44188cec96e5ee5d14830c not found: ID does not exist" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.932196 4997 scope.go:117] "RemoveContainer" containerID="1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5" Dec 05 06:59:06 crc kubenswrapper[4997]: E1205 06:59:06.932565 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\": container with ID starting with 1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5 not found: ID does not exist" containerID="1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.932600 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5"} err="failed to get container status \"1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\": rpc error: code = NotFound desc = could not find container \"1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5\": container with ID starting with 1c7c571225617e31c03f331ad583c2622621bdd7e7e60916902fe9a73120b7f5 not found: ID does not exist" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.932658 4997 scope.go:117] "RemoveContainer" containerID="5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef" Dec 05 06:59:06 crc kubenswrapper[4997]: E1205 06:59:06.933050 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\": container with ID starting with 5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef not found: ID does not exist" containerID="5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef" Dec 05 06:59:06 crc kubenswrapper[4997]: I1205 06:59:06.933112 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef"} err="failed to get container status \"5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\": rpc error: code = NotFound desc = could not find container \"5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef\": container with ID starting with 5ce8c4ca71020f92c9ee71b38ecc169f3617ed10d678a863bb190cacbe5fd5ef not found: ID does not exist" Dec 05 06:59:08 crc kubenswrapper[4997]: E1205 06:59:08.244134 4997 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="3.2s" Dec 05 06:59:08 crc kubenswrapper[4997]: E1205 06:59:08.773285 4997 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.155:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e3f84d8372b0f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-05 06:59:03.771540239 +0000 UTC m=+244.300447490,LastTimestamp:2025-12-05 06:59:03.771540239 +0000 UTC m=+244.300447490,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 05 06:59:08 crc kubenswrapper[4997]: E1205 06:59:08.786704 4997 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.155:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" volumeName="registry-storage" Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.850968 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6d67d224-c501-4501-89a3-120c9f883811-var-lock\") pod \"6d67d224-c501-4501-89a3-120c9f883811\" (UID: \"6d67d224-c501-4501-89a3-120c9f883811\") " Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.851042 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.851114 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6d67d224-c501-4501-89a3-120c9f883811-kube-api-access\") pod \"6d67d224-c501-4501-89a3-120c9f883811\" (UID: \"6d67d224-c501-4501-89a3-120c9f883811\") " Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.851140 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.851154 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d67d224-c501-4501-89a3-120c9f883811-var-lock" (OuterVolumeSpecName: "var-lock") pod "6d67d224-c501-4501-89a3-120c9f883811" (UID: "6d67d224-c501-4501-89a3-120c9f883811"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.851220 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d67d224-c501-4501-89a3-120c9f883811-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6d67d224-c501-4501-89a3-120c9f883811" (UID: "6d67d224-c501-4501-89a3-120c9f883811"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.851188 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6d67d224-c501-4501-89a3-120c9f883811-kubelet-dir\") pod \"6d67d224-c501-4501-89a3-120c9f883811\" (UID: \"6d67d224-c501-4501-89a3-120c9f883811\") " Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.852171 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.851242 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.851260 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.852318 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.852401 4997 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.852418 4997 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.852428 4997 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6d67d224-c501-4501-89a3-120c9f883811-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.852437 4997 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.852445 4997 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6d67d224-c501-4501-89a3-120c9f883811-var-lock\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.855961 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d67d224-c501-4501-89a3-120c9f883811-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6d67d224-c501-4501-89a3-120c9f883811" (UID: "6d67d224-c501-4501-89a3-120c9f883811"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.907783 4997 status_manager.go:851] "Failed to get status for pod" podUID="6d67d224-c501-4501-89a3-120c9f883811" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.908320 4997 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.922232 4997 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.922531 4997 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.922817 4997 status_manager.go:851] "Failed to get status for pod" podUID="6d67d224-c501-4501-89a3-120c9f883811" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:08 crc kubenswrapper[4997]: I1205 06:59:08.954547 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6d67d224-c501-4501-89a3-120c9f883811-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:09 crc kubenswrapper[4997]: I1205 06:59:09.752082 4997 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:09 crc kubenswrapper[4997]: I1205 06:59:09.752645 4997 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:09 crc kubenswrapper[4997]: I1205 06:59:09.752916 4997 status_manager.go:851] "Failed to get status for pod" podUID="6d67d224-c501-4501-89a3-120c9f883811" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:09 crc kubenswrapper[4997]: I1205 06:59:09.758518 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 05 06:59:11 crc kubenswrapper[4997]: E1205 06:59:11.445660 4997 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="6.4s" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.357809 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" podUID="a412ad65-d072-433a-8d6f-7e5390195f89" containerName="oauth-openshift" containerID="cri-o://d894fdd8ca69d1463d83abe8e556551cbff7daa975da4022fb133542d1312cd2" gracePeriod=15 Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.771482 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.772722 4997 status_manager.go:851] "Failed to get status for pod" podUID="a412ad65-d072-433a-8d6f-7e5390195f89" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-wxs8l\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.773395 4997 status_manager.go:851] "Failed to get status for pod" podUID="6d67d224-c501-4501-89a3-120c9f883811" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.773986 4997 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.852290 4997 generic.go:334] "Generic (PLEG): container finished" podID="a412ad65-d072-433a-8d6f-7e5390195f89" containerID="d894fdd8ca69d1463d83abe8e556551cbff7daa975da4022fb133542d1312cd2" exitCode=0 Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.852351 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.852356 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" event={"ID":"a412ad65-d072-433a-8d6f-7e5390195f89","Type":"ContainerDied","Data":"d894fdd8ca69d1463d83abe8e556551cbff7daa975da4022fb133542d1312cd2"} Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.852904 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" event={"ID":"a412ad65-d072-433a-8d6f-7e5390195f89","Type":"ContainerDied","Data":"75211d00b7ee261b58a1f03075a25f2f030695d8d9afbd33fdafb70a239de409"} Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.852947 4997 scope.go:117] "RemoveContainer" containerID="d894fdd8ca69d1463d83abe8e556551cbff7daa975da4022fb133542d1312cd2" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.853311 4997 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.853821 4997 status_manager.go:851] "Failed to get status for pod" podUID="a412ad65-d072-433a-8d6f-7e5390195f89" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-wxs8l\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.854028 4997 status_manager.go:851] "Failed to get status for pod" podUID="6d67d224-c501-4501-89a3-120c9f883811" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.877781 4997 scope.go:117] "RemoveContainer" containerID="d894fdd8ca69d1463d83abe8e556551cbff7daa975da4022fb133542d1312cd2" Dec 05 06:59:12 crc kubenswrapper[4997]: E1205 06:59:12.878662 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d894fdd8ca69d1463d83abe8e556551cbff7daa975da4022fb133542d1312cd2\": container with ID starting with d894fdd8ca69d1463d83abe8e556551cbff7daa975da4022fb133542d1312cd2 not found: ID does not exist" containerID="d894fdd8ca69d1463d83abe8e556551cbff7daa975da4022fb133542d1312cd2" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.878719 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d894fdd8ca69d1463d83abe8e556551cbff7daa975da4022fb133542d1312cd2"} err="failed to get container status \"d894fdd8ca69d1463d83abe8e556551cbff7daa975da4022fb133542d1312cd2\": rpc error: code = NotFound desc = could not find container \"d894fdd8ca69d1463d83abe8e556551cbff7daa975da4022fb133542d1312cd2\": container with ID starting with d894fdd8ca69d1463d83abe8e556551cbff7daa975da4022fb133542d1312cd2 not found: ID does not exist" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.914043 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-ocp-branding-template\") pod \"a412ad65-d072-433a-8d6f-7e5390195f89\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.914108 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxvjd\" (UniqueName: \"kubernetes.io/projected/a412ad65-d072-433a-8d6f-7e5390195f89-kube-api-access-vxvjd\") pod \"a412ad65-d072-433a-8d6f-7e5390195f89\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.914136 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-idp-0-file-data\") pod \"a412ad65-d072-433a-8d6f-7e5390195f89\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.914160 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-session\") pod \"a412ad65-d072-433a-8d6f-7e5390195f89\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.914185 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-error\") pod \"a412ad65-d072-433a-8d6f-7e5390195f89\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.914207 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-provider-selection\") pod \"a412ad65-d072-433a-8d6f-7e5390195f89\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.914230 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-service-ca\") pod \"a412ad65-d072-433a-8d6f-7e5390195f89\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.914258 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-audit-policies\") pod \"a412ad65-d072-433a-8d6f-7e5390195f89\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.914289 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-serving-cert\") pod \"a412ad65-d072-433a-8d6f-7e5390195f89\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.914470 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-trusted-ca-bundle\") pod \"a412ad65-d072-433a-8d6f-7e5390195f89\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.914530 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-cliconfig\") pod \"a412ad65-d072-433a-8d6f-7e5390195f89\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.914562 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a412ad65-d072-433a-8d6f-7e5390195f89-audit-dir\") pod \"a412ad65-d072-433a-8d6f-7e5390195f89\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.914587 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-login\") pod \"a412ad65-d072-433a-8d6f-7e5390195f89\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.914606 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-router-certs\") pod \"a412ad65-d072-433a-8d6f-7e5390195f89\" (UID: \"a412ad65-d072-433a-8d6f-7e5390195f89\") " Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.914895 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a412ad65-d072-433a-8d6f-7e5390195f89-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "a412ad65-d072-433a-8d6f-7e5390195f89" (UID: "a412ad65-d072-433a-8d6f-7e5390195f89"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.915940 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "a412ad65-d072-433a-8d6f-7e5390195f89" (UID: "a412ad65-d072-433a-8d6f-7e5390195f89"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.916362 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "a412ad65-d072-433a-8d6f-7e5390195f89" (UID: "a412ad65-d072-433a-8d6f-7e5390195f89"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.916536 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "a412ad65-d072-433a-8d6f-7e5390195f89" (UID: "a412ad65-d072-433a-8d6f-7e5390195f89"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.917169 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "a412ad65-d072-433a-8d6f-7e5390195f89" (UID: "a412ad65-d072-433a-8d6f-7e5390195f89"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.922927 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "a412ad65-d072-433a-8d6f-7e5390195f89" (UID: "a412ad65-d072-433a-8d6f-7e5390195f89"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.926014 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "a412ad65-d072-433a-8d6f-7e5390195f89" (UID: "a412ad65-d072-433a-8d6f-7e5390195f89"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.928575 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "a412ad65-d072-433a-8d6f-7e5390195f89" (UID: "a412ad65-d072-433a-8d6f-7e5390195f89"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.930850 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "a412ad65-d072-433a-8d6f-7e5390195f89" (UID: "a412ad65-d072-433a-8d6f-7e5390195f89"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.932227 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a412ad65-d072-433a-8d6f-7e5390195f89-kube-api-access-vxvjd" (OuterVolumeSpecName: "kube-api-access-vxvjd") pod "a412ad65-d072-433a-8d6f-7e5390195f89" (UID: "a412ad65-d072-433a-8d6f-7e5390195f89"). InnerVolumeSpecName "kube-api-access-vxvjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.932584 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "a412ad65-d072-433a-8d6f-7e5390195f89" (UID: "a412ad65-d072-433a-8d6f-7e5390195f89"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.932789 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "a412ad65-d072-433a-8d6f-7e5390195f89" (UID: "a412ad65-d072-433a-8d6f-7e5390195f89"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.932931 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "a412ad65-d072-433a-8d6f-7e5390195f89" (UID: "a412ad65-d072-433a-8d6f-7e5390195f89"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:59:12 crc kubenswrapper[4997]: I1205 06:59:12.945449 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "a412ad65-d072-433a-8d6f-7e5390195f89" (UID: "a412ad65-d072-433a-8d6f-7e5390195f89"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 06:59:13 crc kubenswrapper[4997]: I1205 06:59:13.015599 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:13 crc kubenswrapper[4997]: I1205 06:59:13.015803 4997 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a412ad65-d072-433a-8d6f-7e5390195f89-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:13 crc kubenswrapper[4997]: I1205 06:59:13.015817 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:13 crc kubenswrapper[4997]: I1205 06:59:13.015829 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:13 crc kubenswrapper[4997]: I1205 06:59:13.015838 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:13 crc kubenswrapper[4997]: I1205 06:59:13.015847 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxvjd\" (UniqueName: \"kubernetes.io/projected/a412ad65-d072-433a-8d6f-7e5390195f89-kube-api-access-vxvjd\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:13 crc kubenswrapper[4997]: I1205 06:59:13.015864 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:13 crc kubenswrapper[4997]: I1205 06:59:13.015875 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:13 crc kubenswrapper[4997]: I1205 06:59:13.015886 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:13 crc kubenswrapper[4997]: I1205 06:59:13.015895 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:13 crc kubenswrapper[4997]: I1205 06:59:13.015905 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:13 crc kubenswrapper[4997]: I1205 06:59:13.015913 4997 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:13 crc kubenswrapper[4997]: I1205 06:59:13.015922 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:13 crc kubenswrapper[4997]: I1205 06:59:13.015932 4997 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a412ad65-d072-433a-8d6f-7e5390195f89-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:13 crc kubenswrapper[4997]: I1205 06:59:13.165252 4997 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:13 crc kubenswrapper[4997]: I1205 06:59:13.165874 4997 status_manager.go:851] "Failed to get status for pod" podUID="a412ad65-d072-433a-8d6f-7e5390195f89" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-wxs8l\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:13 crc kubenswrapper[4997]: I1205 06:59:13.166148 4997 status_manager.go:851] "Failed to get status for pod" podUID="6d67d224-c501-4501-89a3-120c9f883811" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:14 crc kubenswrapper[4997]: I1205 06:59:14.748501 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:14 crc kubenswrapper[4997]: I1205 06:59:14.752402 4997 status_manager.go:851] "Failed to get status for pod" podUID="a412ad65-d072-433a-8d6f-7e5390195f89" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-wxs8l\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:14 crc kubenswrapper[4997]: I1205 06:59:14.754960 4997 status_manager.go:851] "Failed to get status for pod" podUID="6d67d224-c501-4501-89a3-120c9f883811" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:14 crc kubenswrapper[4997]: I1205 06:59:14.755539 4997 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:14 crc kubenswrapper[4997]: I1205 06:59:14.776723 4997 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="072cc9c7-9e48-446b-99ec-c03892c7e66c" Dec 05 06:59:14 crc kubenswrapper[4997]: I1205 06:59:14.776775 4997 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="072cc9c7-9e48-446b-99ec-c03892c7e66c" Dec 05 06:59:14 crc kubenswrapper[4997]: E1205 06:59:14.777252 4997 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:14 crc kubenswrapper[4997]: I1205 06:59:14.777841 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:14 crc kubenswrapper[4997]: W1205 06:59:14.806365 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-8bf20d49f18f0792cfc045a5cc7cba3f8a84281b8a6c3900de4ec1c9626c5447 WatchSource:0}: Error finding container 8bf20d49f18f0792cfc045a5cc7cba3f8a84281b8a6c3900de4ec1c9626c5447: Status 404 returned error can't find the container with id 8bf20d49f18f0792cfc045a5cc7cba3f8a84281b8a6c3900de4ec1c9626c5447 Dec 05 06:59:14 crc kubenswrapper[4997]: I1205 06:59:14.867251 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8bf20d49f18f0792cfc045a5cc7cba3f8a84281b8a6c3900de4ec1c9626c5447"} Dec 05 06:59:15 crc kubenswrapper[4997]: I1205 06:59:15.878321 4997 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="4416b97408941b2cac31480664842ff0ecfd0edc0c96c42b4f606d22cb9fed5c" exitCode=0 Dec 05 06:59:15 crc kubenswrapper[4997]: I1205 06:59:15.878512 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"4416b97408941b2cac31480664842ff0ecfd0edc0c96c42b4f606d22cb9fed5c"} Dec 05 06:59:15 crc kubenswrapper[4997]: I1205 06:59:15.879126 4997 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="072cc9c7-9e48-446b-99ec-c03892c7e66c" Dec 05 06:59:15 crc kubenswrapper[4997]: I1205 06:59:15.879221 4997 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="072cc9c7-9e48-446b-99ec-c03892c7e66c" Dec 05 06:59:15 crc kubenswrapper[4997]: E1205 06:59:15.879908 4997 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:15 crc kubenswrapper[4997]: I1205 06:59:15.880008 4997 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:15 crc kubenswrapper[4997]: I1205 06:59:15.880562 4997 status_manager.go:851] "Failed to get status for pod" podUID="a412ad65-d072-433a-8d6f-7e5390195f89" pod="openshift-authentication/oauth-openshift-558db77b4-wxs8l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-wxs8l\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:15 crc kubenswrapper[4997]: I1205 06:59:15.881101 4997 status_manager.go:851] "Failed to get status for pod" podUID="6d67d224-c501-4501-89a3-120c9f883811" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 05 06:59:16 crc kubenswrapper[4997]: I1205 06:59:16.888414 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5abcb54741c8707584ad1a529b3de9a2a20ed43d725d953514f41124d51cbc4b"} Dec 05 06:59:16 crc kubenswrapper[4997]: I1205 06:59:16.888804 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"12617de6d9f73b9df00c7371d6da01b60bf2714f7108973cbb568326263222f9"} Dec 05 06:59:16 crc kubenswrapper[4997]: I1205 06:59:16.888818 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a8392af17ed4f2f2132eba7e2b55aba530d7e9c21cf84a8b0463fd8f68964d79"} Dec 05 06:59:16 crc kubenswrapper[4997]: I1205 06:59:16.888831 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"63b8c6bbd6b16550c1a20fa6af6390e561d39a51cfadf6ce46d30f2385e169b7"} Dec 05 06:59:16 crc kubenswrapper[4997]: I1205 06:59:16.895141 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 05 06:59:16 crc kubenswrapper[4997]: I1205 06:59:16.895198 4997 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451" exitCode=1 Dec 05 06:59:16 crc kubenswrapper[4997]: I1205 06:59:16.895257 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451"} Dec 05 06:59:16 crc kubenswrapper[4997]: I1205 06:59:16.895782 4997 scope.go:117] "RemoveContainer" containerID="f999338ed28d632e1dcc24bc88114db16d4b070bc40b84a329437b857c8a7451" Dec 05 06:59:17 crc kubenswrapper[4997]: I1205 06:59:17.904931 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 05 06:59:17 crc kubenswrapper[4997]: I1205 06:59:17.905385 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c587d6dc8af0241df1ac3ae0a17528c15ba01f4a96efd7343c6bbb0dfdc80df8"} Dec 05 06:59:17 crc kubenswrapper[4997]: I1205 06:59:17.908248 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"71370442d4684449ee6eb9c8faf95a6910c1ac2cc3848b5d756b38e045fc9e0d"} Dec 05 06:59:17 crc kubenswrapper[4997]: I1205 06:59:17.908421 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:17 crc kubenswrapper[4997]: I1205 06:59:17.908482 4997 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="072cc9c7-9e48-446b-99ec-c03892c7e66c" Dec 05 06:59:17 crc kubenswrapper[4997]: I1205 06:59:17.908502 4997 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="072cc9c7-9e48-446b-99ec-c03892c7e66c" Dec 05 06:59:19 crc kubenswrapper[4997]: I1205 06:59:19.778218 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:19 crc kubenswrapper[4997]: I1205 06:59:19.778605 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:19 crc kubenswrapper[4997]: I1205 06:59:19.783344 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:22 crc kubenswrapper[4997]: I1205 06:59:22.732847 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:59:22 crc kubenswrapper[4997]: I1205 06:59:22.736635 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:59:22 crc kubenswrapper[4997]: I1205 06:59:22.920787 4997 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:22 crc kubenswrapper[4997]: I1205 06:59:22.936821 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:59:23 crc kubenswrapper[4997]: I1205 06:59:23.941826 4997 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="072cc9c7-9e48-446b-99ec-c03892c7e66c" Dec 05 06:59:23 crc kubenswrapper[4997]: I1205 06:59:23.941857 4997 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="072cc9c7-9e48-446b-99ec-c03892c7e66c" Dec 05 06:59:23 crc kubenswrapper[4997]: I1205 06:59:23.946625 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:23 crc kubenswrapper[4997]: I1205 06:59:23.948876 4997 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="62ec1a84-72ab-4f1f-a8a7-b6a426d7376a" Dec 05 06:59:24 crc kubenswrapper[4997]: I1205 06:59:24.947129 4997 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="072cc9c7-9e48-446b-99ec-c03892c7e66c" Dec 05 06:59:24 crc kubenswrapper[4997]: I1205 06:59:24.947174 4997 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="072cc9c7-9e48-446b-99ec-c03892c7e66c" Dec 05 06:59:29 crc kubenswrapper[4997]: I1205 06:59:29.642383 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 05 06:59:29 crc kubenswrapper[4997]: I1205 06:59:29.681103 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 05 06:59:29 crc kubenswrapper[4997]: I1205 06:59:29.760507 4997 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="62ec1a84-72ab-4f1f-a8a7-b6a426d7376a" Dec 05 06:59:30 crc kubenswrapper[4997]: I1205 06:59:30.146430 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 05 06:59:30 crc kubenswrapper[4997]: I1205 06:59:30.162019 4997 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 05 06:59:30 crc kubenswrapper[4997]: I1205 06:59:30.162351 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=28.162334817 podStartE2EDuration="28.162334817s" podCreationTimestamp="2025-12-05 06:59:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:59:22.71823971 +0000 UTC m=+263.247146971" watchObservedRunningTime="2025-12-05 06:59:30.162334817 +0000 UTC m=+270.691242078" Dec 05 06:59:30 crc kubenswrapper[4997]: I1205 06:59:30.165909 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wxs8l","openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 06:59:30 crc kubenswrapper[4997]: I1205 06:59:30.165996 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 06:59:30 crc kubenswrapper[4997]: I1205 06:59:30.169883 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 06:59:30 crc kubenswrapper[4997]: I1205 06:59:30.185600 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=8.185580827 podStartE2EDuration="8.185580827s" podCreationTimestamp="2025-12-05 06:59:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:59:30.183292323 +0000 UTC m=+270.712199594" watchObservedRunningTime="2025-12-05 06:59:30.185580827 +0000 UTC m=+270.714488098" Dec 05 06:59:30 crc kubenswrapper[4997]: I1205 06:59:30.305778 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 05 06:59:31 crc kubenswrapper[4997]: I1205 06:59:31.757281 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a412ad65-d072-433a-8d6f-7e5390195f89" path="/var/lib/kubelet/pods/a412ad65-d072-433a-8d6f-7e5390195f89/volumes" Dec 05 06:59:32 crc kubenswrapper[4997]: I1205 06:59:32.607011 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 05 06:59:32 crc kubenswrapper[4997]: I1205 06:59:32.959267 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 06:59:33 crc kubenswrapper[4997]: I1205 06:59:33.072607 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 05 06:59:33 crc kubenswrapper[4997]: I1205 06:59:33.141390 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 06:59:33 crc kubenswrapper[4997]: I1205 06:59:33.939905 4997 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 06:59:33 crc kubenswrapper[4997]: I1205 06:59:33.940134 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://c684cfbd0f9840634c2f512dc4d929c06642049eeb4f1e9459c23dcee4bd9ce9" gracePeriod=5 Dec 05 06:59:34 crc kubenswrapper[4997]: I1205 06:59:34.124897 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 05 06:59:34 crc kubenswrapper[4997]: I1205 06:59:34.534300 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 05 06:59:34 crc kubenswrapper[4997]: I1205 06:59:34.834061 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 05 06:59:35 crc kubenswrapper[4997]: I1205 06:59:35.013451 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 05 06:59:35 crc kubenswrapper[4997]: I1205 06:59:35.066523 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 06:59:35 crc kubenswrapper[4997]: I1205 06:59:35.122584 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 05 06:59:35 crc kubenswrapper[4997]: I1205 06:59:35.252908 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 05 06:59:35 crc kubenswrapper[4997]: I1205 06:59:35.320370 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 05 06:59:35 crc kubenswrapper[4997]: I1205 06:59:35.443483 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 05 06:59:35 crc kubenswrapper[4997]: I1205 06:59:35.473267 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 05 06:59:35 crc kubenswrapper[4997]: I1205 06:59:35.623559 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 05 06:59:35 crc kubenswrapper[4997]: I1205 06:59:35.756476 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 05 06:59:36 crc kubenswrapper[4997]: I1205 06:59:36.054728 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 06:59:36 crc kubenswrapper[4997]: I1205 06:59:36.219347 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 05 06:59:36 crc kubenswrapper[4997]: I1205 06:59:36.443223 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 05 06:59:36 crc kubenswrapper[4997]: I1205 06:59:36.450643 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 05 06:59:36 crc kubenswrapper[4997]: I1205 06:59:36.643565 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 05 06:59:36 crc kubenswrapper[4997]: I1205 06:59:36.655931 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 05 06:59:36 crc kubenswrapper[4997]: I1205 06:59:36.767874 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.109593 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5f79ddd96-5d8sx"] Dec 05 06:59:37 crc kubenswrapper[4997]: E1205 06:59:37.109810 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d67d224-c501-4501-89a3-120c9f883811" containerName="installer" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.109821 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d67d224-c501-4501-89a3-120c9f883811" containerName="installer" Dec 05 06:59:37 crc kubenswrapper[4997]: E1205 06:59:37.109830 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.109835 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 05 06:59:37 crc kubenswrapper[4997]: E1205 06:59:37.109847 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a412ad65-d072-433a-8d6f-7e5390195f89" containerName="oauth-openshift" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.109853 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a412ad65-d072-433a-8d6f-7e5390195f89" containerName="oauth-openshift" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.109949 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a412ad65-d072-433a-8d6f-7e5390195f89" containerName="oauth-openshift" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.109961 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.109969 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d67d224-c501-4501-89a3-120c9f883811" containerName="installer" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.110346 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.114213 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.114238 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.114222 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.114349 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.114233 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.114245 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.114595 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.114728 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.114921 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.115134 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.115232 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.117031 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.122478 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.126922 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.127865 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.219458 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-audit-policies\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.219508 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.219553 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.219577 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-user-template-login\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.219595 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-user-template-error\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.219627 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.219650 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-service-ca\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.219666 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-session\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.219689 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.219712 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.219764 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-router-certs\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.219805 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-audit-dir\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.219828 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.219848 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9jtp\" (UniqueName: \"kubernetes.io/projected/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-kube-api-access-s9jtp\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.321259 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.321305 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-user-template-login\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.321345 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-user-template-error\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.321368 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.321403 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-service-ca\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.321423 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-session\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.321448 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.321492 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.321516 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-router-certs\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.321535 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-audit-dir\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.321580 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.321596 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9jtp\" (UniqueName: \"kubernetes.io/projected/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-kube-api-access-s9jtp\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.321627 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-audit-policies\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.321644 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.322362 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-audit-dir\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.322670 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.323113 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-audit-policies\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.323158 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-service-ca\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.323776 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.327054 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.327090 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-user-template-login\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.327305 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-router-certs\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.327506 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.327804 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.337858 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-system-session\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.340227 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.341354 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9jtp\" (UniqueName: \"kubernetes.io/projected/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-kube-api-access-s9jtp\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.344179 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7c8207a2-c64b-4989-adae-8b8aa0f6ceee-v4-0-config-user-template-error\") pod \"oauth-openshift-5f79ddd96-5d8sx\" (UID: \"7c8207a2-c64b-4989-adae-8b8aa0f6ceee\") " pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.432904 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.530731 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.600487 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.600602 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.676457 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.710471 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 05 06:59:37 crc kubenswrapper[4997]: I1205 06:59:37.952821 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 05 06:59:38 crc kubenswrapper[4997]: I1205 06:59:38.228164 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 05 06:59:38 crc kubenswrapper[4997]: I1205 06:59:38.233711 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 05 06:59:38 crc kubenswrapper[4997]: I1205 06:59:38.341590 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 05 06:59:38 crc kubenswrapper[4997]: I1205 06:59:38.356726 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 05 06:59:38 crc kubenswrapper[4997]: I1205 06:59:38.547238 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 05 06:59:38 crc kubenswrapper[4997]: I1205 06:59:38.630068 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5f79ddd96-5d8sx"] Dec 05 06:59:38 crc kubenswrapper[4997]: I1205 06:59:38.669330 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 05 06:59:38 crc kubenswrapper[4997]: I1205 06:59:38.835327 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 05 06:59:38 crc kubenswrapper[4997]: I1205 06:59:38.844558 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 05 06:59:38 crc kubenswrapper[4997]: I1205 06:59:38.903185 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 05 06:59:38 crc kubenswrapper[4997]: I1205 06:59:38.952891 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 05 06:59:39 crc kubenswrapper[4997]: I1205 06:59:39.115022 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 05 06:59:39 crc kubenswrapper[4997]: I1205 06:59:39.192116 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 05 06:59:39 crc kubenswrapper[4997]: I1205 06:59:39.320327 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 05 06:59:39 crc kubenswrapper[4997]: I1205 06:59:39.347409 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 05 06:59:39 crc kubenswrapper[4997]: I1205 06:59:39.420088 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 05 06:59:39 crc kubenswrapper[4997]: I1205 06:59:39.615915 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 05 06:59:39 crc kubenswrapper[4997]: I1205 06:59:39.623786 4997 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 05 06:59:39 crc kubenswrapper[4997]: I1205 06:59:39.980158 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.021764 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.026570 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.026624 4997 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="c684cfbd0f9840634c2f512dc4d929c06642049eeb4f1e9459c23dcee4bd9ce9" exitCode=137 Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.054366 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.265681 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.277509 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.319924 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.319988 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:40 crc kubenswrapper[4997]: E1205 06:59:40.334798 4997 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 06:59:40 crc kubenswrapper[4997]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-5f79ddd96-5d8sx_openshift-authentication_7c8207a2-c64b-4989-adae-8b8aa0f6ceee_0(6f35c732550dc45f5fae4ec768d118399ddbd23f15f6692e72f27ec7d36dc5f2): error adding pod openshift-authentication_oauth-openshift-5f79ddd96-5d8sx to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"6f35c732550dc45f5fae4ec768d118399ddbd23f15f6692e72f27ec7d36dc5f2" Netns:"/var/run/netns/a2a95ca6-b82c-424d-8021-0ac8b97e8c0f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-5f79ddd96-5d8sx;K8S_POD_INFRA_CONTAINER_ID=6f35c732550dc45f5fae4ec768d118399ddbd23f15f6692e72f27ec7d36dc5f2;K8S_POD_UID=7c8207a2-c64b-4989-adae-8b8aa0f6ceee" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-5f79ddd96-5d8sx] networking: Multus: [openshift-authentication/oauth-openshift-5f79ddd96-5d8sx/7c8207a2-c64b-4989-adae-8b8aa0f6ceee]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-5f79ddd96-5d8sx in out of cluster comm: pod "oauth-openshift-5f79ddd96-5d8sx" not found Dec 05 06:59:40 crc kubenswrapper[4997]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 06:59:40 crc kubenswrapper[4997]: > Dec 05 06:59:40 crc kubenswrapper[4997]: E1205 06:59:40.334871 4997 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 06:59:40 crc kubenswrapper[4997]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-5f79ddd96-5d8sx_openshift-authentication_7c8207a2-c64b-4989-adae-8b8aa0f6ceee_0(6f35c732550dc45f5fae4ec768d118399ddbd23f15f6692e72f27ec7d36dc5f2): error adding pod openshift-authentication_oauth-openshift-5f79ddd96-5d8sx to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"6f35c732550dc45f5fae4ec768d118399ddbd23f15f6692e72f27ec7d36dc5f2" Netns:"/var/run/netns/a2a95ca6-b82c-424d-8021-0ac8b97e8c0f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-5f79ddd96-5d8sx;K8S_POD_INFRA_CONTAINER_ID=6f35c732550dc45f5fae4ec768d118399ddbd23f15f6692e72f27ec7d36dc5f2;K8S_POD_UID=7c8207a2-c64b-4989-adae-8b8aa0f6ceee" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-5f79ddd96-5d8sx] networking: Multus: [openshift-authentication/oauth-openshift-5f79ddd96-5d8sx/7c8207a2-c64b-4989-adae-8b8aa0f6ceee]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-5f79ddd96-5d8sx in out of cluster comm: pod "oauth-openshift-5f79ddd96-5d8sx" not found Dec 05 06:59:40 crc kubenswrapper[4997]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 06:59:40 crc kubenswrapper[4997]: > pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:40 crc kubenswrapper[4997]: E1205 06:59:40.334897 4997 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 06:59:40 crc kubenswrapper[4997]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-5f79ddd96-5d8sx_openshift-authentication_7c8207a2-c64b-4989-adae-8b8aa0f6ceee_0(6f35c732550dc45f5fae4ec768d118399ddbd23f15f6692e72f27ec7d36dc5f2): error adding pod openshift-authentication_oauth-openshift-5f79ddd96-5d8sx to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"6f35c732550dc45f5fae4ec768d118399ddbd23f15f6692e72f27ec7d36dc5f2" Netns:"/var/run/netns/a2a95ca6-b82c-424d-8021-0ac8b97e8c0f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-5f79ddd96-5d8sx;K8S_POD_INFRA_CONTAINER_ID=6f35c732550dc45f5fae4ec768d118399ddbd23f15f6692e72f27ec7d36dc5f2;K8S_POD_UID=7c8207a2-c64b-4989-adae-8b8aa0f6ceee" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-5f79ddd96-5d8sx] networking: Multus: [openshift-authentication/oauth-openshift-5f79ddd96-5d8sx/7c8207a2-c64b-4989-adae-8b8aa0f6ceee]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-5f79ddd96-5d8sx in out of cluster comm: pod "oauth-openshift-5f79ddd96-5d8sx" not found Dec 05 06:59:40 crc kubenswrapper[4997]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 06:59:40 crc kubenswrapper[4997]: > pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:40 crc kubenswrapper[4997]: E1205 06:59:40.334958 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"oauth-openshift-5f79ddd96-5d8sx_openshift-authentication(7c8207a2-c64b-4989-adae-8b8aa0f6ceee)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"oauth-openshift-5f79ddd96-5d8sx_openshift-authentication(7c8207a2-c64b-4989-adae-8b8aa0f6ceee)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-5f79ddd96-5d8sx_openshift-authentication_7c8207a2-c64b-4989-adae-8b8aa0f6ceee_0(6f35c732550dc45f5fae4ec768d118399ddbd23f15f6692e72f27ec7d36dc5f2): error adding pod openshift-authentication_oauth-openshift-5f79ddd96-5d8sx to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"6f35c732550dc45f5fae4ec768d118399ddbd23f15f6692e72f27ec7d36dc5f2\\\" Netns:\\\"/var/run/netns/a2a95ca6-b82c-424d-8021-0ac8b97e8c0f\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-5f79ddd96-5d8sx;K8S_POD_INFRA_CONTAINER_ID=6f35c732550dc45f5fae4ec768d118399ddbd23f15f6692e72f27ec7d36dc5f2;K8S_POD_UID=7c8207a2-c64b-4989-adae-8b8aa0f6ceee\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-5f79ddd96-5d8sx] networking: Multus: [openshift-authentication/oauth-openshift-5f79ddd96-5d8sx/7c8207a2-c64b-4989-adae-8b8aa0f6ceee]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-5f79ddd96-5d8sx in out of cluster comm: pod \\\"oauth-openshift-5f79ddd96-5d8sx\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" podUID="7c8207a2-c64b-4989-adae-8b8aa0f6ceee" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.356300 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.361795 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.361849 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.361869 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.361915 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.361915 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.361988 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.362219 4997 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.362277 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.362310 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.362328 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.369754 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.393507 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.395838 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.398869 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.409739 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.463767 4997 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.463803 4997 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.463812 4997 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.463823 4997 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.468952 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.590774 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.605524 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.649078 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.649087 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.661512 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.770405 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.783057 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.796970 4997 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.852472 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.882584 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.892513 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.917984 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.948685 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 05 06:59:40 crc kubenswrapper[4997]: I1205 06:59:40.949341 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.000896 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.034190 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.034284 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.034301 4997 scope.go:117] "RemoveContainer" containerID="c684cfbd0f9840634c2f512dc4d929c06642049eeb4f1e9459c23dcee4bd9ce9" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.034285 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.034678 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.069016 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.137138 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.201389 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.233676 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.303405 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.366452 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.389403 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.412891 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.494839 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.505795 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.575117 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.579118 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.591068 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.683530 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.729132 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.741877 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.747178 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.755972 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.756204 4997 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.765288 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.765790 4997 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="c6f65339-7563-44ff-a2e8-6f263ce5427c" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.768522 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.768565 4997 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="c6f65339-7563-44ff-a2e8-6f263ce5427c" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.806687 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.845865 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.862356 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 05 06:59:41 crc kubenswrapper[4997]: I1205 06:59:41.938649 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 05 06:59:42 crc kubenswrapper[4997]: I1205 06:59:42.011215 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 05 06:59:42 crc kubenswrapper[4997]: I1205 06:59:42.178634 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 05 06:59:42 crc kubenswrapper[4997]: I1205 06:59:42.195128 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 05 06:59:42 crc kubenswrapper[4997]: I1205 06:59:42.200332 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 05 06:59:42 crc kubenswrapper[4997]: I1205 06:59:42.224228 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 05 06:59:42 crc kubenswrapper[4997]: I1205 06:59:42.268551 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 05 06:59:42 crc kubenswrapper[4997]: I1205 06:59:42.404372 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 05 06:59:42 crc kubenswrapper[4997]: I1205 06:59:42.444695 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 05 06:59:42 crc kubenswrapper[4997]: I1205 06:59:42.452750 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 05 06:59:42 crc kubenswrapper[4997]: I1205 06:59:42.647638 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 05 06:59:42 crc kubenswrapper[4997]: I1205 06:59:42.871271 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 05 06:59:42 crc kubenswrapper[4997]: I1205 06:59:42.872327 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 05 06:59:42 crc kubenswrapper[4997]: I1205 06:59:42.874296 4997 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 05 06:59:42 crc kubenswrapper[4997]: I1205 06:59:42.908815 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 05 06:59:43 crc kubenswrapper[4997]: I1205 06:59:43.228218 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 05 06:59:43 crc kubenswrapper[4997]: I1205 06:59:43.274163 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 05 06:59:43 crc kubenswrapper[4997]: I1205 06:59:43.302475 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 05 06:59:43 crc kubenswrapper[4997]: I1205 06:59:43.314354 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 05 06:59:43 crc kubenswrapper[4997]: I1205 06:59:43.362369 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 05 06:59:43 crc kubenswrapper[4997]: I1205 06:59:43.366871 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 05 06:59:43 crc kubenswrapper[4997]: I1205 06:59:43.454828 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 05 06:59:43 crc kubenswrapper[4997]: I1205 06:59:43.813700 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 05 06:59:43 crc kubenswrapper[4997]: I1205 06:59:43.839900 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 05 06:59:43 crc kubenswrapper[4997]: I1205 06:59:43.870218 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 05 06:59:43 crc kubenswrapper[4997]: I1205 06:59:43.936369 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 05 06:59:43 crc kubenswrapper[4997]: E1205 06:59:43.938174 4997 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 06:59:43 crc kubenswrapper[4997]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-5f79ddd96-5d8sx_openshift-authentication_7c8207a2-c64b-4989-adae-8b8aa0f6ceee_0(759cbacfda1fcf25fbd8b94e155b03f2b1276d8d4761f877cafb09b243599caa): error adding pod openshift-authentication_oauth-openshift-5f79ddd96-5d8sx to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"759cbacfda1fcf25fbd8b94e155b03f2b1276d8d4761f877cafb09b243599caa" Netns:"/var/run/netns/df108eeb-2018-45b3-9fd3-dc376c18ba59" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-5f79ddd96-5d8sx;K8S_POD_INFRA_CONTAINER_ID=759cbacfda1fcf25fbd8b94e155b03f2b1276d8d4761f877cafb09b243599caa;K8S_POD_UID=7c8207a2-c64b-4989-adae-8b8aa0f6ceee" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-5f79ddd96-5d8sx] networking: Multus: [openshift-authentication/oauth-openshift-5f79ddd96-5d8sx/7c8207a2-c64b-4989-adae-8b8aa0f6ceee]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-5f79ddd96-5d8sx in out of cluster comm: pod "oauth-openshift-5f79ddd96-5d8sx" not found Dec 05 06:59:43 crc kubenswrapper[4997]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 06:59:43 crc kubenswrapper[4997]: > Dec 05 06:59:43 crc kubenswrapper[4997]: E1205 06:59:43.938216 4997 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 06:59:43 crc kubenswrapper[4997]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-5f79ddd96-5d8sx_openshift-authentication_7c8207a2-c64b-4989-adae-8b8aa0f6ceee_0(759cbacfda1fcf25fbd8b94e155b03f2b1276d8d4761f877cafb09b243599caa): error adding pod openshift-authentication_oauth-openshift-5f79ddd96-5d8sx to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"759cbacfda1fcf25fbd8b94e155b03f2b1276d8d4761f877cafb09b243599caa" Netns:"/var/run/netns/df108eeb-2018-45b3-9fd3-dc376c18ba59" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-5f79ddd96-5d8sx;K8S_POD_INFRA_CONTAINER_ID=759cbacfda1fcf25fbd8b94e155b03f2b1276d8d4761f877cafb09b243599caa;K8S_POD_UID=7c8207a2-c64b-4989-adae-8b8aa0f6ceee" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-5f79ddd96-5d8sx] networking: Multus: [openshift-authentication/oauth-openshift-5f79ddd96-5d8sx/7c8207a2-c64b-4989-adae-8b8aa0f6ceee]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-5f79ddd96-5d8sx in out of cluster comm: pod "oauth-openshift-5f79ddd96-5d8sx" not found Dec 05 06:59:43 crc kubenswrapper[4997]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 06:59:43 crc kubenswrapper[4997]: > pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:43 crc kubenswrapper[4997]: E1205 06:59:43.938234 4997 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 06:59:43 crc kubenswrapper[4997]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-5f79ddd96-5d8sx_openshift-authentication_7c8207a2-c64b-4989-adae-8b8aa0f6ceee_0(759cbacfda1fcf25fbd8b94e155b03f2b1276d8d4761f877cafb09b243599caa): error adding pod openshift-authentication_oauth-openshift-5f79ddd96-5d8sx to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"759cbacfda1fcf25fbd8b94e155b03f2b1276d8d4761f877cafb09b243599caa" Netns:"/var/run/netns/df108eeb-2018-45b3-9fd3-dc376c18ba59" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-5f79ddd96-5d8sx;K8S_POD_INFRA_CONTAINER_ID=759cbacfda1fcf25fbd8b94e155b03f2b1276d8d4761f877cafb09b243599caa;K8S_POD_UID=7c8207a2-c64b-4989-adae-8b8aa0f6ceee" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-5f79ddd96-5d8sx] networking: Multus: [openshift-authentication/oauth-openshift-5f79ddd96-5d8sx/7c8207a2-c64b-4989-adae-8b8aa0f6ceee]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-5f79ddd96-5d8sx in out of cluster comm: pod "oauth-openshift-5f79ddd96-5d8sx" not found Dec 05 06:59:43 crc kubenswrapper[4997]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 06:59:43 crc kubenswrapper[4997]: > pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:43 crc kubenswrapper[4997]: E1205 06:59:43.938280 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"oauth-openshift-5f79ddd96-5d8sx_openshift-authentication(7c8207a2-c64b-4989-adae-8b8aa0f6ceee)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"oauth-openshift-5f79ddd96-5d8sx_openshift-authentication(7c8207a2-c64b-4989-adae-8b8aa0f6ceee)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-5f79ddd96-5d8sx_openshift-authentication_7c8207a2-c64b-4989-adae-8b8aa0f6ceee_0(759cbacfda1fcf25fbd8b94e155b03f2b1276d8d4761f877cafb09b243599caa): error adding pod openshift-authentication_oauth-openshift-5f79ddd96-5d8sx to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"759cbacfda1fcf25fbd8b94e155b03f2b1276d8d4761f877cafb09b243599caa\\\" Netns:\\\"/var/run/netns/df108eeb-2018-45b3-9fd3-dc376c18ba59\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-5f79ddd96-5d8sx;K8S_POD_INFRA_CONTAINER_ID=759cbacfda1fcf25fbd8b94e155b03f2b1276d8d4761f877cafb09b243599caa;K8S_POD_UID=7c8207a2-c64b-4989-adae-8b8aa0f6ceee\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-5f79ddd96-5d8sx] networking: Multus: [openshift-authentication/oauth-openshift-5f79ddd96-5d8sx/7c8207a2-c64b-4989-adae-8b8aa0f6ceee]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-5f79ddd96-5d8sx in out of cluster comm: pod \\\"oauth-openshift-5f79ddd96-5d8sx\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" podUID="7c8207a2-c64b-4989-adae-8b8aa0f6ceee" Dec 05 06:59:44 crc kubenswrapper[4997]: I1205 06:59:44.044438 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 05 06:59:44 crc kubenswrapper[4997]: I1205 06:59:44.086298 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 05 06:59:44 crc kubenswrapper[4997]: I1205 06:59:44.177322 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 05 06:59:44 crc kubenswrapper[4997]: I1205 06:59:44.220730 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 05 06:59:44 crc kubenswrapper[4997]: I1205 06:59:44.231060 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 05 06:59:44 crc kubenswrapper[4997]: I1205 06:59:44.252423 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 05 06:59:44 crc kubenswrapper[4997]: I1205 06:59:44.282438 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 05 06:59:44 crc kubenswrapper[4997]: I1205 06:59:44.372564 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 05 06:59:44 crc kubenswrapper[4997]: I1205 06:59:44.447962 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 05 06:59:44 crc kubenswrapper[4997]: I1205 06:59:44.484968 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 05 06:59:44 crc kubenswrapper[4997]: I1205 06:59:44.486750 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 06:59:44 crc kubenswrapper[4997]: I1205 06:59:44.509585 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 05 06:59:44 crc kubenswrapper[4997]: I1205 06:59:44.639421 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 05 06:59:44 crc kubenswrapper[4997]: I1205 06:59:44.849913 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 05 06:59:44 crc kubenswrapper[4997]: I1205 06:59:44.887159 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 05 06:59:44 crc kubenswrapper[4997]: I1205 06:59:44.965704 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 05 06:59:44 crc kubenswrapper[4997]: I1205 06:59:44.991449 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.001067 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.001126 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.009423 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.037681 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.037798 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.080244 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.119138 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.206592 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.226767 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.300167 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.313649 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.355978 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.420439 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.459387 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.461796 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.474935 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.620418 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.642957 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.667663 4997 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.700963 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.760314 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.823057 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.950011 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 05 06:59:45 crc kubenswrapper[4997]: I1205 06:59:45.996687 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 05 06:59:46 crc kubenswrapper[4997]: I1205 06:59:46.063661 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 05 06:59:46 crc kubenswrapper[4997]: I1205 06:59:46.092646 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 05 06:59:46 crc kubenswrapper[4997]: I1205 06:59:46.178261 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 05 06:59:46 crc kubenswrapper[4997]: I1205 06:59:46.202029 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 05 06:59:46 crc kubenswrapper[4997]: I1205 06:59:46.283761 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 05 06:59:46 crc kubenswrapper[4997]: I1205 06:59:46.328992 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 05 06:59:46 crc kubenswrapper[4997]: I1205 06:59:46.429712 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 05 06:59:46 crc kubenswrapper[4997]: I1205 06:59:46.444147 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 05 06:59:46 crc kubenswrapper[4997]: I1205 06:59:46.512660 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 05 06:59:46 crc kubenswrapper[4997]: I1205 06:59:46.579891 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 05 06:59:46 crc kubenswrapper[4997]: I1205 06:59:46.611784 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 05 06:59:46 crc kubenswrapper[4997]: I1205 06:59:46.616255 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 05 06:59:46 crc kubenswrapper[4997]: I1205 06:59:46.709523 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 05 06:59:46 crc kubenswrapper[4997]: I1205 06:59:46.871033 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 05 06:59:46 crc kubenswrapper[4997]: I1205 06:59:46.919449 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 05 06:59:46 crc kubenswrapper[4997]: I1205 06:59:46.956942 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 05 06:59:47 crc kubenswrapper[4997]: I1205 06:59:47.073809 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 05 06:59:47 crc kubenswrapper[4997]: I1205 06:59:47.078285 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 05 06:59:47 crc kubenswrapper[4997]: I1205 06:59:47.124437 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 05 06:59:47 crc kubenswrapper[4997]: I1205 06:59:47.141588 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 06:59:47 crc kubenswrapper[4997]: I1205 06:59:47.205438 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 05 06:59:47 crc kubenswrapper[4997]: I1205 06:59:47.214691 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 06:59:47 crc kubenswrapper[4997]: I1205 06:59:47.229298 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 05 06:59:47 crc kubenswrapper[4997]: I1205 06:59:47.266501 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 05 06:59:47 crc kubenswrapper[4997]: I1205 06:59:47.392032 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 05 06:59:47 crc kubenswrapper[4997]: I1205 06:59:47.444438 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 05 06:59:47 crc kubenswrapper[4997]: I1205 06:59:47.620642 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 05 06:59:47 crc kubenswrapper[4997]: I1205 06:59:47.771205 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 05 06:59:47 crc kubenswrapper[4997]: I1205 06:59:47.887230 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 05 06:59:47 crc kubenswrapper[4997]: I1205 06:59:47.923634 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 05 06:59:47 crc kubenswrapper[4997]: I1205 06:59:47.924105 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 05 06:59:47 crc kubenswrapper[4997]: I1205 06:59:47.980948 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 05 06:59:48 crc kubenswrapper[4997]: I1205 06:59:48.052426 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 05 06:59:48 crc kubenswrapper[4997]: I1205 06:59:48.165492 4997 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 05 06:59:48 crc kubenswrapper[4997]: I1205 06:59:48.214459 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 05 06:59:48 crc kubenswrapper[4997]: I1205 06:59:48.256560 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 05 06:59:48 crc kubenswrapper[4997]: I1205 06:59:48.390448 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 05 06:59:48 crc kubenswrapper[4997]: I1205 06:59:48.436003 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 05 06:59:48 crc kubenswrapper[4997]: I1205 06:59:48.456899 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 05 06:59:48 crc kubenswrapper[4997]: I1205 06:59:48.480048 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 05 06:59:48 crc kubenswrapper[4997]: I1205 06:59:48.483398 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 05 06:59:48 crc kubenswrapper[4997]: I1205 06:59:48.487773 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 05 06:59:48 crc kubenswrapper[4997]: I1205 06:59:48.540994 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 05 06:59:48 crc kubenswrapper[4997]: I1205 06:59:48.604316 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 05 06:59:48 crc kubenswrapper[4997]: I1205 06:59:48.850190 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 05 06:59:48 crc kubenswrapper[4997]: I1205 06:59:48.917491 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 05 06:59:48 crc kubenswrapper[4997]: I1205 06:59:48.973067 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 05 06:59:49 crc kubenswrapper[4997]: I1205 06:59:49.106391 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 05 06:59:49 crc kubenswrapper[4997]: I1205 06:59:49.209537 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 05 06:59:49 crc kubenswrapper[4997]: I1205 06:59:49.259749 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 05 06:59:49 crc kubenswrapper[4997]: I1205 06:59:49.292045 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 06:59:49 crc kubenswrapper[4997]: I1205 06:59:49.322281 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 05 06:59:49 crc kubenswrapper[4997]: I1205 06:59:49.508311 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 05 06:59:49 crc kubenswrapper[4997]: I1205 06:59:49.523322 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 05 06:59:49 crc kubenswrapper[4997]: I1205 06:59:49.862683 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 05 06:59:49 crc kubenswrapper[4997]: I1205 06:59:49.870324 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 05 06:59:49 crc kubenswrapper[4997]: I1205 06:59:49.997623 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 05 06:59:50 crc kubenswrapper[4997]: I1205 06:59:50.060008 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 05 06:59:50 crc kubenswrapper[4997]: I1205 06:59:50.098588 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 05 06:59:50 crc kubenswrapper[4997]: I1205 06:59:50.478493 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 05 06:59:50 crc kubenswrapper[4997]: I1205 06:59:50.611471 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 05 06:59:50 crc kubenswrapper[4997]: I1205 06:59:50.666558 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 05 06:59:50 crc kubenswrapper[4997]: I1205 06:59:50.746977 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 05 06:59:50 crc kubenswrapper[4997]: I1205 06:59:50.842582 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 05 06:59:50 crc kubenswrapper[4997]: I1205 06:59:50.891215 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 05 06:59:50 crc kubenswrapper[4997]: I1205 06:59:50.975571 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 05 06:59:51 crc kubenswrapper[4997]: I1205 06:59:51.002685 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 05 06:59:51 crc kubenswrapper[4997]: I1205 06:59:51.198576 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 05 06:59:51 crc kubenswrapper[4997]: I1205 06:59:51.488565 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 06:59:51 crc kubenswrapper[4997]: I1205 06:59:51.558398 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 05 06:59:51 crc kubenswrapper[4997]: I1205 06:59:51.660268 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 05 06:59:52 crc kubenswrapper[4997]: I1205 06:59:52.091813 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 05 06:59:52 crc kubenswrapper[4997]: I1205 06:59:52.498180 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 05 06:59:52 crc kubenswrapper[4997]: I1205 06:59:52.886338 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 06:59:57 crc kubenswrapper[4997]: I1205 06:59:57.751395 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:57 crc kubenswrapper[4997]: I1205 06:59:57.752531 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:58 crc kubenswrapper[4997]: I1205 06:59:58.010660 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5f79ddd96-5d8sx"] Dec 05 06:59:58 crc kubenswrapper[4997]: I1205 06:59:58.152240 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" event={"ID":"7c8207a2-c64b-4989-adae-8b8aa0f6ceee","Type":"ContainerStarted","Data":"afe99186791e89b58eadbdee266d2139d307214abeccc8bac4f464b4f1c89229"} Dec 05 06:59:59 crc kubenswrapper[4997]: I1205 06:59:59.160878 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" event={"ID":"7c8207a2-c64b-4989-adae-8b8aa0f6ceee","Type":"ContainerStarted","Data":"792f71800c9a235b702accd96174d205439562b73bdcfcba51c9e6f34fe48a17"} Dec 05 06:59:59 crc kubenswrapper[4997]: I1205 06:59:59.161197 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:59 crc kubenswrapper[4997]: I1205 06:59:59.166114 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" Dec 05 06:59:59 crc kubenswrapper[4997]: I1205 06:59:59.186345 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5f79ddd96-5d8sx" podStartSLOduration=72.186322345 podStartE2EDuration="1m12.186322345s" podCreationTimestamp="2025-12-05 06:58:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 06:59:59.182337696 +0000 UTC m=+299.711244987" watchObservedRunningTime="2025-12-05 06:59:59.186322345 +0000 UTC m=+299.715229606" Dec 05 06:59:59 crc kubenswrapper[4997]: I1205 06:59:59.626173 4997 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 05 07:00:00 crc kubenswrapper[4997]: I1205 07:00:00.173264 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs"] Dec 05 07:00:00 crc kubenswrapper[4997]: I1205 07:00:00.174459 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs" Dec 05 07:00:00 crc kubenswrapper[4997]: I1205 07:00:00.179684 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 07:00:00 crc kubenswrapper[4997]: I1205 07:00:00.179910 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 07:00:00 crc kubenswrapper[4997]: I1205 07:00:00.182334 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs"] Dec 05 07:00:00 crc kubenswrapper[4997]: I1205 07:00:00.316334 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/85d4925b-c3c9-423d-8bf4-0294e312da5c-secret-volume\") pod \"collect-profiles-29415300-kvfgs\" (UID: \"85d4925b-c3c9-423d-8bf4-0294e312da5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs" Dec 05 07:00:00 crc kubenswrapper[4997]: I1205 07:00:00.316416 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85d4925b-c3c9-423d-8bf4-0294e312da5c-config-volume\") pod \"collect-profiles-29415300-kvfgs\" (UID: \"85d4925b-c3c9-423d-8bf4-0294e312da5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs" Dec 05 07:00:00 crc kubenswrapper[4997]: I1205 07:00:00.316674 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rpl6\" (UniqueName: \"kubernetes.io/projected/85d4925b-c3c9-423d-8bf4-0294e312da5c-kube-api-access-9rpl6\") pod \"collect-profiles-29415300-kvfgs\" (UID: \"85d4925b-c3c9-423d-8bf4-0294e312da5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs" Dec 05 07:00:00 crc kubenswrapper[4997]: I1205 07:00:00.418775 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/85d4925b-c3c9-423d-8bf4-0294e312da5c-secret-volume\") pod \"collect-profiles-29415300-kvfgs\" (UID: \"85d4925b-c3c9-423d-8bf4-0294e312da5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs" Dec 05 07:00:00 crc kubenswrapper[4997]: I1205 07:00:00.418862 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85d4925b-c3c9-423d-8bf4-0294e312da5c-config-volume\") pod \"collect-profiles-29415300-kvfgs\" (UID: \"85d4925b-c3c9-423d-8bf4-0294e312da5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs" Dec 05 07:00:00 crc kubenswrapper[4997]: I1205 07:00:00.418914 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rpl6\" (UniqueName: \"kubernetes.io/projected/85d4925b-c3c9-423d-8bf4-0294e312da5c-kube-api-access-9rpl6\") pod \"collect-profiles-29415300-kvfgs\" (UID: \"85d4925b-c3c9-423d-8bf4-0294e312da5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs" Dec 05 07:00:00 crc kubenswrapper[4997]: I1205 07:00:00.420236 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85d4925b-c3c9-423d-8bf4-0294e312da5c-config-volume\") pod \"collect-profiles-29415300-kvfgs\" (UID: \"85d4925b-c3c9-423d-8bf4-0294e312da5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs" Dec 05 07:00:00 crc kubenswrapper[4997]: I1205 07:00:00.428759 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/85d4925b-c3c9-423d-8bf4-0294e312da5c-secret-volume\") pod \"collect-profiles-29415300-kvfgs\" (UID: \"85d4925b-c3c9-423d-8bf4-0294e312da5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs" Dec 05 07:00:00 crc kubenswrapper[4997]: I1205 07:00:00.441893 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rpl6\" (UniqueName: \"kubernetes.io/projected/85d4925b-c3c9-423d-8bf4-0294e312da5c-kube-api-access-9rpl6\") pod \"collect-profiles-29415300-kvfgs\" (UID: \"85d4925b-c3c9-423d-8bf4-0294e312da5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs" Dec 05 07:00:00 crc kubenswrapper[4997]: I1205 07:00:00.491551 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs" Dec 05 07:00:00 crc kubenswrapper[4997]: I1205 07:00:00.718325 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs"] Dec 05 07:00:00 crc kubenswrapper[4997]: W1205 07:00:00.732935 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85d4925b_c3c9_423d_8bf4_0294e312da5c.slice/crio-1d9a4dadcbe9adb4d07fcfc961f80017f8bd2a306a1bbe6780986135543f6404 WatchSource:0}: Error finding container 1d9a4dadcbe9adb4d07fcfc961f80017f8bd2a306a1bbe6780986135543f6404: Status 404 returned error can't find the container with id 1d9a4dadcbe9adb4d07fcfc961f80017f8bd2a306a1bbe6780986135543f6404 Dec 05 07:00:01 crc kubenswrapper[4997]: I1205 07:00:01.175875 4997 generic.go:334] "Generic (PLEG): container finished" podID="85d4925b-c3c9-423d-8bf4-0294e312da5c" containerID="cb4d3efb9e35206f1dade7c89b5f598b5e26bf62b3bbbac259ae49dbcef681e7" exitCode=0 Dec 05 07:00:01 crc kubenswrapper[4997]: I1205 07:00:01.177035 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs" event={"ID":"85d4925b-c3c9-423d-8bf4-0294e312da5c","Type":"ContainerDied","Data":"cb4d3efb9e35206f1dade7c89b5f598b5e26bf62b3bbbac259ae49dbcef681e7"} Dec 05 07:00:01 crc kubenswrapper[4997]: I1205 07:00:01.177070 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs" event={"ID":"85d4925b-c3c9-423d-8bf4-0294e312da5c","Type":"ContainerStarted","Data":"1d9a4dadcbe9adb4d07fcfc961f80017f8bd2a306a1bbe6780986135543f6404"} Dec 05 07:00:02 crc kubenswrapper[4997]: I1205 07:00:02.437388 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs" Dec 05 07:00:02 crc kubenswrapper[4997]: I1205 07:00:02.552858 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85d4925b-c3c9-423d-8bf4-0294e312da5c-config-volume\") pod \"85d4925b-c3c9-423d-8bf4-0294e312da5c\" (UID: \"85d4925b-c3c9-423d-8bf4-0294e312da5c\") " Dec 05 07:00:02 crc kubenswrapper[4997]: I1205 07:00:02.552966 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/85d4925b-c3c9-423d-8bf4-0294e312da5c-secret-volume\") pod \"85d4925b-c3c9-423d-8bf4-0294e312da5c\" (UID: \"85d4925b-c3c9-423d-8bf4-0294e312da5c\") " Dec 05 07:00:02 crc kubenswrapper[4997]: I1205 07:00:02.553094 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rpl6\" (UniqueName: \"kubernetes.io/projected/85d4925b-c3c9-423d-8bf4-0294e312da5c-kube-api-access-9rpl6\") pod \"85d4925b-c3c9-423d-8bf4-0294e312da5c\" (UID: \"85d4925b-c3c9-423d-8bf4-0294e312da5c\") " Dec 05 07:00:02 crc kubenswrapper[4997]: I1205 07:00:02.554279 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85d4925b-c3c9-423d-8bf4-0294e312da5c-config-volume" (OuterVolumeSpecName: "config-volume") pod "85d4925b-c3c9-423d-8bf4-0294e312da5c" (UID: "85d4925b-c3c9-423d-8bf4-0294e312da5c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:00:02 crc kubenswrapper[4997]: I1205 07:00:02.560549 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85d4925b-c3c9-423d-8bf4-0294e312da5c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "85d4925b-c3c9-423d-8bf4-0294e312da5c" (UID: "85d4925b-c3c9-423d-8bf4-0294e312da5c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:00:02 crc kubenswrapper[4997]: I1205 07:00:02.560944 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85d4925b-c3c9-423d-8bf4-0294e312da5c-kube-api-access-9rpl6" (OuterVolumeSpecName: "kube-api-access-9rpl6") pod "85d4925b-c3c9-423d-8bf4-0294e312da5c" (UID: "85d4925b-c3c9-423d-8bf4-0294e312da5c"). InnerVolumeSpecName "kube-api-access-9rpl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:00:02 crc kubenswrapper[4997]: I1205 07:00:02.654269 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rpl6\" (UniqueName: \"kubernetes.io/projected/85d4925b-c3c9-423d-8bf4-0294e312da5c-kube-api-access-9rpl6\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:02 crc kubenswrapper[4997]: I1205 07:00:02.654334 4997 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85d4925b-c3c9-423d-8bf4-0294e312da5c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:02 crc kubenswrapper[4997]: I1205 07:00:02.654344 4997 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/85d4925b-c3c9-423d-8bf4-0294e312da5c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:03 crc kubenswrapper[4997]: I1205 07:00:03.194497 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs" event={"ID":"85d4925b-c3c9-423d-8bf4-0294e312da5c","Type":"ContainerDied","Data":"1d9a4dadcbe9adb4d07fcfc961f80017f8bd2a306a1bbe6780986135543f6404"} Dec 05 07:00:03 crc kubenswrapper[4997]: I1205 07:00:03.195098 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d9a4dadcbe9adb4d07fcfc961f80017f8bd2a306a1bbe6780986135543f6404" Dec 05 07:00:03 crc kubenswrapper[4997]: I1205 07:00:03.194603 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs" Dec 05 07:00:04 crc kubenswrapper[4997]: I1205 07:00:04.204454 4997 generic.go:334] "Generic (PLEG): container finished" podID="17fe9bd5-5973-47bd-945f-55bb8caa9d65" containerID="feb7b9a4853bb63a72df3919b0a2cb6b319deb2104e5684bc736c65bf79d9844" exitCode=0 Dec 05 07:00:04 crc kubenswrapper[4997]: I1205 07:00:04.204507 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" event={"ID":"17fe9bd5-5973-47bd-945f-55bb8caa9d65","Type":"ContainerDied","Data":"feb7b9a4853bb63a72df3919b0a2cb6b319deb2104e5684bc736c65bf79d9844"} Dec 05 07:00:04 crc kubenswrapper[4997]: I1205 07:00:04.205107 4997 scope.go:117] "RemoveContainer" containerID="feb7b9a4853bb63a72df3919b0a2cb6b319deb2104e5684bc736c65bf79d9844" Dec 05 07:00:05 crc kubenswrapper[4997]: I1205 07:00:05.213446 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" event={"ID":"17fe9bd5-5973-47bd-945f-55bb8caa9d65","Type":"ContainerStarted","Data":"99b851f62ee198eeaadf857187e158d3889e72c84fe6d9c43f641d75b03a05df"} Dec 05 07:00:05 crc kubenswrapper[4997]: I1205 07:00:05.214429 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" Dec 05 07:00:05 crc kubenswrapper[4997]: I1205 07:00:05.217534 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" Dec 05 07:00:15 crc kubenswrapper[4997]: I1205 07:00:15.232433 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-57ksd"] Dec 05 07:00:15 crc kubenswrapper[4997]: I1205 07:00:15.233588 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" podUID="e6a372ba-59d9-4fcc-ba21-7b9de5316239" containerName="controller-manager" containerID="cri-o://e185ef630ce8f8eb0f0eafee24a427eab0bff85e1965dff9023bab13a95480f3" gracePeriod=30 Dec 05 07:00:15 crc kubenswrapper[4997]: I1205 07:00:15.326954 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh"] Dec 05 07:00:15 crc kubenswrapper[4997]: I1205 07:00:15.327481 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" podUID="0a38ab13-190d-40c5-9a69-389e2fae6884" containerName="route-controller-manager" containerID="cri-o://a5fe8cbc600bdc6d50b247ce9eeb3580052821f375efb95eb0d296e7169ffde4" gracePeriod=30 Dec 05 07:00:15 crc kubenswrapper[4997]: I1205 07:00:15.676193 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 07:00:15 crc kubenswrapper[4997]: I1205 07:00:15.777139 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a38ab13-190d-40c5-9a69-389e2fae6884-config\") pod \"0a38ab13-190d-40c5-9a69-389e2fae6884\" (UID: \"0a38ab13-190d-40c5-9a69-389e2fae6884\") " Dec 05 07:00:15 crc kubenswrapper[4997]: I1205 07:00:15.777281 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a38ab13-190d-40c5-9a69-389e2fae6884-serving-cert\") pod \"0a38ab13-190d-40c5-9a69-389e2fae6884\" (UID: \"0a38ab13-190d-40c5-9a69-389e2fae6884\") " Dec 05 07:00:15 crc kubenswrapper[4997]: I1205 07:00:15.777410 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzjz6\" (UniqueName: \"kubernetes.io/projected/0a38ab13-190d-40c5-9a69-389e2fae6884-kube-api-access-rzjz6\") pod \"0a38ab13-190d-40c5-9a69-389e2fae6884\" (UID: \"0a38ab13-190d-40c5-9a69-389e2fae6884\") " Dec 05 07:00:15 crc kubenswrapper[4997]: I1205 07:00:15.777463 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0a38ab13-190d-40c5-9a69-389e2fae6884-client-ca\") pod \"0a38ab13-190d-40c5-9a69-389e2fae6884\" (UID: \"0a38ab13-190d-40c5-9a69-389e2fae6884\") " Dec 05 07:00:15 crc kubenswrapper[4997]: I1205 07:00:15.778105 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a38ab13-190d-40c5-9a69-389e2fae6884-config" (OuterVolumeSpecName: "config") pod "0a38ab13-190d-40c5-9a69-389e2fae6884" (UID: "0a38ab13-190d-40c5-9a69-389e2fae6884"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:00:15 crc kubenswrapper[4997]: I1205 07:00:15.778714 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a38ab13-190d-40c5-9a69-389e2fae6884-client-ca" (OuterVolumeSpecName: "client-ca") pod "0a38ab13-190d-40c5-9a69-389e2fae6884" (UID: "0a38ab13-190d-40c5-9a69-389e2fae6884"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:00:15 crc kubenswrapper[4997]: I1205 07:00:15.784972 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a38ab13-190d-40c5-9a69-389e2fae6884-kube-api-access-rzjz6" (OuterVolumeSpecName: "kube-api-access-rzjz6") pod "0a38ab13-190d-40c5-9a69-389e2fae6884" (UID: "0a38ab13-190d-40c5-9a69-389e2fae6884"). InnerVolumeSpecName "kube-api-access-rzjz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:00:15 crc kubenswrapper[4997]: I1205 07:00:15.785962 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a38ab13-190d-40c5-9a69-389e2fae6884-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0a38ab13-190d-40c5-9a69-389e2fae6884" (UID: "0a38ab13-190d-40c5-9a69-389e2fae6884"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:00:15 crc kubenswrapper[4997]: I1205 07:00:15.879262 4997 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0a38ab13-190d-40c5-9a69-389e2fae6884-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:15 crc kubenswrapper[4997]: I1205 07:00:15.879326 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a38ab13-190d-40c5-9a69-389e2fae6884-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:15 crc kubenswrapper[4997]: I1205 07:00:15.879336 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a38ab13-190d-40c5-9a69-389e2fae6884-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:15 crc kubenswrapper[4997]: I1205 07:00:15.879350 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzjz6\" (UniqueName: \"kubernetes.io/projected/0a38ab13-190d-40c5-9a69-389e2fae6884-kube-api-access-rzjz6\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.086935 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.182559 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsqbk\" (UniqueName: \"kubernetes.io/projected/e6a372ba-59d9-4fcc-ba21-7b9de5316239-kube-api-access-jsqbk\") pod \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.182715 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-proxy-ca-bundles\") pod \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.182892 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6a372ba-59d9-4fcc-ba21-7b9de5316239-serving-cert\") pod \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.182947 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-client-ca\") pod \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.182997 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-config\") pod \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\" (UID: \"e6a372ba-59d9-4fcc-ba21-7b9de5316239\") " Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.184014 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e6a372ba-59d9-4fcc-ba21-7b9de5316239" (UID: "e6a372ba-59d9-4fcc-ba21-7b9de5316239"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.184037 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-client-ca" (OuterVolumeSpecName: "client-ca") pod "e6a372ba-59d9-4fcc-ba21-7b9de5316239" (UID: "e6a372ba-59d9-4fcc-ba21-7b9de5316239"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.184407 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-config" (OuterVolumeSpecName: "config") pod "e6a372ba-59d9-4fcc-ba21-7b9de5316239" (UID: "e6a372ba-59d9-4fcc-ba21-7b9de5316239"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.187317 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6a372ba-59d9-4fcc-ba21-7b9de5316239-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e6a372ba-59d9-4fcc-ba21-7b9de5316239" (UID: "e6a372ba-59d9-4fcc-ba21-7b9de5316239"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.188722 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6a372ba-59d9-4fcc-ba21-7b9de5316239-kube-api-access-jsqbk" (OuterVolumeSpecName: "kube-api-access-jsqbk") pod "e6a372ba-59d9-4fcc-ba21-7b9de5316239" (UID: "e6a372ba-59d9-4fcc-ba21-7b9de5316239"). InnerVolumeSpecName "kube-api-access-jsqbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.280826 4997 generic.go:334] "Generic (PLEG): container finished" podID="0a38ab13-190d-40c5-9a69-389e2fae6884" containerID="a5fe8cbc600bdc6d50b247ce9eeb3580052821f375efb95eb0d296e7169ffde4" exitCode=0 Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.280907 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" event={"ID":"0a38ab13-190d-40c5-9a69-389e2fae6884","Type":"ContainerDied","Data":"a5fe8cbc600bdc6d50b247ce9eeb3580052821f375efb95eb0d296e7169ffde4"} Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.280937 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" event={"ID":"0a38ab13-190d-40c5-9a69-389e2fae6884","Type":"ContainerDied","Data":"5ebbff2803e4c29ed107d52e30a61afd3238b1e455a3b09537e671087018c3f3"} Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.280956 4997 scope.go:117] "RemoveContainer" containerID="a5fe8cbc600bdc6d50b247ce9eeb3580052821f375efb95eb0d296e7169ffde4" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.280980 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.284846 4997 generic.go:334] "Generic (PLEG): container finished" podID="e6a372ba-59d9-4fcc-ba21-7b9de5316239" containerID="e185ef630ce8f8eb0f0eafee24a427eab0bff85e1965dff9023bab13a95480f3" exitCode=0 Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.284890 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.284928 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" event={"ID":"e6a372ba-59d9-4fcc-ba21-7b9de5316239","Type":"ContainerDied","Data":"e185ef630ce8f8eb0f0eafee24a427eab0bff85e1965dff9023bab13a95480f3"} Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.285016 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-57ksd" event={"ID":"e6a372ba-59d9-4fcc-ba21-7b9de5316239","Type":"ContainerDied","Data":"e6ef3c4bbe66efeeea728677ebda5fc005e6b7cb8728228bcc91270d9605c7f2"} Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.285392 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsqbk\" (UniqueName: \"kubernetes.io/projected/e6a372ba-59d9-4fcc-ba21-7b9de5316239-kube-api-access-jsqbk\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.285802 4997 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.285859 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6a372ba-59d9-4fcc-ba21-7b9de5316239-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.285890 4997 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.285916 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6a372ba-59d9-4fcc-ba21-7b9de5316239-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.298064 4997 scope.go:117] "RemoveContainer" containerID="a5fe8cbc600bdc6d50b247ce9eeb3580052821f375efb95eb0d296e7169ffde4" Dec 05 07:00:16 crc kubenswrapper[4997]: E1205 07:00:16.298633 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5fe8cbc600bdc6d50b247ce9eeb3580052821f375efb95eb0d296e7169ffde4\": container with ID starting with a5fe8cbc600bdc6d50b247ce9eeb3580052821f375efb95eb0d296e7169ffde4 not found: ID does not exist" containerID="a5fe8cbc600bdc6d50b247ce9eeb3580052821f375efb95eb0d296e7169ffde4" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.298728 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5fe8cbc600bdc6d50b247ce9eeb3580052821f375efb95eb0d296e7169ffde4"} err="failed to get container status \"a5fe8cbc600bdc6d50b247ce9eeb3580052821f375efb95eb0d296e7169ffde4\": rpc error: code = NotFound desc = could not find container \"a5fe8cbc600bdc6d50b247ce9eeb3580052821f375efb95eb0d296e7169ffde4\": container with ID starting with a5fe8cbc600bdc6d50b247ce9eeb3580052821f375efb95eb0d296e7169ffde4 not found: ID does not exist" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.298785 4997 scope.go:117] "RemoveContainer" containerID="e185ef630ce8f8eb0f0eafee24a427eab0bff85e1965dff9023bab13a95480f3" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.325949 4997 scope.go:117] "RemoveContainer" containerID="e185ef630ce8f8eb0f0eafee24a427eab0bff85e1965dff9023bab13a95480f3" Dec 05 07:00:16 crc kubenswrapper[4997]: E1205 07:00:16.328737 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e185ef630ce8f8eb0f0eafee24a427eab0bff85e1965dff9023bab13a95480f3\": container with ID starting with e185ef630ce8f8eb0f0eafee24a427eab0bff85e1965dff9023bab13a95480f3 not found: ID does not exist" containerID="e185ef630ce8f8eb0f0eafee24a427eab0bff85e1965dff9023bab13a95480f3" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.328925 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e185ef630ce8f8eb0f0eafee24a427eab0bff85e1965dff9023bab13a95480f3"} err="failed to get container status \"e185ef630ce8f8eb0f0eafee24a427eab0bff85e1965dff9023bab13a95480f3\": rpc error: code = NotFound desc = could not find container \"e185ef630ce8f8eb0f0eafee24a427eab0bff85e1965dff9023bab13a95480f3\": container with ID starting with e185ef630ce8f8eb0f0eafee24a427eab0bff85e1965dff9023bab13a95480f3 not found: ID does not exist" Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.329302 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh"] Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.334161 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-db5rh"] Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.337737 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-57ksd"] Dec 05 07:00:16 crc kubenswrapper[4997]: I1205 07:00:16.341087 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-57ksd"] Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.031652 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-56b87fd594-p55qg"] Dec 05 07:00:17 crc kubenswrapper[4997]: E1205 07:00:17.033245 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a38ab13-190d-40c5-9a69-389e2fae6884" containerName="route-controller-manager" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.033423 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a38ab13-190d-40c5-9a69-389e2fae6884" containerName="route-controller-manager" Dec 05 07:00:17 crc kubenswrapper[4997]: E1205 07:00:17.033540 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85d4925b-c3c9-423d-8bf4-0294e312da5c" containerName="collect-profiles" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.033715 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="85d4925b-c3c9-423d-8bf4-0294e312da5c" containerName="collect-profiles" Dec 05 07:00:17 crc kubenswrapper[4997]: E1205 07:00:17.033840 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6a372ba-59d9-4fcc-ba21-7b9de5316239" containerName="controller-manager" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.033930 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6a372ba-59d9-4fcc-ba21-7b9de5316239" containerName="controller-manager" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.034208 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="85d4925b-c3c9-423d-8bf4-0294e312da5c" containerName="collect-profiles" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.034494 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6a372ba-59d9-4fcc-ba21-7b9de5316239" containerName="controller-manager" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.034672 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a38ab13-190d-40c5-9a69-389e2fae6884" containerName="route-controller-manager" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.035442 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw"] Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.035707 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.036913 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.042986 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.043141 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.043754 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.043762 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.043877 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.044215 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.044417 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.044468 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.044430 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.044774 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.044890 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.045303 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.052251 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.052365 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw"] Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.056139 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56b87fd594-p55qg"] Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.199023 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-config\") pod \"route-controller-manager-558cd6b58-5d6cw\" (UID: \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.199097 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-serving-cert\") pod \"route-controller-manager-558cd6b58-5d6cw\" (UID: \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.199131 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flhn2\" (UniqueName: \"kubernetes.io/projected/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-kube-api-access-flhn2\") pod \"route-controller-manager-558cd6b58-5d6cw\" (UID: \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.199402 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-serving-cert\") pod \"controller-manager-56b87fd594-p55qg\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.199609 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcdzb\" (UniqueName: \"kubernetes.io/projected/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-kube-api-access-hcdzb\") pod \"controller-manager-56b87fd594-p55qg\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.199851 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-proxy-ca-bundles\") pod \"controller-manager-56b87fd594-p55qg\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.199905 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-config\") pod \"controller-manager-56b87fd594-p55qg\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.199946 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-client-ca\") pod \"route-controller-manager-558cd6b58-5d6cw\" (UID: \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.200011 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-client-ca\") pod \"controller-manager-56b87fd594-p55qg\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.301281 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-config\") pod \"controller-manager-56b87fd594-p55qg\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.301321 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-client-ca\") pod \"route-controller-manager-558cd6b58-5d6cw\" (UID: \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.301344 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-client-ca\") pod \"controller-manager-56b87fd594-p55qg\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.301381 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-config\") pod \"route-controller-manager-558cd6b58-5d6cw\" (UID: \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.301407 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-serving-cert\") pod \"route-controller-manager-558cd6b58-5d6cw\" (UID: \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.301423 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flhn2\" (UniqueName: \"kubernetes.io/projected/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-kube-api-access-flhn2\") pod \"route-controller-manager-558cd6b58-5d6cw\" (UID: \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.301463 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-serving-cert\") pod \"controller-manager-56b87fd594-p55qg\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.302868 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcdzb\" (UniqueName: \"kubernetes.io/projected/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-kube-api-access-hcdzb\") pod \"controller-manager-56b87fd594-p55qg\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.302898 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-proxy-ca-bundles\") pod \"controller-manager-56b87fd594-p55qg\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.303047 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-client-ca\") pod \"controller-manager-56b87fd594-p55qg\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.303462 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-config\") pod \"route-controller-manager-558cd6b58-5d6cw\" (UID: \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.303715 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-client-ca\") pod \"route-controller-manager-558cd6b58-5d6cw\" (UID: \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.304343 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-proxy-ca-bundles\") pod \"controller-manager-56b87fd594-p55qg\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.304418 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-config\") pod \"controller-manager-56b87fd594-p55qg\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.307095 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-serving-cert\") pod \"controller-manager-56b87fd594-p55qg\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.307656 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-serving-cert\") pod \"route-controller-manager-558cd6b58-5d6cw\" (UID: \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.319405 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flhn2\" (UniqueName: \"kubernetes.io/projected/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-kube-api-access-flhn2\") pod \"route-controller-manager-558cd6b58-5d6cw\" (UID: \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.319450 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcdzb\" (UniqueName: \"kubernetes.io/projected/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-kube-api-access-hcdzb\") pod \"controller-manager-56b87fd594-p55qg\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.423042 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.431588 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.700196 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56b87fd594-p55qg"] Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.731728 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw"] Dec 05 07:00:17 crc kubenswrapper[4997]: W1205 07:00:17.738128 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51cf49c8_f30d_4ba9_80b0_6410130ebbf0.slice/crio-30f8b5f79a390aebf52b05cfd854ada5a2ed4eb06055e2d3e066f1ca349c389a WatchSource:0}: Error finding container 30f8b5f79a390aebf52b05cfd854ada5a2ed4eb06055e2d3e066f1ca349c389a: Status 404 returned error can't find the container with id 30f8b5f79a390aebf52b05cfd854ada5a2ed4eb06055e2d3e066f1ca349c389a Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.756377 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a38ab13-190d-40c5-9a69-389e2fae6884" path="/var/lib/kubelet/pods/0a38ab13-190d-40c5-9a69-389e2fae6884/volumes" Dec 05 07:00:17 crc kubenswrapper[4997]: I1205 07:00:17.757250 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6a372ba-59d9-4fcc-ba21-7b9de5316239" path="/var/lib/kubelet/pods/e6a372ba-59d9-4fcc-ba21-7b9de5316239/volumes" Dec 05 07:00:18 crc kubenswrapper[4997]: I1205 07:00:18.304290 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" event={"ID":"51cf49c8-f30d-4ba9-80b0-6410130ebbf0","Type":"ContainerStarted","Data":"4d879c4e4f45fdf05b3c272753ee2e88743ecc94952a5c6dd7d0f654602d832c"} Dec 05 07:00:18 crc kubenswrapper[4997]: I1205 07:00:18.304854 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" event={"ID":"51cf49c8-f30d-4ba9-80b0-6410130ebbf0","Type":"ContainerStarted","Data":"30f8b5f79a390aebf52b05cfd854ada5a2ed4eb06055e2d3e066f1ca349c389a"} Dec 05 07:00:18 crc kubenswrapper[4997]: I1205 07:00:18.305175 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:18 crc kubenswrapper[4997]: I1205 07:00:18.306314 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" event={"ID":"d4d8cdd5-5a56-4fa0-bb23-919746ea7468","Type":"ContainerStarted","Data":"60cafea78376e1becb0dd836a153b3e0e773a5b66b50a0003f39c79331a1af74"} Dec 05 07:00:18 crc kubenswrapper[4997]: I1205 07:00:18.306373 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" event={"ID":"d4d8cdd5-5a56-4fa0-bb23-919746ea7468","Type":"ContainerStarted","Data":"1291d5c9471cba286f2fc8760e853db5fd36cbdda14fa0f5c4656d409936a0ce"} Dec 05 07:00:18 crc kubenswrapper[4997]: I1205 07:00:18.306679 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:18 crc kubenswrapper[4997]: I1205 07:00:18.311927 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:18 crc kubenswrapper[4997]: I1205 07:00:18.323474 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" podStartSLOduration=3.323448952 podStartE2EDuration="3.323448952s" podCreationTimestamp="2025-12-05 07:00:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:00:18.323277417 +0000 UTC m=+318.852184678" watchObservedRunningTime="2025-12-05 07:00:18.323448952 +0000 UTC m=+318.852356223" Dec 05 07:00:18 crc kubenswrapper[4997]: I1205 07:00:18.346977 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" podStartSLOduration=3.346951238 podStartE2EDuration="3.346951238s" podCreationTimestamp="2025-12-05 07:00:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:00:18.342373203 +0000 UTC m=+318.871280484" watchObservedRunningTime="2025-12-05 07:00:18.346951238 +0000 UTC m=+318.875858489" Dec 05 07:00:18 crc kubenswrapper[4997]: I1205 07:00:18.597033 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:23 crc kubenswrapper[4997]: I1205 07:00:23.606846 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-56b87fd594-p55qg"] Dec 05 07:00:23 crc kubenswrapper[4997]: I1205 07:00:23.607838 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" podUID="d4d8cdd5-5a56-4fa0-bb23-919746ea7468" containerName="controller-manager" containerID="cri-o://60cafea78376e1becb0dd836a153b3e0e773a5b66b50a0003f39c79331a1af74" gracePeriod=30 Dec 05 07:00:23 crc kubenswrapper[4997]: I1205 07:00:23.617384 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw"] Dec 05 07:00:23 crc kubenswrapper[4997]: I1205 07:00:23.617650 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" podUID="51cf49c8-f30d-4ba9-80b0-6410130ebbf0" containerName="route-controller-manager" containerID="cri-o://4d879c4e4f45fdf05b3c272753ee2e88743ecc94952a5c6dd7d0f654602d832c" gracePeriod=30 Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.152855 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.215337 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.220730 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flhn2\" (UniqueName: \"kubernetes.io/projected/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-kube-api-access-flhn2\") pod \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\" (UID: \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\") " Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.220778 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-client-ca\") pod \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\" (UID: \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\") " Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.220924 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-config\") pod \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\" (UID: \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\") " Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.220996 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-serving-cert\") pod \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\" (UID: \"51cf49c8-f30d-4ba9-80b0-6410130ebbf0\") " Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.221750 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-client-ca" (OuterVolumeSpecName: "client-ca") pod "51cf49c8-f30d-4ba9-80b0-6410130ebbf0" (UID: "51cf49c8-f30d-4ba9-80b0-6410130ebbf0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.221874 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-config" (OuterVolumeSpecName: "config") pod "51cf49c8-f30d-4ba9-80b0-6410130ebbf0" (UID: "51cf49c8-f30d-4ba9-80b0-6410130ebbf0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.222342 4997 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.222364 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.232269 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "51cf49c8-f30d-4ba9-80b0-6410130ebbf0" (UID: "51cf49c8-f30d-4ba9-80b0-6410130ebbf0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.232307 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-kube-api-access-flhn2" (OuterVolumeSpecName: "kube-api-access-flhn2") pod "51cf49c8-f30d-4ba9-80b0-6410130ebbf0" (UID: "51cf49c8-f30d-4ba9-80b0-6410130ebbf0"). InnerVolumeSpecName "kube-api-access-flhn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.322782 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcdzb\" (UniqueName: \"kubernetes.io/projected/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-kube-api-access-hcdzb\") pod \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.322864 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-config\") pod \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.322893 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-proxy-ca-bundles\") pod \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.322943 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-client-ca\") pod \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.323006 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-serving-cert\") pod \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\" (UID: \"d4d8cdd5-5a56-4fa0-bb23-919746ea7468\") " Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.323245 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.323262 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flhn2\" (UniqueName: \"kubernetes.io/projected/51cf49c8-f30d-4ba9-80b0-6410130ebbf0-kube-api-access-flhn2\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.324848 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d4d8cdd5-5a56-4fa0-bb23-919746ea7468" (UID: "d4d8cdd5-5a56-4fa0-bb23-919746ea7468"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.324977 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-config" (OuterVolumeSpecName: "config") pod "d4d8cdd5-5a56-4fa0-bb23-919746ea7468" (UID: "d4d8cdd5-5a56-4fa0-bb23-919746ea7468"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.325085 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-client-ca" (OuterVolumeSpecName: "client-ca") pod "d4d8cdd5-5a56-4fa0-bb23-919746ea7468" (UID: "d4d8cdd5-5a56-4fa0-bb23-919746ea7468"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.327642 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d4d8cdd5-5a56-4fa0-bb23-919746ea7468" (UID: "d4d8cdd5-5a56-4fa0-bb23-919746ea7468"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.327862 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-kube-api-access-hcdzb" (OuterVolumeSpecName: "kube-api-access-hcdzb") pod "d4d8cdd5-5a56-4fa0-bb23-919746ea7468" (UID: "d4d8cdd5-5a56-4fa0-bb23-919746ea7468"). InnerVolumeSpecName "kube-api-access-hcdzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.341963 4997 generic.go:334] "Generic (PLEG): container finished" podID="51cf49c8-f30d-4ba9-80b0-6410130ebbf0" containerID="4d879c4e4f45fdf05b3c272753ee2e88743ecc94952a5c6dd7d0f654602d832c" exitCode=0 Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.342018 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.342093 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" event={"ID":"51cf49c8-f30d-4ba9-80b0-6410130ebbf0","Type":"ContainerDied","Data":"4d879c4e4f45fdf05b3c272753ee2e88743ecc94952a5c6dd7d0f654602d832c"} Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.342166 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw" event={"ID":"51cf49c8-f30d-4ba9-80b0-6410130ebbf0","Type":"ContainerDied","Data":"30f8b5f79a390aebf52b05cfd854ada5a2ed4eb06055e2d3e066f1ca349c389a"} Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.342202 4997 scope.go:117] "RemoveContainer" containerID="4d879c4e4f45fdf05b3c272753ee2e88743ecc94952a5c6dd7d0f654602d832c" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.345064 4997 generic.go:334] "Generic (PLEG): container finished" podID="d4d8cdd5-5a56-4fa0-bb23-919746ea7468" containerID="60cafea78376e1becb0dd836a153b3e0e773a5b66b50a0003f39c79331a1af74" exitCode=0 Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.345107 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" event={"ID":"d4d8cdd5-5a56-4fa0-bb23-919746ea7468","Type":"ContainerDied","Data":"60cafea78376e1becb0dd836a153b3e0e773a5b66b50a0003f39c79331a1af74"} Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.345146 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" event={"ID":"d4d8cdd5-5a56-4fa0-bb23-919746ea7468","Type":"ContainerDied","Data":"1291d5c9471cba286f2fc8760e853db5fd36cbdda14fa0f5c4656d409936a0ce"} Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.345218 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56b87fd594-p55qg" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.370249 4997 scope.go:117] "RemoveContainer" containerID="4d879c4e4f45fdf05b3c272753ee2e88743ecc94952a5c6dd7d0f654602d832c" Dec 05 07:00:24 crc kubenswrapper[4997]: E1205 07:00:24.370775 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d879c4e4f45fdf05b3c272753ee2e88743ecc94952a5c6dd7d0f654602d832c\": container with ID starting with 4d879c4e4f45fdf05b3c272753ee2e88743ecc94952a5c6dd7d0f654602d832c not found: ID does not exist" containerID="4d879c4e4f45fdf05b3c272753ee2e88743ecc94952a5c6dd7d0f654602d832c" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.370812 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d879c4e4f45fdf05b3c272753ee2e88743ecc94952a5c6dd7d0f654602d832c"} err="failed to get container status \"4d879c4e4f45fdf05b3c272753ee2e88743ecc94952a5c6dd7d0f654602d832c\": rpc error: code = NotFound desc = could not find container \"4d879c4e4f45fdf05b3c272753ee2e88743ecc94952a5c6dd7d0f654602d832c\": container with ID starting with 4d879c4e4f45fdf05b3c272753ee2e88743ecc94952a5c6dd7d0f654602d832c not found: ID does not exist" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.370840 4997 scope.go:117] "RemoveContainer" containerID="60cafea78376e1becb0dd836a153b3e0e773a5b66b50a0003f39c79331a1af74" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.382689 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-56b87fd594-p55qg"] Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.386859 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-56b87fd594-p55qg"] Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.389581 4997 scope.go:117] "RemoveContainer" containerID="60cafea78376e1becb0dd836a153b3e0e773a5b66b50a0003f39c79331a1af74" Dec 05 07:00:24 crc kubenswrapper[4997]: E1205 07:00:24.390089 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60cafea78376e1becb0dd836a153b3e0e773a5b66b50a0003f39c79331a1af74\": container with ID starting with 60cafea78376e1becb0dd836a153b3e0e773a5b66b50a0003f39c79331a1af74 not found: ID does not exist" containerID="60cafea78376e1becb0dd836a153b3e0e773a5b66b50a0003f39c79331a1af74" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.390146 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60cafea78376e1becb0dd836a153b3e0e773a5b66b50a0003f39c79331a1af74"} err="failed to get container status \"60cafea78376e1becb0dd836a153b3e0e773a5b66b50a0003f39c79331a1af74\": rpc error: code = NotFound desc = could not find container \"60cafea78376e1becb0dd836a153b3e0e773a5b66b50a0003f39c79331a1af74\": container with ID starting with 60cafea78376e1becb0dd836a153b3e0e773a5b66b50a0003f39c79331a1af74 not found: ID does not exist" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.401438 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw"] Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.406450 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-558cd6b58-5d6cw"] Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.424549 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.424595 4997 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.424625 4997 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.424636 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:24 crc kubenswrapper[4997]: I1205 07:00:24.424647 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcdzb\" (UniqueName: \"kubernetes.io/projected/d4d8cdd5-5a56-4fa0-bb23-919746ea7468-kube-api-access-hcdzb\") on node \"crc\" DevicePath \"\"" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.042167 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5967c84899-qhb68"] Dec 05 07:00:25 crc kubenswrapper[4997]: E1205 07:00:25.042730 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4d8cdd5-5a56-4fa0-bb23-919746ea7468" containerName="controller-manager" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.042750 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4d8cdd5-5a56-4fa0-bb23-919746ea7468" containerName="controller-manager" Dec 05 07:00:25 crc kubenswrapper[4997]: E1205 07:00:25.042785 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51cf49c8-f30d-4ba9-80b0-6410130ebbf0" containerName="route-controller-manager" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.042794 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="51cf49c8-f30d-4ba9-80b0-6410130ebbf0" containerName="route-controller-manager" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.043052 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="51cf49c8-f30d-4ba9-80b0-6410130ebbf0" containerName="route-controller-manager" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.043077 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4d8cdd5-5a56-4fa0-bb23-919746ea7468" containerName="controller-manager" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.043823 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.045194 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq"] Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.048460 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.048683 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.049565 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.049904 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.057214 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.061600 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.061923 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.062128 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.062891 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.063326 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.063414 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.063596 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.063844 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.076847 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.079692 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5967c84899-qhb68"] Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.085486 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq"] Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.132960 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-serving-cert\") pod \"controller-manager-5967c84899-qhb68\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.133029 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnvzw\" (UniqueName: \"kubernetes.io/projected/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-kube-api-access-gnvzw\") pod \"controller-manager-5967c84899-qhb68\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.133067 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-proxy-ca-bundles\") pod \"controller-manager-5967c84899-qhb68\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.133105 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-config\") pod \"controller-manager-5967c84899-qhb68\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.133177 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa72e73f-826e-43ac-b7a3-004fbda587a7-client-ca\") pod \"route-controller-manager-76b6b65598-r2rmq\" (UID: \"fa72e73f-826e-43ac-b7a3-004fbda587a7\") " pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.133202 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-client-ca\") pod \"controller-manager-5967c84899-qhb68\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.133244 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa72e73f-826e-43ac-b7a3-004fbda587a7-config\") pod \"route-controller-manager-76b6b65598-r2rmq\" (UID: \"fa72e73f-826e-43ac-b7a3-004fbda587a7\") " pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.133266 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqvhs\" (UniqueName: \"kubernetes.io/projected/fa72e73f-826e-43ac-b7a3-004fbda587a7-kube-api-access-hqvhs\") pod \"route-controller-manager-76b6b65598-r2rmq\" (UID: \"fa72e73f-826e-43ac-b7a3-004fbda587a7\") " pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.133372 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa72e73f-826e-43ac-b7a3-004fbda587a7-serving-cert\") pod \"route-controller-manager-76b6b65598-r2rmq\" (UID: \"fa72e73f-826e-43ac-b7a3-004fbda587a7\") " pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.234227 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa72e73f-826e-43ac-b7a3-004fbda587a7-client-ca\") pod \"route-controller-manager-76b6b65598-r2rmq\" (UID: \"fa72e73f-826e-43ac-b7a3-004fbda587a7\") " pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.234296 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-client-ca\") pod \"controller-manager-5967c84899-qhb68\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.234348 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa72e73f-826e-43ac-b7a3-004fbda587a7-config\") pod \"route-controller-manager-76b6b65598-r2rmq\" (UID: \"fa72e73f-826e-43ac-b7a3-004fbda587a7\") " pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.234369 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqvhs\" (UniqueName: \"kubernetes.io/projected/fa72e73f-826e-43ac-b7a3-004fbda587a7-kube-api-access-hqvhs\") pod \"route-controller-manager-76b6b65598-r2rmq\" (UID: \"fa72e73f-826e-43ac-b7a3-004fbda587a7\") " pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.234397 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa72e73f-826e-43ac-b7a3-004fbda587a7-serving-cert\") pod \"route-controller-manager-76b6b65598-r2rmq\" (UID: \"fa72e73f-826e-43ac-b7a3-004fbda587a7\") " pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.234444 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-serving-cert\") pod \"controller-manager-5967c84899-qhb68\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.234466 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnvzw\" (UniqueName: \"kubernetes.io/projected/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-kube-api-access-gnvzw\") pod \"controller-manager-5967c84899-qhb68\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.234493 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-proxy-ca-bundles\") pod \"controller-manager-5967c84899-qhb68\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.234521 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-config\") pod \"controller-manager-5967c84899-qhb68\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.235575 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-client-ca\") pod \"controller-manager-5967c84899-qhb68\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.235911 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa72e73f-826e-43ac-b7a3-004fbda587a7-client-ca\") pod \"route-controller-manager-76b6b65598-r2rmq\" (UID: \"fa72e73f-826e-43ac-b7a3-004fbda587a7\") " pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.236352 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-config\") pod \"controller-manager-5967c84899-qhb68\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.236727 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-proxy-ca-bundles\") pod \"controller-manager-5967c84899-qhb68\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.237876 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa72e73f-826e-43ac-b7a3-004fbda587a7-config\") pod \"route-controller-manager-76b6b65598-r2rmq\" (UID: \"fa72e73f-826e-43ac-b7a3-004fbda587a7\") " pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.238339 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-serving-cert\") pod \"controller-manager-5967c84899-qhb68\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.239753 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa72e73f-826e-43ac-b7a3-004fbda587a7-serving-cert\") pod \"route-controller-manager-76b6b65598-r2rmq\" (UID: \"fa72e73f-826e-43ac-b7a3-004fbda587a7\") " pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.258415 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnvzw\" (UniqueName: \"kubernetes.io/projected/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-kube-api-access-gnvzw\") pod \"controller-manager-5967c84899-qhb68\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.259033 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqvhs\" (UniqueName: \"kubernetes.io/projected/fa72e73f-826e-43ac-b7a3-004fbda587a7-kube-api-access-hqvhs\") pod \"route-controller-manager-76b6b65598-r2rmq\" (UID: \"fa72e73f-826e-43ac-b7a3-004fbda587a7\") " pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.380076 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.397676 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.614043 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5967c84899-qhb68"] Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.697583 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq"] Dec 05 07:00:25 crc kubenswrapper[4997]: W1205 07:00:25.702660 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa72e73f_826e_43ac_b7a3_004fbda587a7.slice/crio-fcf99ea3127029958274f4a22cb2953d244004dc039bb2b0dd9a3f53cb458fc6 WatchSource:0}: Error finding container fcf99ea3127029958274f4a22cb2953d244004dc039bb2b0dd9a3f53cb458fc6: Status 404 returned error can't find the container with id fcf99ea3127029958274f4a22cb2953d244004dc039bb2b0dd9a3f53cb458fc6 Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.757238 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51cf49c8-f30d-4ba9-80b0-6410130ebbf0" path="/var/lib/kubelet/pods/51cf49c8-f30d-4ba9-80b0-6410130ebbf0/volumes" Dec 05 07:00:25 crc kubenswrapper[4997]: I1205 07:00:25.758332 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4d8cdd5-5a56-4fa0-bb23-919746ea7468" path="/var/lib/kubelet/pods/d4d8cdd5-5a56-4fa0-bb23-919746ea7468/volumes" Dec 05 07:00:26 crc kubenswrapper[4997]: I1205 07:00:26.361296 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" event={"ID":"496d9d2d-9c90-431b-b3bd-0cddc82a0edc","Type":"ContainerStarted","Data":"6517d0810e851f41656cafc729cb97cb92d7faf1d065ff8af7ca1e728a465064"} Dec 05 07:00:26 crc kubenswrapper[4997]: I1205 07:00:26.361740 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" event={"ID":"496d9d2d-9c90-431b-b3bd-0cddc82a0edc","Type":"ContainerStarted","Data":"2555407e4db4cfbd834ac7c5485c4a7b767a4ee9f54463b92d155afb88b7d7f4"} Dec 05 07:00:26 crc kubenswrapper[4997]: I1205 07:00:26.361763 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:26 crc kubenswrapper[4997]: I1205 07:00:26.363123 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" event={"ID":"fa72e73f-826e-43ac-b7a3-004fbda587a7","Type":"ContainerStarted","Data":"b886daea218d5413c47447ef2db87e73d2cf22fb2d7be4d49f24074af86b7edd"} Dec 05 07:00:26 crc kubenswrapper[4997]: I1205 07:00:26.363149 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" event={"ID":"fa72e73f-826e-43ac-b7a3-004fbda587a7","Type":"ContainerStarted","Data":"fcf99ea3127029958274f4a22cb2953d244004dc039bb2b0dd9a3f53cb458fc6"} Dec 05 07:00:26 crc kubenswrapper[4997]: I1205 07:00:26.363382 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" Dec 05 07:00:26 crc kubenswrapper[4997]: I1205 07:00:26.369091 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" Dec 05 07:00:26 crc kubenswrapper[4997]: I1205 07:00:26.370724 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:00:26 crc kubenswrapper[4997]: I1205 07:00:26.400132 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" podStartSLOduration=3.400095103 podStartE2EDuration="3.400095103s" podCreationTimestamp="2025-12-05 07:00:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:00:26.396686954 +0000 UTC m=+326.925594235" watchObservedRunningTime="2025-12-05 07:00:26.400095103 +0000 UTC m=+326.929002374" Dec 05 07:00:26 crc kubenswrapper[4997]: I1205 07:00:26.478670 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-76b6b65598-r2rmq" podStartSLOduration=3.478648268 podStartE2EDuration="3.478648268s" podCreationTimestamp="2025-12-05 07:00:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:00:26.47420919 +0000 UTC m=+327.003116471" watchObservedRunningTime="2025-12-05 07:00:26.478648268 +0000 UTC m=+327.007555539" Dec 05 07:00:49 crc kubenswrapper[4997]: I1205 07:00:49.770481 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:00:49 crc kubenswrapper[4997]: I1205 07:00:49.771122 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.339805 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-h8lqf"] Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.341767 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.356868 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-h8lqf"] Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.430535 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96e0bc9b-7166-424c-b256-85a0edb60092-trusted-ca\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.430598 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mz75\" (UniqueName: \"kubernetes.io/projected/96e0bc9b-7166-424c-b256-85a0edb60092-kube-api-access-7mz75\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.430703 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.430741 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96e0bc9b-7166-424c-b256-85a0edb60092-bound-sa-token\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.430867 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/96e0bc9b-7166-424c-b256-85a0edb60092-registry-tls\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.430946 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/96e0bc9b-7166-424c-b256-85a0edb60092-installation-pull-secrets\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.431005 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/96e0bc9b-7166-424c-b256-85a0edb60092-ca-trust-extracted\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.431028 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/96e0bc9b-7166-424c-b256-85a0edb60092-registry-certificates\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.459930 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.534208 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mz75\" (UniqueName: \"kubernetes.io/projected/96e0bc9b-7166-424c-b256-85a0edb60092-kube-api-access-7mz75\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.534293 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96e0bc9b-7166-424c-b256-85a0edb60092-bound-sa-token\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.534347 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/96e0bc9b-7166-424c-b256-85a0edb60092-registry-tls\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.534374 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/96e0bc9b-7166-424c-b256-85a0edb60092-installation-pull-secrets\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.534403 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/96e0bc9b-7166-424c-b256-85a0edb60092-ca-trust-extracted\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.534426 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/96e0bc9b-7166-424c-b256-85a0edb60092-registry-certificates\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.534455 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96e0bc9b-7166-424c-b256-85a0edb60092-trusted-ca\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.535732 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96e0bc9b-7166-424c-b256-85a0edb60092-trusted-ca\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.536166 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/96e0bc9b-7166-424c-b256-85a0edb60092-ca-trust-extracted\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.536823 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/96e0bc9b-7166-424c-b256-85a0edb60092-registry-certificates\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.545319 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/96e0bc9b-7166-424c-b256-85a0edb60092-installation-pull-secrets\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.557819 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/96e0bc9b-7166-424c-b256-85a0edb60092-registry-tls\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.558963 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96e0bc9b-7166-424c-b256-85a0edb60092-bound-sa-token\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.565144 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mz75\" (UniqueName: \"kubernetes.io/projected/96e0bc9b-7166-424c-b256-85a0edb60092-kube-api-access-7mz75\") pod \"image-registry-66df7c8f76-h8lqf\" (UID: \"96e0bc9b-7166-424c-b256-85a0edb60092\") " pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:04 crc kubenswrapper[4997]: I1205 07:01:04.658470 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:05 crc kubenswrapper[4997]: I1205 07:01:05.097757 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-h8lqf"] Dec 05 07:01:05 crc kubenswrapper[4997]: I1205 07:01:05.601549 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" event={"ID":"96e0bc9b-7166-424c-b256-85a0edb60092","Type":"ContainerStarted","Data":"5ac2d73a3321b4a59d2742f1d2fdba84762f1720719be1bbb52c6455d214b957"} Dec 05 07:01:05 crc kubenswrapper[4997]: I1205 07:01:05.601982 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:05 crc kubenswrapper[4997]: I1205 07:01:05.602198 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" event={"ID":"96e0bc9b-7166-424c-b256-85a0edb60092","Type":"ContainerStarted","Data":"5d50439013af6f7a59cf1ae40534e846f179fb14edf0f782e8a7f71f87cf82a0"} Dec 05 07:01:05 crc kubenswrapper[4997]: I1205 07:01:05.624453 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" podStartSLOduration=1.624432221 podStartE2EDuration="1.624432221s" podCreationTimestamp="2025-12-05 07:01:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:01:05.621538938 +0000 UTC m=+366.150446219" watchObservedRunningTime="2025-12-05 07:01:05.624432221 +0000 UTC m=+366.153339482" Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.230194 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5967c84899-qhb68"] Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.231541 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" podUID="496d9d2d-9c90-431b-b3bd-0cddc82a0edc" containerName="controller-manager" containerID="cri-o://6517d0810e851f41656cafc729cb97cb92d7faf1d065ff8af7ca1e728a465064" gracePeriod=30 Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.380868 4997 patch_prober.go:28] interesting pod/controller-manager-5967c84899-qhb68 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.61:8443/healthz\": dial tcp 10.217.0.61:8443: connect: connection refused" start-of-body= Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.381295 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" podUID="496d9d2d-9c90-431b-b3bd-0cddc82a0edc" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.61:8443/healthz\": dial tcp 10.217.0.61:8443: connect: connection refused" Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.604574 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.660231 4997 generic.go:334] "Generic (PLEG): container finished" podID="496d9d2d-9c90-431b-b3bd-0cddc82a0edc" containerID="6517d0810e851f41656cafc729cb97cb92d7faf1d065ff8af7ca1e728a465064" exitCode=0 Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.660282 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" event={"ID":"496d9d2d-9c90-431b-b3bd-0cddc82a0edc","Type":"ContainerDied","Data":"6517d0810e851f41656cafc729cb97cb92d7faf1d065ff8af7ca1e728a465064"} Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.660320 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.660334 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5967c84899-qhb68" event={"ID":"496d9d2d-9c90-431b-b3bd-0cddc82a0edc","Type":"ContainerDied","Data":"2555407e4db4cfbd834ac7c5485c4a7b767a4ee9f54463b92d155afb88b7d7f4"} Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.660353 4997 scope.go:117] "RemoveContainer" containerID="6517d0810e851f41656cafc729cb97cb92d7faf1d065ff8af7ca1e728a465064" Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.678039 4997 scope.go:117] "RemoveContainer" containerID="6517d0810e851f41656cafc729cb97cb92d7faf1d065ff8af7ca1e728a465064" Dec 05 07:01:15 crc kubenswrapper[4997]: E1205 07:01:15.678665 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6517d0810e851f41656cafc729cb97cb92d7faf1d065ff8af7ca1e728a465064\": container with ID starting with 6517d0810e851f41656cafc729cb97cb92d7faf1d065ff8af7ca1e728a465064 not found: ID does not exist" containerID="6517d0810e851f41656cafc729cb97cb92d7faf1d065ff8af7ca1e728a465064" Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.678714 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6517d0810e851f41656cafc729cb97cb92d7faf1d065ff8af7ca1e728a465064"} err="failed to get container status \"6517d0810e851f41656cafc729cb97cb92d7faf1d065ff8af7ca1e728a465064\": rpc error: code = NotFound desc = could not find container \"6517d0810e851f41656cafc729cb97cb92d7faf1d065ff8af7ca1e728a465064\": container with ID starting with 6517d0810e851f41656cafc729cb97cb92d7faf1d065ff8af7ca1e728a465064 not found: ID does not exist" Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.804348 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-serving-cert\") pod \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.804438 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnvzw\" (UniqueName: \"kubernetes.io/projected/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-kube-api-access-gnvzw\") pod \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.804464 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-proxy-ca-bundles\") pod \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.804938 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-config\") pod \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.805099 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-client-ca\") pod \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\" (UID: \"496d9d2d-9c90-431b-b3bd-0cddc82a0edc\") " Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.805740 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "496d9d2d-9c90-431b-b3bd-0cddc82a0edc" (UID: "496d9d2d-9c90-431b-b3bd-0cddc82a0edc"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.805987 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-config" (OuterVolumeSpecName: "config") pod "496d9d2d-9c90-431b-b3bd-0cddc82a0edc" (UID: "496d9d2d-9c90-431b-b3bd-0cddc82a0edc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.806660 4997 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.807009 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-client-ca" (OuterVolumeSpecName: "client-ca") pod "496d9d2d-9c90-431b-b3bd-0cddc82a0edc" (UID: "496d9d2d-9c90-431b-b3bd-0cddc82a0edc"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.811542 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-kube-api-access-gnvzw" (OuterVolumeSpecName: "kube-api-access-gnvzw") pod "496d9d2d-9c90-431b-b3bd-0cddc82a0edc" (UID: "496d9d2d-9c90-431b-b3bd-0cddc82a0edc"). InnerVolumeSpecName "kube-api-access-gnvzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.812657 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496d9d2d-9c90-431b-b3bd-0cddc82a0edc" (UID: "496d9d2d-9c90-431b-b3bd-0cddc82a0edc"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.907303 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnvzw\" (UniqueName: \"kubernetes.io/projected/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-kube-api-access-gnvzw\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.907341 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.907352 4997 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:15 crc kubenswrapper[4997]: I1205 07:01:15.907363 4997 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496d9d2d-9c90-431b-b3bd-0cddc82a0edc-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:16 crc kubenswrapper[4997]: I1205 07:01:16.016682 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5967c84899-qhb68"] Dec 05 07:01:16 crc kubenswrapper[4997]: I1205 07:01:16.021922 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5967c84899-qhb68"] Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.079445 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-59675fc847-vsf42"] Dec 05 07:01:17 crc kubenswrapper[4997]: E1205 07:01:17.080260 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="496d9d2d-9c90-431b-b3bd-0cddc82a0edc" containerName="controller-manager" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.080282 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="496d9d2d-9c90-431b-b3bd-0cddc82a0edc" containerName="controller-manager" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.080419 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="496d9d2d-9c90-431b-b3bd-0cddc82a0edc" containerName="controller-manager" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.080951 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.083418 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.083789 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.083868 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.083964 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.084183 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.084317 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.094154 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-59675fc847-vsf42"] Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.098174 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.228119 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b8cfae45-96a6-4f80-acc0-3188f451cd95-proxy-ca-bundles\") pod \"controller-manager-59675fc847-vsf42\" (UID: \"b8cfae45-96a6-4f80-acc0-3188f451cd95\") " pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.228176 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8cfae45-96a6-4f80-acc0-3188f451cd95-client-ca\") pod \"controller-manager-59675fc847-vsf42\" (UID: \"b8cfae45-96a6-4f80-acc0-3188f451cd95\") " pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.228283 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8cfae45-96a6-4f80-acc0-3188f451cd95-serving-cert\") pod \"controller-manager-59675fc847-vsf42\" (UID: \"b8cfae45-96a6-4f80-acc0-3188f451cd95\") " pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.228423 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8cfae45-96a6-4f80-acc0-3188f451cd95-config\") pod \"controller-manager-59675fc847-vsf42\" (UID: \"b8cfae45-96a6-4f80-acc0-3188f451cd95\") " pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.228510 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-877v8\" (UniqueName: \"kubernetes.io/projected/b8cfae45-96a6-4f80-acc0-3188f451cd95-kube-api-access-877v8\") pod \"controller-manager-59675fc847-vsf42\" (UID: \"b8cfae45-96a6-4f80-acc0-3188f451cd95\") " pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.329414 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b8cfae45-96a6-4f80-acc0-3188f451cd95-proxy-ca-bundles\") pod \"controller-manager-59675fc847-vsf42\" (UID: \"b8cfae45-96a6-4f80-acc0-3188f451cd95\") " pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.329473 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8cfae45-96a6-4f80-acc0-3188f451cd95-client-ca\") pod \"controller-manager-59675fc847-vsf42\" (UID: \"b8cfae45-96a6-4f80-acc0-3188f451cd95\") " pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.329529 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8cfae45-96a6-4f80-acc0-3188f451cd95-serving-cert\") pod \"controller-manager-59675fc847-vsf42\" (UID: \"b8cfae45-96a6-4f80-acc0-3188f451cd95\") " pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.329549 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8cfae45-96a6-4f80-acc0-3188f451cd95-config\") pod \"controller-manager-59675fc847-vsf42\" (UID: \"b8cfae45-96a6-4f80-acc0-3188f451cd95\") " pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.330663 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-877v8\" (UniqueName: \"kubernetes.io/projected/b8cfae45-96a6-4f80-acc0-3188f451cd95-kube-api-access-877v8\") pod \"controller-manager-59675fc847-vsf42\" (UID: \"b8cfae45-96a6-4f80-acc0-3188f451cd95\") " pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.331160 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8cfae45-96a6-4f80-acc0-3188f451cd95-client-ca\") pod \"controller-manager-59675fc847-vsf42\" (UID: \"b8cfae45-96a6-4f80-acc0-3188f451cd95\") " pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.331339 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8cfae45-96a6-4f80-acc0-3188f451cd95-config\") pod \"controller-manager-59675fc847-vsf42\" (UID: \"b8cfae45-96a6-4f80-acc0-3188f451cd95\") " pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.331663 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b8cfae45-96a6-4f80-acc0-3188f451cd95-proxy-ca-bundles\") pod \"controller-manager-59675fc847-vsf42\" (UID: \"b8cfae45-96a6-4f80-acc0-3188f451cd95\") " pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.339510 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8cfae45-96a6-4f80-acc0-3188f451cd95-serving-cert\") pod \"controller-manager-59675fc847-vsf42\" (UID: \"b8cfae45-96a6-4f80-acc0-3188f451cd95\") " pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.362997 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-877v8\" (UniqueName: \"kubernetes.io/projected/b8cfae45-96a6-4f80-acc0-3188f451cd95-kube-api-access-877v8\") pod \"controller-manager-59675fc847-vsf42\" (UID: \"b8cfae45-96a6-4f80-acc0-3188f451cd95\") " pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.399346 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.718368 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-59675fc847-vsf42"] Dec 05 07:01:17 crc kubenswrapper[4997]: I1205 07:01:17.770220 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496d9d2d-9c90-431b-b3bd-0cddc82a0edc" path="/var/lib/kubelet/pods/496d9d2d-9c90-431b-b3bd-0cddc82a0edc/volumes" Dec 05 07:01:18 crc kubenswrapper[4997]: I1205 07:01:18.686407 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" event={"ID":"b8cfae45-96a6-4f80-acc0-3188f451cd95","Type":"ContainerStarted","Data":"f1f65ef48ad303104509bea534aaf9790c7613e357f8fe672f07d6c7acd097f4"} Dec 05 07:01:18 crc kubenswrapper[4997]: I1205 07:01:18.686803 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:18 crc kubenswrapper[4997]: I1205 07:01:18.686816 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" event={"ID":"b8cfae45-96a6-4f80-acc0-3188f451cd95","Type":"ContainerStarted","Data":"415fdba65fde68f497b604e4fe294cf7ab74c69558d69c679f8cff567913710d"} Dec 05 07:01:18 crc kubenswrapper[4997]: I1205 07:01:18.694688 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" Dec 05 07:01:18 crc kubenswrapper[4997]: I1205 07:01:18.733912 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-59675fc847-vsf42" podStartSLOduration=3.733889654 podStartE2EDuration="3.733889654s" podCreationTimestamp="2025-12-05 07:01:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:01:18.710788798 +0000 UTC m=+379.239696059" watchObservedRunningTime="2025-12-05 07:01:18.733889654 +0000 UTC m=+379.262796905" Dec 05 07:01:19 crc kubenswrapper[4997]: I1205 07:01:19.770398 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:01:19 crc kubenswrapper[4997]: I1205 07:01:19.770927 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:01:24 crc kubenswrapper[4997]: I1205 07:01:24.666085 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-h8lqf" Dec 05 07:01:24 crc kubenswrapper[4997]: I1205 07:01:24.733036 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5xbd9"] Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.667796 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n6tmw"] Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.669210 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n6tmw" podUID="c0ca09da-0456-4ab9-8260-03c64893a6dd" containerName="registry-server" containerID="cri-o://ebb32219ca7297fc6f6e84a88a97517ec975997f7f3538a900818823da9b3a34" gracePeriod=30 Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.674966 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2pls6"] Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.675316 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2pls6" podUID="3ec5dfd3-3221-4141-a2d4-aea02733a985" containerName="registry-server" containerID="cri-o://4e95c9a0122f9f97de73329b76fe8d535a38995655562840fb0aafe6fd7824a1" gracePeriod=30 Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.688035 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k9twj"] Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.688481 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" podUID="17fe9bd5-5973-47bd-945f-55bb8caa9d65" containerName="marketplace-operator" containerID="cri-o://99b851f62ee198eeaadf857187e158d3889e72c84fe6d9c43f641d75b03a05df" gracePeriod=30 Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.704376 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6zt6h"] Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.704771 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6zt6h" podUID="19ea766c-8ab7-4499-a51b-cd6968a62a76" containerName="registry-server" containerID="cri-o://485e08d238a48edaa6a607645ffb2b538173a81e733c867397c153bde926b98c" gracePeriod=30 Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.707597 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k4n89"] Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.708329 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k4n89" Dec 05 07:01:44 crc kubenswrapper[4997]: E1205 07:01:44.709039 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4e95c9a0122f9f97de73329b76fe8d535a38995655562840fb0aafe6fd7824a1" cmd=["grpc_health_probe","-addr=:50051"] Dec 05 07:01:44 crc kubenswrapper[4997]: E1205 07:01:44.715584 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4e95c9a0122f9f97de73329b76fe8d535a38995655562840fb0aafe6fd7824a1" cmd=["grpc_health_probe","-addr=:50051"] Dec 05 07:01:44 crc kubenswrapper[4997]: E1205 07:01:44.719199 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4e95c9a0122f9f97de73329b76fe8d535a38995655562840fb0aafe6fd7824a1 is running failed: container process not found" containerID="4e95c9a0122f9f97de73329b76fe8d535a38995655562840fb0aafe6fd7824a1" cmd=["grpc_health_probe","-addr=:50051"] Dec 05 07:01:44 crc kubenswrapper[4997]: E1205 07:01:44.719262 4997 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4e95c9a0122f9f97de73329b76fe8d535a38995655562840fb0aafe6fd7824a1 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-2pls6" podUID="3ec5dfd3-3221-4141-a2d4-aea02733a985" containerName="registry-server" Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.719978 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sqj49"] Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.720211 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sqj49" podUID="f67be9f3-d0bd-439c-ab66-72a6a29b3104" containerName="registry-server" containerID="cri-o://bb9a8fd6335f6cdc2749befa3b807f9f84860609bd15b8e49189cf40c00dc69f" gracePeriod=30 Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.737506 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k4n89"] Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.847398 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14ea843e-5c09-4461-88b4-4810e6d28cca-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k4n89\" (UID: \"14ea843e-5c09-4461-88b4-4810e6d28cca\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4n89" Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.847466 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/14ea843e-5c09-4461-88b4-4810e6d28cca-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k4n89\" (UID: \"14ea843e-5c09-4461-88b4-4810e6d28cca\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4n89" Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.847703 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxsrt\" (UniqueName: \"kubernetes.io/projected/14ea843e-5c09-4461-88b4-4810e6d28cca-kube-api-access-mxsrt\") pod \"marketplace-operator-79b997595-k4n89\" (UID: \"14ea843e-5c09-4461-88b4-4810e6d28cca\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4n89" Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.900010 4997 generic.go:334] "Generic (PLEG): container finished" podID="c0ca09da-0456-4ab9-8260-03c64893a6dd" containerID="ebb32219ca7297fc6f6e84a88a97517ec975997f7f3538a900818823da9b3a34" exitCode=0 Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.900086 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6tmw" event={"ID":"c0ca09da-0456-4ab9-8260-03c64893a6dd","Type":"ContainerDied","Data":"ebb32219ca7297fc6f6e84a88a97517ec975997f7f3538a900818823da9b3a34"} Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.929259 4997 generic.go:334] "Generic (PLEG): container finished" podID="f67be9f3-d0bd-439c-ab66-72a6a29b3104" containerID="bb9a8fd6335f6cdc2749befa3b807f9f84860609bd15b8e49189cf40c00dc69f" exitCode=0 Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.929372 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqj49" event={"ID":"f67be9f3-d0bd-439c-ab66-72a6a29b3104","Type":"ContainerDied","Data":"bb9a8fd6335f6cdc2749befa3b807f9f84860609bd15b8e49189cf40c00dc69f"} Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.949299 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14ea843e-5c09-4461-88b4-4810e6d28cca-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k4n89\" (UID: \"14ea843e-5c09-4461-88b4-4810e6d28cca\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4n89" Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.949366 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/14ea843e-5c09-4461-88b4-4810e6d28cca-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k4n89\" (UID: \"14ea843e-5c09-4461-88b4-4810e6d28cca\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4n89" Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.949430 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxsrt\" (UniqueName: \"kubernetes.io/projected/14ea843e-5c09-4461-88b4-4810e6d28cca-kube-api-access-mxsrt\") pod \"marketplace-operator-79b997595-k4n89\" (UID: \"14ea843e-5c09-4461-88b4-4810e6d28cca\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4n89" Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.950750 4997 generic.go:334] "Generic (PLEG): container finished" podID="3ec5dfd3-3221-4141-a2d4-aea02733a985" containerID="4e95c9a0122f9f97de73329b76fe8d535a38995655562840fb0aafe6fd7824a1" exitCode=0 Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.950865 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2pls6" event={"ID":"3ec5dfd3-3221-4141-a2d4-aea02733a985","Type":"ContainerDied","Data":"4e95c9a0122f9f97de73329b76fe8d535a38995655562840fb0aafe6fd7824a1"} Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.951121 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14ea843e-5c09-4461-88b4-4810e6d28cca-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k4n89\" (UID: \"14ea843e-5c09-4461-88b4-4810e6d28cca\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4n89" Dec 05 07:01:44 crc kubenswrapper[4997]: E1205 07:01:44.952230 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ebb32219ca7297fc6f6e84a88a97517ec975997f7f3538a900818823da9b3a34 is running failed: container process not found" containerID="ebb32219ca7297fc6f6e84a88a97517ec975997f7f3538a900818823da9b3a34" cmd=["grpc_health_probe","-addr=:50051"] Dec 05 07:01:44 crc kubenswrapper[4997]: E1205 07:01:44.953019 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ebb32219ca7297fc6f6e84a88a97517ec975997f7f3538a900818823da9b3a34 is running failed: container process not found" containerID="ebb32219ca7297fc6f6e84a88a97517ec975997f7f3538a900818823da9b3a34" cmd=["grpc_health_probe","-addr=:50051"] Dec 05 07:01:44 crc kubenswrapper[4997]: E1205 07:01:44.953742 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ebb32219ca7297fc6f6e84a88a97517ec975997f7f3538a900818823da9b3a34 is running failed: container process not found" containerID="ebb32219ca7297fc6f6e84a88a97517ec975997f7f3538a900818823da9b3a34" cmd=["grpc_health_probe","-addr=:50051"] Dec 05 07:01:44 crc kubenswrapper[4997]: E1205 07:01:44.953781 4997 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ebb32219ca7297fc6f6e84a88a97517ec975997f7f3538a900818823da9b3a34 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-n6tmw" podUID="c0ca09da-0456-4ab9-8260-03c64893a6dd" containerName="registry-server" Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.955694 4997 generic.go:334] "Generic (PLEG): container finished" podID="17fe9bd5-5973-47bd-945f-55bb8caa9d65" containerID="99b851f62ee198eeaadf857187e158d3889e72c84fe6d9c43f641d75b03a05df" exitCode=0 Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.955771 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" event={"ID":"17fe9bd5-5973-47bd-945f-55bb8caa9d65","Type":"ContainerDied","Data":"99b851f62ee198eeaadf857187e158d3889e72c84fe6d9c43f641d75b03a05df"} Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.956519 4997 scope.go:117] "RemoveContainer" containerID="feb7b9a4853bb63a72df3919b0a2cb6b319deb2104e5684bc736c65bf79d9844" Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.959349 4997 generic.go:334] "Generic (PLEG): container finished" podID="19ea766c-8ab7-4499-a51b-cd6968a62a76" containerID="485e08d238a48edaa6a607645ffb2b538173a81e733c867397c153bde926b98c" exitCode=0 Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.959457 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zt6h" event={"ID":"19ea766c-8ab7-4499-a51b-cd6968a62a76","Type":"ContainerDied","Data":"485e08d238a48edaa6a607645ffb2b538173a81e733c867397c153bde926b98c"} Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.962902 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/14ea843e-5c09-4461-88b4-4810e6d28cca-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k4n89\" (UID: \"14ea843e-5c09-4461-88b4-4810e6d28cca\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4n89" Dec 05 07:01:44 crc kubenswrapper[4997]: I1205 07:01:44.968105 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxsrt\" (UniqueName: \"kubernetes.io/projected/14ea843e-5c09-4461-88b4-4810e6d28cca-kube-api-access-mxsrt\") pod \"marketplace-operator-79b997595-k4n89\" (UID: \"14ea843e-5c09-4461-88b4-4810e6d28cca\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4n89" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.037816 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k4n89" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.284033 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6tmw" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.461796 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9vmz\" (UniqueName: \"kubernetes.io/projected/c0ca09da-0456-4ab9-8260-03c64893a6dd-kube-api-access-t9vmz\") pod \"c0ca09da-0456-4ab9-8260-03c64893a6dd\" (UID: \"c0ca09da-0456-4ab9-8260-03c64893a6dd\") " Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.461899 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0ca09da-0456-4ab9-8260-03c64893a6dd-catalog-content\") pod \"c0ca09da-0456-4ab9-8260-03c64893a6dd\" (UID: \"c0ca09da-0456-4ab9-8260-03c64893a6dd\") " Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.461964 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0ca09da-0456-4ab9-8260-03c64893a6dd-utilities\") pod \"c0ca09da-0456-4ab9-8260-03c64893a6dd\" (UID: \"c0ca09da-0456-4ab9-8260-03c64893a6dd\") " Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.465747 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0ca09da-0456-4ab9-8260-03c64893a6dd-utilities" (OuterVolumeSpecName: "utilities") pod "c0ca09da-0456-4ab9-8260-03c64893a6dd" (UID: "c0ca09da-0456-4ab9-8260-03c64893a6dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.467937 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0ca09da-0456-4ab9-8260-03c64893a6dd-kube-api-access-t9vmz" (OuterVolumeSpecName: "kube-api-access-t9vmz") pod "c0ca09da-0456-4ab9-8260-03c64893a6dd" (UID: "c0ca09da-0456-4ab9-8260-03c64893a6dd"). InnerVolumeSpecName "kube-api-access-t9vmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.487949 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6zt6h" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.499824 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2pls6" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.503952 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sqj49" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.530102 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.540057 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0ca09da-0456-4ab9-8260-03c64893a6dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c0ca09da-0456-4ab9-8260-03c64893a6dd" (UID: "c0ca09da-0456-4ab9-8260-03c64893a6dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.563092 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0ca09da-0456-4ab9-8260-03c64893a6dd-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.563133 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9vmz\" (UniqueName: \"kubernetes.io/projected/c0ca09da-0456-4ab9-8260-03c64893a6dd-kube-api-access-t9vmz\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.563147 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0ca09da-0456-4ab9-8260-03c64893a6dd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.664420 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17fe9bd5-5973-47bd-945f-55bb8caa9d65-marketplace-trusted-ca\") pod \"17fe9bd5-5973-47bd-945f-55bb8caa9d65\" (UID: \"17fe9bd5-5973-47bd-945f-55bb8caa9d65\") " Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.664476 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ec5dfd3-3221-4141-a2d4-aea02733a985-catalog-content\") pod \"3ec5dfd3-3221-4141-a2d4-aea02733a985\" (UID: \"3ec5dfd3-3221-4141-a2d4-aea02733a985\") " Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.664511 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19ea766c-8ab7-4499-a51b-cd6968a62a76-catalog-content\") pod \"19ea766c-8ab7-4499-a51b-cd6968a62a76\" (UID: \"19ea766c-8ab7-4499-a51b-cd6968a62a76\") " Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.664541 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdh4n\" (UniqueName: \"kubernetes.io/projected/f67be9f3-d0bd-439c-ab66-72a6a29b3104-kube-api-access-qdh4n\") pod \"f67be9f3-d0bd-439c-ab66-72a6a29b3104\" (UID: \"f67be9f3-d0bd-439c-ab66-72a6a29b3104\") " Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.664571 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2s94\" (UniqueName: \"kubernetes.io/projected/3ec5dfd3-3221-4141-a2d4-aea02733a985-kube-api-access-r2s94\") pod \"3ec5dfd3-3221-4141-a2d4-aea02733a985\" (UID: \"3ec5dfd3-3221-4141-a2d4-aea02733a985\") " Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.664602 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f67be9f3-d0bd-439c-ab66-72a6a29b3104-catalog-content\") pod \"f67be9f3-d0bd-439c-ab66-72a6a29b3104\" (UID: \"f67be9f3-d0bd-439c-ab66-72a6a29b3104\") " Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.664643 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/17fe9bd5-5973-47bd-945f-55bb8caa9d65-marketplace-operator-metrics\") pod \"17fe9bd5-5973-47bd-945f-55bb8caa9d65\" (UID: \"17fe9bd5-5973-47bd-945f-55bb8caa9d65\") " Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.664675 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ec5dfd3-3221-4141-a2d4-aea02733a985-utilities\") pod \"3ec5dfd3-3221-4141-a2d4-aea02733a985\" (UID: \"3ec5dfd3-3221-4141-a2d4-aea02733a985\") " Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.664688 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19ea766c-8ab7-4499-a51b-cd6968a62a76-utilities\") pod \"19ea766c-8ab7-4499-a51b-cd6968a62a76\" (UID: \"19ea766c-8ab7-4499-a51b-cd6968a62a76\") " Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.664708 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmrmp\" (UniqueName: \"kubernetes.io/projected/17fe9bd5-5973-47bd-945f-55bb8caa9d65-kube-api-access-jmrmp\") pod \"17fe9bd5-5973-47bd-945f-55bb8caa9d65\" (UID: \"17fe9bd5-5973-47bd-945f-55bb8caa9d65\") " Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.664744 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f67be9f3-d0bd-439c-ab66-72a6a29b3104-utilities\") pod \"f67be9f3-d0bd-439c-ab66-72a6a29b3104\" (UID: \"f67be9f3-d0bd-439c-ab66-72a6a29b3104\") " Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.664791 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmfbl\" (UniqueName: \"kubernetes.io/projected/19ea766c-8ab7-4499-a51b-cd6968a62a76-kube-api-access-qmfbl\") pod \"19ea766c-8ab7-4499-a51b-cd6968a62a76\" (UID: \"19ea766c-8ab7-4499-a51b-cd6968a62a76\") " Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.665176 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17fe9bd5-5973-47bd-945f-55bb8caa9d65-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "17fe9bd5-5973-47bd-945f-55bb8caa9d65" (UID: "17fe9bd5-5973-47bd-945f-55bb8caa9d65"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.665786 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19ea766c-8ab7-4499-a51b-cd6968a62a76-utilities" (OuterVolumeSpecName: "utilities") pod "19ea766c-8ab7-4499-a51b-cd6968a62a76" (UID: "19ea766c-8ab7-4499-a51b-cd6968a62a76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.666009 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f67be9f3-d0bd-439c-ab66-72a6a29b3104-utilities" (OuterVolumeSpecName: "utilities") pod "f67be9f3-d0bd-439c-ab66-72a6a29b3104" (UID: "f67be9f3-d0bd-439c-ab66-72a6a29b3104"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.667489 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ec5dfd3-3221-4141-a2d4-aea02733a985-utilities" (OuterVolumeSpecName: "utilities") pod "3ec5dfd3-3221-4141-a2d4-aea02733a985" (UID: "3ec5dfd3-3221-4141-a2d4-aea02733a985"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.671843 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f67be9f3-d0bd-439c-ab66-72a6a29b3104-kube-api-access-qdh4n" (OuterVolumeSpecName: "kube-api-access-qdh4n") pod "f67be9f3-d0bd-439c-ab66-72a6a29b3104" (UID: "f67be9f3-d0bd-439c-ab66-72a6a29b3104"). InnerVolumeSpecName "kube-api-access-qdh4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.672171 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17fe9bd5-5973-47bd-945f-55bb8caa9d65-kube-api-access-jmrmp" (OuterVolumeSpecName: "kube-api-access-jmrmp") pod "17fe9bd5-5973-47bd-945f-55bb8caa9d65" (UID: "17fe9bd5-5973-47bd-945f-55bb8caa9d65"). InnerVolumeSpecName "kube-api-access-jmrmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.673837 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17fe9bd5-5973-47bd-945f-55bb8caa9d65-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "17fe9bd5-5973-47bd-945f-55bb8caa9d65" (UID: "17fe9bd5-5973-47bd-945f-55bb8caa9d65"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.683806 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ec5dfd3-3221-4141-a2d4-aea02733a985-kube-api-access-r2s94" (OuterVolumeSpecName: "kube-api-access-r2s94") pod "3ec5dfd3-3221-4141-a2d4-aea02733a985" (UID: "3ec5dfd3-3221-4141-a2d4-aea02733a985"). InnerVolumeSpecName "kube-api-access-r2s94". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.685576 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19ea766c-8ab7-4499-a51b-cd6968a62a76-kube-api-access-qmfbl" (OuterVolumeSpecName: "kube-api-access-qmfbl") pod "19ea766c-8ab7-4499-a51b-cd6968a62a76" (UID: "19ea766c-8ab7-4499-a51b-cd6968a62a76"). InnerVolumeSpecName "kube-api-access-qmfbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.691282 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k4n89"] Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.692892 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19ea766c-8ab7-4499-a51b-cd6968a62a76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19ea766c-8ab7-4499-a51b-cd6968a62a76" (UID: "19ea766c-8ab7-4499-a51b-cd6968a62a76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.737467 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ec5dfd3-3221-4141-a2d4-aea02733a985-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3ec5dfd3-3221-4141-a2d4-aea02733a985" (UID: "3ec5dfd3-3221-4141-a2d4-aea02733a985"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.765928 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmfbl\" (UniqueName: \"kubernetes.io/projected/19ea766c-8ab7-4499-a51b-cd6968a62a76-kube-api-access-qmfbl\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.765959 4997 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17fe9bd5-5973-47bd-945f-55bb8caa9d65-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.765969 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ec5dfd3-3221-4141-a2d4-aea02733a985-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.765978 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19ea766c-8ab7-4499-a51b-cd6968a62a76-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.765986 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdh4n\" (UniqueName: \"kubernetes.io/projected/f67be9f3-d0bd-439c-ab66-72a6a29b3104-kube-api-access-qdh4n\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.765998 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2s94\" (UniqueName: \"kubernetes.io/projected/3ec5dfd3-3221-4141-a2d4-aea02733a985-kube-api-access-r2s94\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.766008 4997 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/17fe9bd5-5973-47bd-945f-55bb8caa9d65-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.766017 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ec5dfd3-3221-4141-a2d4-aea02733a985-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.766024 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19ea766c-8ab7-4499-a51b-cd6968a62a76-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.766032 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmrmp\" (UniqueName: \"kubernetes.io/projected/17fe9bd5-5973-47bd-945f-55bb8caa9d65-kube-api-access-jmrmp\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.766040 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f67be9f3-d0bd-439c-ab66-72a6a29b3104-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.810899 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f67be9f3-d0bd-439c-ab66-72a6a29b3104-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f67be9f3-d0bd-439c-ab66-72a6a29b3104" (UID: "f67be9f3-d0bd-439c-ab66-72a6a29b3104"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.867028 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f67be9f3-d0bd-439c-ab66-72a6a29b3104-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.984828 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k4n89" event={"ID":"14ea843e-5c09-4461-88b4-4810e6d28cca","Type":"ContainerStarted","Data":"a95123c575c9eff70646c83f17b8c7f094863b351913141df2109cc6181d3755"} Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.985372 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k4n89" event={"ID":"14ea843e-5c09-4461-88b4-4810e6d28cca","Type":"ContainerStarted","Data":"06638bf766ef275d8b2d41f40e81d4cb27c0e271e9066e50f5bad07c0e056812"} Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.985914 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-k4n89" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.988057 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zt6h" event={"ID":"19ea766c-8ab7-4499-a51b-cd6968a62a76","Type":"ContainerDied","Data":"d9e165880b88eb79dad36357ea95ee4cfbd2ec5e4d25341631943616d98e3e28"} Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.988126 4997 scope.go:117] "RemoveContainer" containerID="485e08d238a48edaa6a607645ffb2b538173a81e733c867397c153bde926b98c" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.988233 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6zt6h" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.988903 4997 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-k4n89 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.65:8080/healthz\": dial tcp 10.217.0.65:8080: connect: connection refused" start-of-body= Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.988936 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-k4n89" podUID="14ea843e-5c09-4461-88b4-4810e6d28cca" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.65:8080/healthz\": dial tcp 10.217.0.65:8080: connect: connection refused" Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.992600 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqj49" event={"ID":"f67be9f3-d0bd-439c-ab66-72a6a29b3104","Type":"ContainerDied","Data":"dcc1955a3a7954ed1e2a9c92876736cd8c77e3ece2c376570f0b0cca1bc294b7"} Dec 05 07:01:45 crc kubenswrapper[4997]: I1205 07:01:45.992679 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sqj49" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.003673 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6tmw" event={"ID":"c0ca09da-0456-4ab9-8260-03c64893a6dd","Type":"ContainerDied","Data":"fe654c25c33b74a9a9df34f0abf0ce44035a96ac33fb0b5b2ec6def7b4510fac"} Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.003097 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6tmw" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.004831 4997 scope.go:117] "RemoveContainer" containerID="072d16b6f5b85a1dc575341b25ee9b439d5cbd8391d9c089da4d9be91d991414" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.009033 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-k4n89" podStartSLOduration=2.009008775 podStartE2EDuration="2.009008775s" podCreationTimestamp="2025-12-05 07:01:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:01:45.999866542 +0000 UTC m=+406.528773813" watchObservedRunningTime="2025-12-05 07:01:46.009008775 +0000 UTC m=+406.537916056" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.018702 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2pls6" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.018799 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2pls6" event={"ID":"3ec5dfd3-3221-4141-a2d4-aea02733a985","Type":"ContainerDied","Data":"b857b36770f2e4fa834bc88a9e2f2cff0bc6813ceaba068b977d90eb0e5d6c7e"} Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.021883 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6zt6h"] Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.024009 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" event={"ID":"17fe9bd5-5973-47bd-945f-55bb8caa9d65","Type":"ContainerDied","Data":"fd4cffdda303ba5d75ec82b29977cb9148b2ea1268e3de5b24b55dc798848fae"} Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.024091 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k9twj" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.035659 4997 scope.go:117] "RemoveContainer" containerID="8aa4e42c3fc0a15620790818f529b5aeb899bb0c726805e692b09bb9f940209a" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.035799 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6zt6h"] Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.042492 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sqj49"] Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.045699 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sqj49"] Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.050484 4997 scope.go:117] "RemoveContainer" containerID="bb9a8fd6335f6cdc2749befa3b807f9f84860609bd15b8e49189cf40c00dc69f" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.055276 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n6tmw"] Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.063349 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n6tmw"] Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.068665 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k9twj"] Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.075096 4997 scope.go:117] "RemoveContainer" containerID="d145b208473d5074697d7999b4e8745198267a2d6c50f5818a14d646533803b8" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.076365 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k9twj"] Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.081891 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2pls6"] Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.085061 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2pls6"] Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.100849 4997 scope.go:117] "RemoveContainer" containerID="e0d79dda4df3f7129af8e06d8e2d17503ae56ed340b4833f0e360903e10a69bc" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.114584 4997 scope.go:117] "RemoveContainer" containerID="ebb32219ca7297fc6f6e84a88a97517ec975997f7f3538a900818823da9b3a34" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.131833 4997 scope.go:117] "RemoveContainer" containerID="a9229007162d88dc7222c56646b799e696134f1ac031cc24746970e7065276c1" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.148769 4997 scope.go:117] "RemoveContainer" containerID="88139d49c80dee570c01f582f9d8414dd3cdde56cd4dbdbc2e2eab5539c440ee" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.167105 4997 scope.go:117] "RemoveContainer" containerID="4e95c9a0122f9f97de73329b76fe8d535a38995655562840fb0aafe6fd7824a1" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.193890 4997 scope.go:117] "RemoveContainer" containerID="e7f84a21bad53bb6f660d0d6f35db8bf64bdbe637269e05bf17bfac3a68c2ab5" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.208486 4997 scope.go:117] "RemoveContainer" containerID="08691ef05fca2b5aa6f6734da33f74f246a7c44cc89dd5b7afa16ec3c67cb724" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.223326 4997 scope.go:117] "RemoveContainer" containerID="99b851f62ee198eeaadf857187e158d3889e72c84fe6d9c43f641d75b03a05df" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.887548 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qxz27"] Dec 05 07:01:46 crc kubenswrapper[4997]: E1205 07:01:46.888004 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ec5dfd3-3221-4141-a2d4-aea02733a985" containerName="registry-server" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888021 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ec5dfd3-3221-4141-a2d4-aea02733a985" containerName="registry-server" Dec 05 07:01:46 crc kubenswrapper[4997]: E1205 07:01:46.888042 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0ca09da-0456-4ab9-8260-03c64893a6dd" containerName="extract-utilities" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888049 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0ca09da-0456-4ab9-8260-03c64893a6dd" containerName="extract-utilities" Dec 05 07:01:46 crc kubenswrapper[4997]: E1205 07:01:46.888077 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ea766c-8ab7-4499-a51b-cd6968a62a76" containerName="extract-utilities" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888083 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ea766c-8ab7-4499-a51b-cd6968a62a76" containerName="extract-utilities" Dec 05 07:01:46 crc kubenswrapper[4997]: E1205 07:01:46.888097 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ea766c-8ab7-4499-a51b-cd6968a62a76" containerName="extract-content" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888106 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ea766c-8ab7-4499-a51b-cd6968a62a76" containerName="extract-content" Dec 05 07:01:46 crc kubenswrapper[4997]: E1205 07:01:46.888121 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17fe9bd5-5973-47bd-945f-55bb8caa9d65" containerName="marketplace-operator" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888129 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="17fe9bd5-5973-47bd-945f-55bb8caa9d65" containerName="marketplace-operator" Dec 05 07:01:46 crc kubenswrapper[4997]: E1205 07:01:46.888142 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0ca09da-0456-4ab9-8260-03c64893a6dd" containerName="registry-server" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888149 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0ca09da-0456-4ab9-8260-03c64893a6dd" containerName="registry-server" Dec 05 07:01:46 crc kubenswrapper[4997]: E1205 07:01:46.888165 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17fe9bd5-5973-47bd-945f-55bb8caa9d65" containerName="marketplace-operator" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888172 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="17fe9bd5-5973-47bd-945f-55bb8caa9d65" containerName="marketplace-operator" Dec 05 07:01:46 crc kubenswrapper[4997]: E1205 07:01:46.888216 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ec5dfd3-3221-4141-a2d4-aea02733a985" containerName="extract-utilities" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888224 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ec5dfd3-3221-4141-a2d4-aea02733a985" containerName="extract-utilities" Dec 05 07:01:46 crc kubenswrapper[4997]: E1205 07:01:46.888239 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0ca09da-0456-4ab9-8260-03c64893a6dd" containerName="extract-content" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888251 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0ca09da-0456-4ab9-8260-03c64893a6dd" containerName="extract-content" Dec 05 07:01:46 crc kubenswrapper[4997]: E1205 07:01:46.888273 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f67be9f3-d0bd-439c-ab66-72a6a29b3104" containerName="registry-server" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888282 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f67be9f3-d0bd-439c-ab66-72a6a29b3104" containerName="registry-server" Dec 05 07:01:46 crc kubenswrapper[4997]: E1205 07:01:46.888301 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ea766c-8ab7-4499-a51b-cd6968a62a76" containerName="registry-server" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888309 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ea766c-8ab7-4499-a51b-cd6968a62a76" containerName="registry-server" Dec 05 07:01:46 crc kubenswrapper[4997]: E1205 07:01:46.888320 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f67be9f3-d0bd-439c-ab66-72a6a29b3104" containerName="extract-utilities" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888335 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f67be9f3-d0bd-439c-ab66-72a6a29b3104" containerName="extract-utilities" Dec 05 07:01:46 crc kubenswrapper[4997]: E1205 07:01:46.888355 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ec5dfd3-3221-4141-a2d4-aea02733a985" containerName="extract-content" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888364 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ec5dfd3-3221-4141-a2d4-aea02733a985" containerName="extract-content" Dec 05 07:01:46 crc kubenswrapper[4997]: E1205 07:01:46.888379 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f67be9f3-d0bd-439c-ab66-72a6a29b3104" containerName="extract-content" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888387 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f67be9f3-d0bd-439c-ab66-72a6a29b3104" containerName="extract-content" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888651 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ec5dfd3-3221-4141-a2d4-aea02733a985" containerName="registry-server" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888671 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="17fe9bd5-5973-47bd-945f-55bb8caa9d65" containerName="marketplace-operator" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888679 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f67be9f3-d0bd-439c-ab66-72a6a29b3104" containerName="registry-server" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888701 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0ca09da-0456-4ab9-8260-03c64893a6dd" containerName="registry-server" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.888712 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="19ea766c-8ab7-4499-a51b-cd6968a62a76" containerName="registry-server" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.889201 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="17fe9bd5-5973-47bd-945f-55bb8caa9d65" containerName="marketplace-operator" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.891788 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qxz27" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.896485 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.900734 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qxz27"] Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.998382 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/016b7cd0-efde-4d2c-940c-181108006a33-utilities\") pod \"certified-operators-qxz27\" (UID: \"016b7cd0-efde-4d2c-940c-181108006a33\") " pod="openshift-marketplace/certified-operators-qxz27" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.998440 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/016b7cd0-efde-4d2c-940c-181108006a33-catalog-content\") pod \"certified-operators-qxz27\" (UID: \"016b7cd0-efde-4d2c-940c-181108006a33\") " pod="openshift-marketplace/certified-operators-qxz27" Dec 05 07:01:46 crc kubenswrapper[4997]: I1205 07:01:46.998469 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85m26\" (UniqueName: \"kubernetes.io/projected/016b7cd0-efde-4d2c-940c-181108006a33-kube-api-access-85m26\") pod \"certified-operators-qxz27\" (UID: \"016b7cd0-efde-4d2c-940c-181108006a33\") " pod="openshift-marketplace/certified-operators-qxz27" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.035205 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-k4n89" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.077045 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zdk9s"] Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.086237 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zdk9s" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.093949 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.098091 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zdk9s"] Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.099207 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/016b7cd0-efde-4d2c-940c-181108006a33-utilities\") pod \"certified-operators-qxz27\" (UID: \"016b7cd0-efde-4d2c-940c-181108006a33\") " pod="openshift-marketplace/certified-operators-qxz27" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.099265 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/016b7cd0-efde-4d2c-940c-181108006a33-catalog-content\") pod \"certified-operators-qxz27\" (UID: \"016b7cd0-efde-4d2c-940c-181108006a33\") " pod="openshift-marketplace/certified-operators-qxz27" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.099296 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3274e3c-2ad5-4756-8d3f-bae751f1e719-utilities\") pod \"community-operators-zdk9s\" (UID: \"d3274e3c-2ad5-4756-8d3f-bae751f1e719\") " pod="openshift-marketplace/community-operators-zdk9s" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.099320 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rt55\" (UniqueName: \"kubernetes.io/projected/d3274e3c-2ad5-4756-8d3f-bae751f1e719-kube-api-access-9rt55\") pod \"community-operators-zdk9s\" (UID: \"d3274e3c-2ad5-4756-8d3f-bae751f1e719\") " pod="openshift-marketplace/community-operators-zdk9s" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.099343 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3274e3c-2ad5-4756-8d3f-bae751f1e719-catalog-content\") pod \"community-operators-zdk9s\" (UID: \"d3274e3c-2ad5-4756-8d3f-bae751f1e719\") " pod="openshift-marketplace/community-operators-zdk9s" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.099374 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85m26\" (UniqueName: \"kubernetes.io/projected/016b7cd0-efde-4d2c-940c-181108006a33-kube-api-access-85m26\") pod \"certified-operators-qxz27\" (UID: \"016b7cd0-efde-4d2c-940c-181108006a33\") " pod="openshift-marketplace/certified-operators-qxz27" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.100285 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/016b7cd0-efde-4d2c-940c-181108006a33-catalog-content\") pod \"certified-operators-qxz27\" (UID: \"016b7cd0-efde-4d2c-940c-181108006a33\") " pod="openshift-marketplace/certified-operators-qxz27" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.100536 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/016b7cd0-efde-4d2c-940c-181108006a33-utilities\") pod \"certified-operators-qxz27\" (UID: \"016b7cd0-efde-4d2c-940c-181108006a33\") " pod="openshift-marketplace/certified-operators-qxz27" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.129517 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85m26\" (UniqueName: \"kubernetes.io/projected/016b7cd0-efde-4d2c-940c-181108006a33-kube-api-access-85m26\") pod \"certified-operators-qxz27\" (UID: \"016b7cd0-efde-4d2c-940c-181108006a33\") " pod="openshift-marketplace/certified-operators-qxz27" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.200686 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3274e3c-2ad5-4756-8d3f-bae751f1e719-utilities\") pod \"community-operators-zdk9s\" (UID: \"d3274e3c-2ad5-4756-8d3f-bae751f1e719\") " pod="openshift-marketplace/community-operators-zdk9s" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.200730 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rt55\" (UniqueName: \"kubernetes.io/projected/d3274e3c-2ad5-4756-8d3f-bae751f1e719-kube-api-access-9rt55\") pod \"community-operators-zdk9s\" (UID: \"d3274e3c-2ad5-4756-8d3f-bae751f1e719\") " pod="openshift-marketplace/community-operators-zdk9s" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.200752 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3274e3c-2ad5-4756-8d3f-bae751f1e719-catalog-content\") pod \"community-operators-zdk9s\" (UID: \"d3274e3c-2ad5-4756-8d3f-bae751f1e719\") " pod="openshift-marketplace/community-operators-zdk9s" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.201271 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3274e3c-2ad5-4756-8d3f-bae751f1e719-catalog-content\") pod \"community-operators-zdk9s\" (UID: \"d3274e3c-2ad5-4756-8d3f-bae751f1e719\") " pod="openshift-marketplace/community-operators-zdk9s" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.201466 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3274e3c-2ad5-4756-8d3f-bae751f1e719-utilities\") pod \"community-operators-zdk9s\" (UID: \"d3274e3c-2ad5-4756-8d3f-bae751f1e719\") " pod="openshift-marketplace/community-operators-zdk9s" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.215915 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qxz27" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.219974 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rt55\" (UniqueName: \"kubernetes.io/projected/d3274e3c-2ad5-4756-8d3f-bae751f1e719-kube-api-access-9rt55\") pod \"community-operators-zdk9s\" (UID: \"d3274e3c-2ad5-4756-8d3f-bae751f1e719\") " pod="openshift-marketplace/community-operators-zdk9s" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.413472 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zdk9s" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.630006 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qxz27"] Dec 05 07:01:47 crc kubenswrapper[4997]: W1205 07:01:47.639628 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod016b7cd0_efde_4d2c_940c_181108006a33.slice/crio-6b5f884e76f29d3b65abca91b8b5e5ffe97e044651cacf681877d8d4e514ba6c WatchSource:0}: Error finding container 6b5f884e76f29d3b65abca91b8b5e5ffe97e044651cacf681877d8d4e514ba6c: Status 404 returned error can't find the container with id 6b5f884e76f29d3b65abca91b8b5e5ffe97e044651cacf681877d8d4e514ba6c Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.758261 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17fe9bd5-5973-47bd-945f-55bb8caa9d65" path="/var/lib/kubelet/pods/17fe9bd5-5973-47bd-945f-55bb8caa9d65/volumes" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.758878 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19ea766c-8ab7-4499-a51b-cd6968a62a76" path="/var/lib/kubelet/pods/19ea766c-8ab7-4499-a51b-cd6968a62a76/volumes" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.759432 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ec5dfd3-3221-4141-a2d4-aea02733a985" path="/var/lib/kubelet/pods/3ec5dfd3-3221-4141-a2d4-aea02733a985/volumes" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.760575 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0ca09da-0456-4ab9-8260-03c64893a6dd" path="/var/lib/kubelet/pods/c0ca09da-0456-4ab9-8260-03c64893a6dd/volumes" Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.761155 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f67be9f3-d0bd-439c-ab66-72a6a29b3104" path="/var/lib/kubelet/pods/f67be9f3-d0bd-439c-ab66-72a6a29b3104/volumes" Dec 05 07:01:47 crc kubenswrapper[4997]: W1205 07:01:47.895717 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3274e3c_2ad5_4756_8d3f_bae751f1e719.slice/crio-af636190be01e54fa8360c220dce646c7204ad819c54aaba92271c59e2e10e67 WatchSource:0}: Error finding container af636190be01e54fa8360c220dce646c7204ad819c54aaba92271c59e2e10e67: Status 404 returned error can't find the container with id af636190be01e54fa8360c220dce646c7204ad819c54aaba92271c59e2e10e67 Dec 05 07:01:47 crc kubenswrapper[4997]: I1205 07:01:47.903804 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zdk9s"] Dec 05 07:01:48 crc kubenswrapper[4997]: I1205 07:01:48.043248 4997 generic.go:334] "Generic (PLEG): container finished" podID="016b7cd0-efde-4d2c-940c-181108006a33" containerID="a9c68c2270409df9305fea61bfb523736272b3d4a5aff279aefe99e07a9aaf93" exitCode=0 Dec 05 07:01:48 crc kubenswrapper[4997]: I1205 07:01:48.043387 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxz27" event={"ID":"016b7cd0-efde-4d2c-940c-181108006a33","Type":"ContainerDied","Data":"a9c68c2270409df9305fea61bfb523736272b3d4a5aff279aefe99e07a9aaf93"} Dec 05 07:01:48 crc kubenswrapper[4997]: I1205 07:01:48.043483 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxz27" event={"ID":"016b7cd0-efde-4d2c-940c-181108006a33","Type":"ContainerStarted","Data":"6b5f884e76f29d3b65abca91b8b5e5ffe97e044651cacf681877d8d4e514ba6c"} Dec 05 07:01:48 crc kubenswrapper[4997]: I1205 07:01:48.053067 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zdk9s" event={"ID":"d3274e3c-2ad5-4756-8d3f-bae751f1e719","Type":"ContainerStarted","Data":"bc9de9eb01b3ac9c658ad129e5487b12cb3a6a4f387cee91a441aba69d50a5d9"} Dec 05 07:01:48 crc kubenswrapper[4997]: I1205 07:01:48.053129 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zdk9s" event={"ID":"d3274e3c-2ad5-4756-8d3f-bae751f1e719","Type":"ContainerStarted","Data":"af636190be01e54fa8360c220dce646c7204ad819c54aaba92271c59e2e10e67"} Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.063285 4997 generic.go:334] "Generic (PLEG): container finished" podID="d3274e3c-2ad5-4756-8d3f-bae751f1e719" containerID="bc9de9eb01b3ac9c658ad129e5487b12cb3a6a4f387cee91a441aba69d50a5d9" exitCode=0 Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.063391 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zdk9s" event={"ID":"d3274e3c-2ad5-4756-8d3f-bae751f1e719","Type":"ContainerDied","Data":"bc9de9eb01b3ac9c658ad129e5487b12cb3a6a4f387cee91a441aba69d50a5d9"} Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.068504 4997 generic.go:334] "Generic (PLEG): container finished" podID="016b7cd0-efde-4d2c-940c-181108006a33" containerID="75938ac1bf709a5faa682693493ce26c1217ed8e5b6fef11bb387a7a4b04ba74" exitCode=0 Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.068537 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxz27" event={"ID":"016b7cd0-efde-4d2c-940c-181108006a33","Type":"ContainerDied","Data":"75938ac1bf709a5faa682693493ce26c1217ed8e5b6fef11bb387a7a4b04ba74"} Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.270428 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c46hv"] Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.271467 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c46hv" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.275058 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.284797 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c46hv"] Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.333015 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b51e69f-d870-494a-adc2-b912ec12e2e6-catalog-content\") pod \"redhat-marketplace-c46hv\" (UID: \"0b51e69f-d870-494a-adc2-b912ec12e2e6\") " pod="openshift-marketplace/redhat-marketplace-c46hv" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.333086 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggcp7\" (UniqueName: \"kubernetes.io/projected/0b51e69f-d870-494a-adc2-b912ec12e2e6-kube-api-access-ggcp7\") pod \"redhat-marketplace-c46hv\" (UID: \"0b51e69f-d870-494a-adc2-b912ec12e2e6\") " pod="openshift-marketplace/redhat-marketplace-c46hv" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.333109 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b51e69f-d870-494a-adc2-b912ec12e2e6-utilities\") pod \"redhat-marketplace-c46hv\" (UID: \"0b51e69f-d870-494a-adc2-b912ec12e2e6\") " pod="openshift-marketplace/redhat-marketplace-c46hv" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.435176 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b51e69f-d870-494a-adc2-b912ec12e2e6-catalog-content\") pod \"redhat-marketplace-c46hv\" (UID: \"0b51e69f-d870-494a-adc2-b912ec12e2e6\") " pod="openshift-marketplace/redhat-marketplace-c46hv" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.435256 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggcp7\" (UniqueName: \"kubernetes.io/projected/0b51e69f-d870-494a-adc2-b912ec12e2e6-kube-api-access-ggcp7\") pod \"redhat-marketplace-c46hv\" (UID: \"0b51e69f-d870-494a-adc2-b912ec12e2e6\") " pod="openshift-marketplace/redhat-marketplace-c46hv" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.435295 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b51e69f-d870-494a-adc2-b912ec12e2e6-utilities\") pod \"redhat-marketplace-c46hv\" (UID: \"0b51e69f-d870-494a-adc2-b912ec12e2e6\") " pod="openshift-marketplace/redhat-marketplace-c46hv" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.435819 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b51e69f-d870-494a-adc2-b912ec12e2e6-catalog-content\") pod \"redhat-marketplace-c46hv\" (UID: \"0b51e69f-d870-494a-adc2-b912ec12e2e6\") " pod="openshift-marketplace/redhat-marketplace-c46hv" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.437960 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b51e69f-d870-494a-adc2-b912ec12e2e6-utilities\") pod \"redhat-marketplace-c46hv\" (UID: \"0b51e69f-d870-494a-adc2-b912ec12e2e6\") " pod="openshift-marketplace/redhat-marketplace-c46hv" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.459965 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggcp7\" (UniqueName: \"kubernetes.io/projected/0b51e69f-d870-494a-adc2-b912ec12e2e6-kube-api-access-ggcp7\") pod \"redhat-marketplace-c46hv\" (UID: \"0b51e69f-d870-494a-adc2-b912ec12e2e6\") " pod="openshift-marketplace/redhat-marketplace-c46hv" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.477959 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ktvff"] Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.479488 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ktvff" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.481775 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.492470 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ktvff"] Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.536340 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76xbn\" (UniqueName: \"kubernetes.io/projected/e5f08766-f19b-4b4f-b1f8-f25b4214457e-kube-api-access-76xbn\") pod \"redhat-operators-ktvff\" (UID: \"e5f08766-f19b-4b4f-b1f8-f25b4214457e\") " pod="openshift-marketplace/redhat-operators-ktvff" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.536406 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5f08766-f19b-4b4f-b1f8-f25b4214457e-catalog-content\") pod \"redhat-operators-ktvff\" (UID: \"e5f08766-f19b-4b4f-b1f8-f25b4214457e\") " pod="openshift-marketplace/redhat-operators-ktvff" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.536527 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5f08766-f19b-4b4f-b1f8-f25b4214457e-utilities\") pod \"redhat-operators-ktvff\" (UID: \"e5f08766-f19b-4b4f-b1f8-f25b4214457e\") " pod="openshift-marketplace/redhat-operators-ktvff" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.602741 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c46hv" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.638066 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5f08766-f19b-4b4f-b1f8-f25b4214457e-catalog-content\") pod \"redhat-operators-ktvff\" (UID: \"e5f08766-f19b-4b4f-b1f8-f25b4214457e\") " pod="openshift-marketplace/redhat-operators-ktvff" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.638198 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5f08766-f19b-4b4f-b1f8-f25b4214457e-utilities\") pod \"redhat-operators-ktvff\" (UID: \"e5f08766-f19b-4b4f-b1f8-f25b4214457e\") " pod="openshift-marketplace/redhat-operators-ktvff" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.638229 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76xbn\" (UniqueName: \"kubernetes.io/projected/e5f08766-f19b-4b4f-b1f8-f25b4214457e-kube-api-access-76xbn\") pod \"redhat-operators-ktvff\" (UID: \"e5f08766-f19b-4b4f-b1f8-f25b4214457e\") " pod="openshift-marketplace/redhat-operators-ktvff" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.639256 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5f08766-f19b-4b4f-b1f8-f25b4214457e-catalog-content\") pod \"redhat-operators-ktvff\" (UID: \"e5f08766-f19b-4b4f-b1f8-f25b4214457e\") " pod="openshift-marketplace/redhat-operators-ktvff" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.639710 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5f08766-f19b-4b4f-b1f8-f25b4214457e-utilities\") pod \"redhat-operators-ktvff\" (UID: \"e5f08766-f19b-4b4f-b1f8-f25b4214457e\") " pod="openshift-marketplace/redhat-operators-ktvff" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.667110 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76xbn\" (UniqueName: \"kubernetes.io/projected/e5f08766-f19b-4b4f-b1f8-f25b4214457e-kube-api-access-76xbn\") pod \"redhat-operators-ktvff\" (UID: \"e5f08766-f19b-4b4f-b1f8-f25b4214457e\") " pod="openshift-marketplace/redhat-operators-ktvff" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.770493 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.770554 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.770605 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.771225 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2f6d912e8d30078a8f27c58e585fd50e33dbd4541657a6aef0990fa4ea86b3b6"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.771292 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://2f6d912e8d30078a8f27c58e585fd50e33dbd4541657a6aef0990fa4ea86b3b6" gracePeriod=600 Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.789938 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" podUID="cd09450e-798d-41d5-b925-90ab15df8dac" containerName="registry" containerID="cri-o://d489a9092e2cdddebf8931a4ffd9ead2066e0f6d2c9185c1800bd007548a2afb" gracePeriod=30 Dec 05 07:01:49 crc kubenswrapper[4997]: I1205 07:01:49.805763 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ktvff" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.082317 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c46hv"] Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.104600 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zdk9s" event={"ID":"d3274e3c-2ad5-4756-8d3f-bae751f1e719","Type":"ContainerStarted","Data":"f33f8296c34e4683e59da1ac65dc94102f8aa3f8cb8c40b4784a9cc92ecd4771"} Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.110073 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="2f6d912e8d30078a8f27c58e585fd50e33dbd4541657a6aef0990fa4ea86b3b6" exitCode=0 Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.110132 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"2f6d912e8d30078a8f27c58e585fd50e33dbd4541657a6aef0990fa4ea86b3b6"} Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.110166 4997 scope.go:117] "RemoveContainer" containerID="6a33c5992c398c663ac144809110cfc330622bd58a90e131324c19120492ba9a" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.115683 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxz27" event={"ID":"016b7cd0-efde-4d2c-940c-181108006a33","Type":"ContainerStarted","Data":"487fd2b294c68db72810d2d7c43a20ef768624e86a9529c833861fdd3b696b6c"} Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.118653 4997 generic.go:334] "Generic (PLEG): container finished" podID="cd09450e-798d-41d5-b925-90ab15df8dac" containerID="d489a9092e2cdddebf8931a4ffd9ead2066e0f6d2c9185c1800bd007548a2afb" exitCode=0 Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.118710 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" event={"ID":"cd09450e-798d-41d5-b925-90ab15df8dac","Type":"ContainerDied","Data":"d489a9092e2cdddebf8931a4ffd9ead2066e0f6d2c9185c1800bd007548a2afb"} Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.152487 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qxz27" podStartSLOduration=2.641140004 podStartE2EDuration="4.152457439s" podCreationTimestamp="2025-12-05 07:01:46 +0000 UTC" firstStartedPulling="2025-12-05 07:01:48.047222478 +0000 UTC m=+408.576129779" lastFinishedPulling="2025-12-05 07:01:49.558539953 +0000 UTC m=+410.087447214" observedRunningTime="2025-12-05 07:01:50.14258924 +0000 UTC m=+410.671496511" watchObservedRunningTime="2025-12-05 07:01:50.152457439 +0000 UTC m=+410.681364700" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.264174 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ktvff"] Dec 05 07:01:50 crc kubenswrapper[4997]: W1205 07:01:50.300439 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5f08766_f19b_4b4f_b1f8_f25b4214457e.slice/crio-e940064204a5dd99cbc9ef2fd378638a574a1dd8db41ef3ad07fb3da67317dcd WatchSource:0}: Error finding container e940064204a5dd99cbc9ef2fd378638a574a1dd8db41ef3ad07fb3da67317dcd: Status 404 returned error can't find the container with id e940064204a5dd99cbc9ef2fd378638a574a1dd8db41ef3ad07fb3da67317dcd Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.332232 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.353997 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cd09450e-798d-41d5-b925-90ab15df8dac-registry-certificates\") pod \"cd09450e-798d-41d5-b925-90ab15df8dac\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.354164 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"cd09450e-798d-41d5-b925-90ab15df8dac\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.354256 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cd09450e-798d-41d5-b925-90ab15df8dac-installation-pull-secrets\") pod \"cd09450e-798d-41d5-b925-90ab15df8dac\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.354320 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cd09450e-798d-41d5-b925-90ab15df8dac-ca-trust-extracted\") pod \"cd09450e-798d-41d5-b925-90ab15df8dac\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.354350 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-bound-sa-token\") pod \"cd09450e-798d-41d5-b925-90ab15df8dac\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.354367 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkrch\" (UniqueName: \"kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-kube-api-access-vkrch\") pod \"cd09450e-798d-41d5-b925-90ab15df8dac\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.354402 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cd09450e-798d-41d5-b925-90ab15df8dac-trusted-ca\") pod \"cd09450e-798d-41d5-b925-90ab15df8dac\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.354432 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-registry-tls\") pod \"cd09450e-798d-41d5-b925-90ab15df8dac\" (UID: \"cd09450e-798d-41d5-b925-90ab15df8dac\") " Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.355647 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd09450e-798d-41d5-b925-90ab15df8dac-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "cd09450e-798d-41d5-b925-90ab15df8dac" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.356677 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd09450e-798d-41d5-b925-90ab15df8dac-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "cd09450e-798d-41d5-b925-90ab15df8dac" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.366816 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "cd09450e-798d-41d5-b925-90ab15df8dac" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.374119 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd09450e-798d-41d5-b925-90ab15df8dac-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "cd09450e-798d-41d5-b925-90ab15df8dac" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.374263 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "cd09450e-798d-41d5-b925-90ab15df8dac" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.379461 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd09450e-798d-41d5-b925-90ab15df8dac-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "cd09450e-798d-41d5-b925-90ab15df8dac" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.381564 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-kube-api-access-vkrch" (OuterVolumeSpecName: "kube-api-access-vkrch") pod "cd09450e-798d-41d5-b925-90ab15df8dac" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac"). InnerVolumeSpecName "kube-api-access-vkrch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.389426 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "cd09450e-798d-41d5-b925-90ab15df8dac" (UID: "cd09450e-798d-41d5-b925-90ab15df8dac"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.458497 4997 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cd09450e-798d-41d5-b925-90ab15df8dac-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.458538 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkrch\" (UniqueName: \"kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-kube-api-access-vkrch\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.458553 4997 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.458563 4997 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cd09450e-798d-41d5-b925-90ab15df8dac-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.458572 4997 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cd09450e-798d-41d5-b925-90ab15df8dac-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.458580 4997 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cd09450e-798d-41d5-b925-90ab15df8dac-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:50 crc kubenswrapper[4997]: I1205 07:01:50.458590 4997 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cd09450e-798d-41d5-b925-90ab15df8dac-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 05 07:01:51 crc kubenswrapper[4997]: I1205 07:01:51.126487 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"349f1cbf90f454974a9d2a5680e2433c8f6509c0411a6ed6079bff64bb6e9434"} Dec 05 07:01:51 crc kubenswrapper[4997]: I1205 07:01:51.128855 4997 generic.go:334] "Generic (PLEG): container finished" podID="0b51e69f-d870-494a-adc2-b912ec12e2e6" containerID="88a2d3145ee558236139a8b160b4d0a53d93b3d4f41409e656fbf753093d7e84" exitCode=0 Dec 05 07:01:51 crc kubenswrapper[4997]: I1205 07:01:51.128939 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c46hv" event={"ID":"0b51e69f-d870-494a-adc2-b912ec12e2e6","Type":"ContainerDied","Data":"88a2d3145ee558236139a8b160b4d0a53d93b3d4f41409e656fbf753093d7e84"} Dec 05 07:01:51 crc kubenswrapper[4997]: I1205 07:01:51.128971 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c46hv" event={"ID":"0b51e69f-d870-494a-adc2-b912ec12e2e6","Type":"ContainerStarted","Data":"7a7ec2ae0f89be0d50f53c6c589cd90a8f354ba08da04644ceec5cb817232b98"} Dec 05 07:01:51 crc kubenswrapper[4997]: I1205 07:01:51.130303 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" event={"ID":"cd09450e-798d-41d5-b925-90ab15df8dac","Type":"ContainerDied","Data":"dd16f29590ca291a136e4ff7ad0143adfb64361ee01979c09a146a027afcd726"} Dec 05 07:01:51 crc kubenswrapper[4997]: I1205 07:01:51.130354 4997 scope.go:117] "RemoveContainer" containerID="d489a9092e2cdddebf8931a4ffd9ead2066e0f6d2c9185c1800bd007548a2afb" Dec 05 07:01:51 crc kubenswrapper[4997]: I1205 07:01:51.130484 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5xbd9" Dec 05 07:01:51 crc kubenswrapper[4997]: I1205 07:01:51.134345 4997 generic.go:334] "Generic (PLEG): container finished" podID="d3274e3c-2ad5-4756-8d3f-bae751f1e719" containerID="f33f8296c34e4683e59da1ac65dc94102f8aa3f8cb8c40b4784a9cc92ecd4771" exitCode=0 Dec 05 07:01:51 crc kubenswrapper[4997]: I1205 07:01:51.134397 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zdk9s" event={"ID":"d3274e3c-2ad5-4756-8d3f-bae751f1e719","Type":"ContainerDied","Data":"f33f8296c34e4683e59da1ac65dc94102f8aa3f8cb8c40b4784a9cc92ecd4771"} Dec 05 07:01:51 crc kubenswrapper[4997]: I1205 07:01:51.136843 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktvff" event={"ID":"e5f08766-f19b-4b4f-b1f8-f25b4214457e","Type":"ContainerDied","Data":"dbbdc7aefa80b908c109acb2702b7d33759516f63f802b57d3d8c2eeadac4b95"} Dec 05 07:01:51 crc kubenswrapper[4997]: I1205 07:01:51.136797 4997 generic.go:334] "Generic (PLEG): container finished" podID="e5f08766-f19b-4b4f-b1f8-f25b4214457e" containerID="dbbdc7aefa80b908c109acb2702b7d33759516f63f802b57d3d8c2eeadac4b95" exitCode=0 Dec 05 07:01:51 crc kubenswrapper[4997]: I1205 07:01:51.137165 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktvff" event={"ID":"e5f08766-f19b-4b4f-b1f8-f25b4214457e","Type":"ContainerStarted","Data":"e940064204a5dd99cbc9ef2fd378638a574a1dd8db41ef3ad07fb3da67317dcd"} Dec 05 07:01:51 crc kubenswrapper[4997]: I1205 07:01:51.221724 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5xbd9"] Dec 05 07:01:51 crc kubenswrapper[4997]: I1205 07:01:51.227505 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5xbd9"] Dec 05 07:01:51 crc kubenswrapper[4997]: I1205 07:01:51.756474 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd09450e-798d-41d5-b925-90ab15df8dac" path="/var/lib/kubelet/pods/cd09450e-798d-41d5-b925-90ab15df8dac/volumes" Dec 05 07:01:52 crc kubenswrapper[4997]: I1205 07:01:52.144210 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zdk9s" event={"ID":"d3274e3c-2ad5-4756-8d3f-bae751f1e719","Type":"ContainerStarted","Data":"9c2afaa8dfee20860b91a43a1030662f1087186da9f0d30e7927675d98f0bf48"} Dec 05 07:01:52 crc kubenswrapper[4997]: I1205 07:01:52.147129 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktvff" event={"ID":"e5f08766-f19b-4b4f-b1f8-f25b4214457e","Type":"ContainerStarted","Data":"1886300a7d9bec175e9c018183aecb4feec11989c8efe076edfb9fa4fb1444a5"} Dec 05 07:01:52 crc kubenswrapper[4997]: I1205 07:01:52.148877 4997 generic.go:334] "Generic (PLEG): container finished" podID="0b51e69f-d870-494a-adc2-b912ec12e2e6" containerID="cc8cb2e3205858304671252edbd0a14d98bb53bdfc89c45415f14797e71c961c" exitCode=0 Dec 05 07:01:52 crc kubenswrapper[4997]: I1205 07:01:52.149829 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c46hv" event={"ID":"0b51e69f-d870-494a-adc2-b912ec12e2e6","Type":"ContainerDied","Data":"cc8cb2e3205858304671252edbd0a14d98bb53bdfc89c45415f14797e71c961c"} Dec 05 07:01:52 crc kubenswrapper[4997]: I1205 07:01:52.178270 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zdk9s" podStartSLOduration=2.668330273 podStartE2EDuration="5.178257238s" podCreationTimestamp="2025-12-05 07:01:47 +0000 UTC" firstStartedPulling="2025-12-05 07:01:49.065248464 +0000 UTC m=+409.594155735" lastFinishedPulling="2025-12-05 07:01:51.575175439 +0000 UTC m=+412.104082700" observedRunningTime="2025-12-05 07:01:52.176642475 +0000 UTC m=+412.705549746" watchObservedRunningTime="2025-12-05 07:01:52.178257238 +0000 UTC m=+412.707164499" Dec 05 07:01:53 crc kubenswrapper[4997]: I1205 07:01:53.156677 4997 generic.go:334] "Generic (PLEG): container finished" podID="e5f08766-f19b-4b4f-b1f8-f25b4214457e" containerID="1886300a7d9bec175e9c018183aecb4feec11989c8efe076edfb9fa4fb1444a5" exitCode=0 Dec 05 07:01:53 crc kubenswrapper[4997]: I1205 07:01:53.156769 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktvff" event={"ID":"e5f08766-f19b-4b4f-b1f8-f25b4214457e","Type":"ContainerDied","Data":"1886300a7d9bec175e9c018183aecb4feec11989c8efe076edfb9fa4fb1444a5"} Dec 05 07:01:53 crc kubenswrapper[4997]: I1205 07:01:53.159637 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c46hv" event={"ID":"0b51e69f-d870-494a-adc2-b912ec12e2e6","Type":"ContainerStarted","Data":"8a2e28f66eed828799ceedf21b17c31f351bf16b2025b6d48c68d47be814fdde"} Dec 05 07:01:53 crc kubenswrapper[4997]: I1205 07:01:53.205516 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c46hv" podStartSLOduration=2.801915168 podStartE2EDuration="4.205498111s" podCreationTimestamp="2025-12-05 07:01:49 +0000 UTC" firstStartedPulling="2025-12-05 07:01:51.135976547 +0000 UTC m=+411.664883808" lastFinishedPulling="2025-12-05 07:01:52.53955949 +0000 UTC m=+413.068466751" observedRunningTime="2025-12-05 07:01:53.20397469 +0000 UTC m=+413.732881971" watchObservedRunningTime="2025-12-05 07:01:53.205498111 +0000 UTC m=+413.734405372" Dec 05 07:01:54 crc kubenswrapper[4997]: I1205 07:01:54.174020 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktvff" event={"ID":"e5f08766-f19b-4b4f-b1f8-f25b4214457e","Type":"ContainerStarted","Data":"0a398d0771633d79cdb7807a72990ec23f7a33da1068a769c8ca556d250636fa"} Dec 05 07:01:54 crc kubenswrapper[4997]: I1205 07:01:54.196260 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ktvff" podStartSLOduration=2.771125633 podStartE2EDuration="5.196240709s" podCreationTimestamp="2025-12-05 07:01:49 +0000 UTC" firstStartedPulling="2025-12-05 07:01:51.137922111 +0000 UTC m=+411.666829372" lastFinishedPulling="2025-12-05 07:01:53.563037187 +0000 UTC m=+414.091944448" observedRunningTime="2025-12-05 07:01:54.194137279 +0000 UTC m=+414.723044550" watchObservedRunningTime="2025-12-05 07:01:54.196240709 +0000 UTC m=+414.725147960" Dec 05 07:01:57 crc kubenswrapper[4997]: I1205 07:01:57.217230 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qxz27" Dec 05 07:01:57 crc kubenswrapper[4997]: I1205 07:01:57.219889 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qxz27" Dec 05 07:01:57 crc kubenswrapper[4997]: I1205 07:01:57.269578 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qxz27" Dec 05 07:01:57 crc kubenswrapper[4997]: I1205 07:01:57.413696 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zdk9s" Dec 05 07:01:57 crc kubenswrapper[4997]: I1205 07:01:57.413941 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zdk9s" Dec 05 07:01:57 crc kubenswrapper[4997]: I1205 07:01:57.455289 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zdk9s" Dec 05 07:01:58 crc kubenswrapper[4997]: I1205 07:01:58.230687 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qxz27" Dec 05 07:01:58 crc kubenswrapper[4997]: I1205 07:01:58.230883 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zdk9s" Dec 05 07:01:59 crc kubenswrapper[4997]: I1205 07:01:59.603795 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c46hv" Dec 05 07:01:59 crc kubenswrapper[4997]: I1205 07:01:59.603837 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c46hv" Dec 05 07:01:59 crc kubenswrapper[4997]: I1205 07:01:59.652797 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c46hv" Dec 05 07:01:59 crc kubenswrapper[4997]: I1205 07:01:59.806894 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ktvff" Dec 05 07:01:59 crc kubenswrapper[4997]: I1205 07:01:59.806946 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ktvff" Dec 05 07:01:59 crc kubenswrapper[4997]: I1205 07:01:59.847116 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ktvff" Dec 05 07:02:00 crc kubenswrapper[4997]: I1205 07:02:00.257675 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ktvff" Dec 05 07:02:00 crc kubenswrapper[4997]: I1205 07:02:00.264271 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c46hv" Dec 05 07:04:19 crc kubenswrapper[4997]: I1205 07:04:19.769838 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:04:19 crc kubenswrapper[4997]: I1205 07:04:19.770670 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:04:49 crc kubenswrapper[4997]: I1205 07:04:49.770971 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:04:49 crc kubenswrapper[4997]: I1205 07:04:49.772930 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:05:19 crc kubenswrapper[4997]: I1205 07:05:19.770245 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:05:19 crc kubenswrapper[4997]: I1205 07:05:19.770895 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:05:19 crc kubenswrapper[4997]: I1205 07:05:19.770960 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 07:05:19 crc kubenswrapper[4997]: I1205 07:05:19.771803 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"349f1cbf90f454974a9d2a5680e2433c8f6509c0411a6ed6079bff64bb6e9434"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 07:05:19 crc kubenswrapper[4997]: I1205 07:05:19.771875 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://349f1cbf90f454974a9d2a5680e2433c8f6509c0411a6ed6079bff64bb6e9434" gracePeriod=600 Dec 05 07:05:20 crc kubenswrapper[4997]: I1205 07:05:20.896224 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="349f1cbf90f454974a9d2a5680e2433c8f6509c0411a6ed6079bff64bb6e9434" exitCode=0 Dec 05 07:05:20 crc kubenswrapper[4997]: I1205 07:05:20.896368 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"349f1cbf90f454974a9d2a5680e2433c8f6509c0411a6ed6079bff64bb6e9434"} Dec 05 07:05:20 crc kubenswrapper[4997]: I1205 07:05:20.896822 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"1ff8d1fdf0f1aff6bb5ff2695a5046a42e548eeb16bf2838702262c451fcb00b"} Dec 05 07:05:20 crc kubenswrapper[4997]: I1205 07:05:20.896853 4997 scope.go:117] "RemoveContainer" containerID="2f6d912e8d30078a8f27c58e585fd50e33dbd4541657a6aef0990fa4ea86b3b6" Dec 05 07:07:40 crc kubenswrapper[4997]: I1205 07:07:40.868386 4997 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 07:07:49 crc kubenswrapper[4997]: I1205 07:07:49.770009 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:07:49 crc kubenswrapper[4997]: I1205 07:07:49.771942 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:08:19 crc kubenswrapper[4997]: I1205 07:08:19.770899 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:08:19 crc kubenswrapper[4997]: I1205 07:08:19.771894 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:08:49 crc kubenswrapper[4997]: I1205 07:08:49.770498 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:08:49 crc kubenswrapper[4997]: I1205 07:08:49.771423 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:08:49 crc kubenswrapper[4997]: I1205 07:08:49.771487 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 07:08:49 crc kubenswrapper[4997]: I1205 07:08:49.772407 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1ff8d1fdf0f1aff6bb5ff2695a5046a42e548eeb16bf2838702262c451fcb00b"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 07:08:49 crc kubenswrapper[4997]: I1205 07:08:49.772499 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://1ff8d1fdf0f1aff6bb5ff2695a5046a42e548eeb16bf2838702262c451fcb00b" gracePeriod=600 Dec 05 07:08:50 crc kubenswrapper[4997]: I1205 07:08:50.340930 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="1ff8d1fdf0f1aff6bb5ff2695a5046a42e548eeb16bf2838702262c451fcb00b" exitCode=0 Dec 05 07:08:50 crc kubenswrapper[4997]: I1205 07:08:50.341017 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"1ff8d1fdf0f1aff6bb5ff2695a5046a42e548eeb16bf2838702262c451fcb00b"} Dec 05 07:08:50 crc kubenswrapper[4997]: I1205 07:08:50.341773 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"046abc9c3b110cde88610f054e681c42bd85fac9c9e032fa744fe8940f595f0c"} Dec 05 07:08:50 crc kubenswrapper[4997]: I1205 07:08:50.341799 4997 scope.go:117] "RemoveContainer" containerID="349f1cbf90f454974a9d2a5680e2433c8f6509c0411a6ed6079bff64bb6e9434" Dec 05 07:10:05 crc kubenswrapper[4997]: I1205 07:10:05.916929 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pcr86"] Dec 05 07:10:05 crc kubenswrapper[4997]: E1205 07:10:05.919464 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd09450e-798d-41d5-b925-90ab15df8dac" containerName="registry" Dec 05 07:10:05 crc kubenswrapper[4997]: I1205 07:10:05.920181 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd09450e-798d-41d5-b925-90ab15df8dac" containerName="registry" Dec 05 07:10:05 crc kubenswrapper[4997]: I1205 07:10:05.920435 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd09450e-798d-41d5-b925-90ab15df8dac" containerName="registry" Dec 05 07:10:05 crc kubenswrapper[4997]: I1205 07:10:05.922193 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pcr86" Dec 05 07:10:05 crc kubenswrapper[4997]: I1205 07:10:05.937811 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pcr86"] Dec 05 07:10:05 crc kubenswrapper[4997]: I1205 07:10:05.972366 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-catalog-content\") pod \"redhat-marketplace-pcr86\" (UID: \"79a0f59e-c3c1-4252-94d3-d5a34f12ef31\") " pod="openshift-marketplace/redhat-marketplace-pcr86" Dec 05 07:10:05 crc kubenswrapper[4997]: I1205 07:10:05.972593 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-utilities\") pod \"redhat-marketplace-pcr86\" (UID: \"79a0f59e-c3c1-4252-94d3-d5a34f12ef31\") " pod="openshift-marketplace/redhat-marketplace-pcr86" Dec 05 07:10:05 crc kubenswrapper[4997]: I1205 07:10:05.972731 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh7rv\" (UniqueName: \"kubernetes.io/projected/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-kube-api-access-sh7rv\") pod \"redhat-marketplace-pcr86\" (UID: \"79a0f59e-c3c1-4252-94d3-d5a34f12ef31\") " pod="openshift-marketplace/redhat-marketplace-pcr86" Dec 05 07:10:06 crc kubenswrapper[4997]: I1205 07:10:06.074162 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-utilities\") pod \"redhat-marketplace-pcr86\" (UID: \"79a0f59e-c3c1-4252-94d3-d5a34f12ef31\") " pod="openshift-marketplace/redhat-marketplace-pcr86" Dec 05 07:10:06 crc kubenswrapper[4997]: I1205 07:10:06.074236 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh7rv\" (UniqueName: \"kubernetes.io/projected/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-kube-api-access-sh7rv\") pod \"redhat-marketplace-pcr86\" (UID: \"79a0f59e-c3c1-4252-94d3-d5a34f12ef31\") " pod="openshift-marketplace/redhat-marketplace-pcr86" Dec 05 07:10:06 crc kubenswrapper[4997]: I1205 07:10:06.074286 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-catalog-content\") pod \"redhat-marketplace-pcr86\" (UID: \"79a0f59e-c3c1-4252-94d3-d5a34f12ef31\") " pod="openshift-marketplace/redhat-marketplace-pcr86" Dec 05 07:10:06 crc kubenswrapper[4997]: I1205 07:10:06.074816 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-utilities\") pod \"redhat-marketplace-pcr86\" (UID: \"79a0f59e-c3c1-4252-94d3-d5a34f12ef31\") " pod="openshift-marketplace/redhat-marketplace-pcr86" Dec 05 07:10:06 crc kubenswrapper[4997]: I1205 07:10:06.074894 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-catalog-content\") pod \"redhat-marketplace-pcr86\" (UID: \"79a0f59e-c3c1-4252-94d3-d5a34f12ef31\") " pod="openshift-marketplace/redhat-marketplace-pcr86" Dec 05 07:10:06 crc kubenswrapper[4997]: I1205 07:10:06.097117 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh7rv\" (UniqueName: \"kubernetes.io/projected/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-kube-api-access-sh7rv\") pod \"redhat-marketplace-pcr86\" (UID: \"79a0f59e-c3c1-4252-94d3-d5a34f12ef31\") " pod="openshift-marketplace/redhat-marketplace-pcr86" Dec 05 07:10:06 crc kubenswrapper[4997]: I1205 07:10:06.254466 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pcr86" Dec 05 07:10:06 crc kubenswrapper[4997]: I1205 07:10:06.480987 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pcr86"] Dec 05 07:10:06 crc kubenswrapper[4997]: I1205 07:10:06.836900 4997 generic.go:334] "Generic (PLEG): container finished" podID="79a0f59e-c3c1-4252-94d3-d5a34f12ef31" containerID="a5aab3541b878dc9e58573c0a4607c74bf6139444679149b1f8273a4b271c214" exitCode=0 Dec 05 07:10:06 crc kubenswrapper[4997]: I1205 07:10:06.836987 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcr86" event={"ID":"79a0f59e-c3c1-4252-94d3-d5a34f12ef31","Type":"ContainerDied","Data":"a5aab3541b878dc9e58573c0a4607c74bf6139444679149b1f8273a4b271c214"} Dec 05 07:10:06 crc kubenswrapper[4997]: I1205 07:10:06.837072 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcr86" event={"ID":"79a0f59e-c3c1-4252-94d3-d5a34f12ef31","Type":"ContainerStarted","Data":"c6fb67d4f5a76b2fe0a451afede0ee43e28424dc609fb431984d715f85bcaf2c"} Dec 05 07:10:06 crc kubenswrapper[4997]: I1205 07:10:06.839758 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 07:10:07 crc kubenswrapper[4997]: I1205 07:10:07.845548 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcr86" event={"ID":"79a0f59e-c3c1-4252-94d3-d5a34f12ef31","Type":"ContainerStarted","Data":"1cfd4914dfe2cc3ab5d46c191b64c7fd75bb5b97668d055a8e104a14cbccfd2e"} Dec 05 07:10:08 crc kubenswrapper[4997]: I1205 07:10:08.856945 4997 generic.go:334] "Generic (PLEG): container finished" podID="79a0f59e-c3c1-4252-94d3-d5a34f12ef31" containerID="1cfd4914dfe2cc3ab5d46c191b64c7fd75bb5b97668d055a8e104a14cbccfd2e" exitCode=0 Dec 05 07:10:08 crc kubenswrapper[4997]: I1205 07:10:08.857080 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcr86" event={"ID":"79a0f59e-c3c1-4252-94d3-d5a34f12ef31","Type":"ContainerDied","Data":"1cfd4914dfe2cc3ab5d46c191b64c7fd75bb5b97668d055a8e104a14cbccfd2e"} Dec 05 07:10:09 crc kubenswrapper[4997]: I1205 07:10:09.868442 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcr86" event={"ID":"79a0f59e-c3c1-4252-94d3-d5a34f12ef31","Type":"ContainerStarted","Data":"716649fa22a242741a153fecce6104021e37396c1164cfde7074eb09f4889277"} Dec 05 07:10:09 crc kubenswrapper[4997]: I1205 07:10:09.897775 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pcr86" podStartSLOduration=2.455286622 podStartE2EDuration="4.897726788s" podCreationTimestamp="2025-12-05 07:10:05 +0000 UTC" firstStartedPulling="2025-12-05 07:10:06.839413073 +0000 UTC m=+907.368320334" lastFinishedPulling="2025-12-05 07:10:09.281853189 +0000 UTC m=+909.810760500" observedRunningTime="2025-12-05 07:10:09.894241628 +0000 UTC m=+910.423148889" watchObservedRunningTime="2025-12-05 07:10:09.897726788 +0000 UTC m=+910.426634129" Dec 05 07:10:16 crc kubenswrapper[4997]: I1205 07:10:16.255660 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pcr86" Dec 05 07:10:16 crc kubenswrapper[4997]: I1205 07:10:16.256484 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pcr86" Dec 05 07:10:16 crc kubenswrapper[4997]: I1205 07:10:16.320535 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pcr86" Dec 05 07:10:16 crc kubenswrapper[4997]: I1205 07:10:16.950438 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pcr86" Dec 05 07:10:16 crc kubenswrapper[4997]: I1205 07:10:16.996805 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pcr86"] Dec 05 07:10:18 crc kubenswrapper[4997]: I1205 07:10:18.915982 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pcr86" podUID="79a0f59e-c3c1-4252-94d3-d5a34f12ef31" containerName="registry-server" containerID="cri-o://716649fa22a242741a153fecce6104021e37396c1164cfde7074eb09f4889277" gracePeriod=2 Dec 05 07:10:20 crc kubenswrapper[4997]: I1205 07:10:20.928856 4997 generic.go:334] "Generic (PLEG): container finished" podID="79a0f59e-c3c1-4252-94d3-d5a34f12ef31" containerID="716649fa22a242741a153fecce6104021e37396c1164cfde7074eb09f4889277" exitCode=0 Dec 05 07:10:20 crc kubenswrapper[4997]: I1205 07:10:20.929305 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcr86" event={"ID":"79a0f59e-c3c1-4252-94d3-d5a34f12ef31","Type":"ContainerDied","Data":"716649fa22a242741a153fecce6104021e37396c1164cfde7074eb09f4889277"} Dec 05 07:10:21 crc kubenswrapper[4997]: I1205 07:10:21.137641 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pcr86" Dec 05 07:10:21 crc kubenswrapper[4997]: I1205 07:10:21.198024 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-utilities\") pod \"79a0f59e-c3c1-4252-94d3-d5a34f12ef31\" (UID: \"79a0f59e-c3c1-4252-94d3-d5a34f12ef31\") " Dec 05 07:10:21 crc kubenswrapper[4997]: I1205 07:10:21.198084 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sh7rv\" (UniqueName: \"kubernetes.io/projected/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-kube-api-access-sh7rv\") pod \"79a0f59e-c3c1-4252-94d3-d5a34f12ef31\" (UID: \"79a0f59e-c3c1-4252-94d3-d5a34f12ef31\") " Dec 05 07:10:21 crc kubenswrapper[4997]: I1205 07:10:21.198110 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-catalog-content\") pod \"79a0f59e-c3c1-4252-94d3-d5a34f12ef31\" (UID: \"79a0f59e-c3c1-4252-94d3-d5a34f12ef31\") " Dec 05 07:10:21 crc kubenswrapper[4997]: I1205 07:10:21.199045 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-utilities" (OuterVolumeSpecName: "utilities") pod "79a0f59e-c3c1-4252-94d3-d5a34f12ef31" (UID: "79a0f59e-c3c1-4252-94d3-d5a34f12ef31"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:10:21 crc kubenswrapper[4997]: I1205 07:10:21.205839 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-kube-api-access-sh7rv" (OuterVolumeSpecName: "kube-api-access-sh7rv") pod "79a0f59e-c3c1-4252-94d3-d5a34f12ef31" (UID: "79a0f59e-c3c1-4252-94d3-d5a34f12ef31"). InnerVolumeSpecName "kube-api-access-sh7rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:10:21 crc kubenswrapper[4997]: I1205 07:10:21.217883 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "79a0f59e-c3c1-4252-94d3-d5a34f12ef31" (UID: "79a0f59e-c3c1-4252-94d3-d5a34f12ef31"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:10:21 crc kubenswrapper[4997]: I1205 07:10:21.312182 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sh7rv\" (UniqueName: \"kubernetes.io/projected/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-kube-api-access-sh7rv\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:21 crc kubenswrapper[4997]: I1205 07:10:21.312245 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:21 crc kubenswrapper[4997]: I1205 07:10:21.312258 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79a0f59e-c3c1-4252-94d3-d5a34f12ef31-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:21 crc kubenswrapper[4997]: E1205 07:10:21.886307 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79a0f59e_c3c1_4252_94d3_d5a34f12ef31.slice/crio-c6fb67d4f5a76b2fe0a451afede0ee43e28424dc609fb431984d715f85bcaf2c\": RecentStats: unable to find data in memory cache]" Dec 05 07:10:21 crc kubenswrapper[4997]: I1205 07:10:21.939243 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcr86" event={"ID":"79a0f59e-c3c1-4252-94d3-d5a34f12ef31","Type":"ContainerDied","Data":"c6fb67d4f5a76b2fe0a451afede0ee43e28424dc609fb431984d715f85bcaf2c"} Dec 05 07:10:21 crc kubenswrapper[4997]: I1205 07:10:21.939304 4997 scope.go:117] "RemoveContainer" containerID="716649fa22a242741a153fecce6104021e37396c1164cfde7074eb09f4889277" Dec 05 07:10:21 crc kubenswrapper[4997]: I1205 07:10:21.939455 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pcr86" Dec 05 07:10:21 crc kubenswrapper[4997]: I1205 07:10:21.969101 4997 scope.go:117] "RemoveContainer" containerID="1cfd4914dfe2cc3ab5d46c191b64c7fd75bb5b97668d055a8e104a14cbccfd2e" Dec 05 07:10:21 crc kubenswrapper[4997]: I1205 07:10:21.969746 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pcr86"] Dec 05 07:10:21 crc kubenswrapper[4997]: I1205 07:10:21.978982 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pcr86"] Dec 05 07:10:21 crc kubenswrapper[4997]: I1205 07:10:21.992919 4997 scope.go:117] "RemoveContainer" containerID="a5aab3541b878dc9e58573c0a4607c74bf6139444679149b1f8273a4b271c214" Dec 05 07:10:23 crc kubenswrapper[4997]: I1205 07:10:23.758934 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79a0f59e-c3c1-4252-94d3-d5a34f12ef31" path="/var/lib/kubelet/pods/79a0f59e-c3c1-4252-94d3-d5a34f12ef31/volumes" Dec 05 07:10:41 crc kubenswrapper[4997]: I1205 07:10:41.744079 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bchpk"] Dec 05 07:10:41 crc kubenswrapper[4997]: I1205 07:10:41.745229 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovn-controller" containerID="cri-o://eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798" gracePeriod=30 Dec 05 07:10:41 crc kubenswrapper[4997]: I1205 07:10:41.745327 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="nbdb" containerID="cri-o://8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b" gracePeriod=30 Dec 05 07:10:41 crc kubenswrapper[4997]: I1205 07:10:41.745429 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb" gracePeriod=30 Dec 05 07:10:41 crc kubenswrapper[4997]: I1205 07:10:41.745460 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="northd" containerID="cri-o://ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34" gracePeriod=30 Dec 05 07:10:41 crc kubenswrapper[4997]: I1205 07:10:41.745485 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="kube-rbac-proxy-node" containerID="cri-o://46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf" gracePeriod=30 Dec 05 07:10:41 crc kubenswrapper[4997]: I1205 07:10:41.745475 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="sbdb" containerID="cri-o://742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c" gracePeriod=30 Dec 05 07:10:41 crc kubenswrapper[4997]: I1205 07:10:41.745556 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovn-acl-logging" containerID="cri-o://cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c" gracePeriod=30 Dec 05 07:10:41 crc kubenswrapper[4997]: I1205 07:10:41.772205 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovnkube-controller" containerID="cri-o://9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa" gracePeriod=30 Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.032987 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovnkube-controller/3.log" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.036154 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovn-acl-logging/0.log" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.036657 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovn-controller/0.log" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.037203 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.069837 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovnkube-controller/3.log" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.071878 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovn-acl-logging/0.log" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.072440 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bchpk_28f65c70-f59b-4393-b9c9-8163bb596f6c/ovn-controller/0.log" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073038 4997 generic.go:334] "Generic (PLEG): container finished" podID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerID="9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa" exitCode=0 Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073085 4997 generic.go:334] "Generic (PLEG): container finished" podID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerID="742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c" exitCode=0 Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073097 4997 generic.go:334] "Generic (PLEG): container finished" podID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerID="8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b" exitCode=0 Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073107 4997 generic.go:334] "Generic (PLEG): container finished" podID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerID="ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34" exitCode=0 Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073115 4997 generic.go:334] "Generic (PLEG): container finished" podID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerID="efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb" exitCode=0 Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073121 4997 generic.go:334] "Generic (PLEG): container finished" podID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerID="46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf" exitCode=0 Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073127 4997 generic.go:334] "Generic (PLEG): container finished" podID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerID="cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c" exitCode=143 Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073134 4997 generic.go:334] "Generic (PLEG): container finished" podID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerID="eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798" exitCode=143 Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073148 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073185 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerDied","Data":"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073224 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerDied","Data":"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073237 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerDied","Data":"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073246 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerDied","Data":"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073255 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerDied","Data":"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073266 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerDied","Data":"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073277 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073288 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073294 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073299 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073305 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073311 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073317 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073322 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073327 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073335 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerDied","Data":"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073342 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073349 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073385 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073393 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073398 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073403 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073409 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073414 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073420 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073426 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073434 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerDied","Data":"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073444 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073450 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073456 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073461 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073467 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073473 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073479 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073485 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073490 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073495 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073503 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bchpk" event={"ID":"28f65c70-f59b-4393-b9c9-8163bb596f6c","Type":"ContainerDied","Data":"897a7454e3968995ef11164249ce88aa5e11662142735a63b34b5d3056e89dbd"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073512 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073518 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073524 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073530 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073536 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073611 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073635 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073641 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073647 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073652 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.073550 4997 scope.go:117] "RemoveContainer" containerID="9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.081371 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-th9dq_85f05eb3-df7f-4c46-a0e2-451766bb3297/kube-multus/2.log" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.082044 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-th9dq_85f05eb3-df7f-4c46-a0e2-451766bb3297/kube-multus/1.log" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.082084 4997 generic.go:334] "Generic (PLEG): container finished" podID="85f05eb3-df7f-4c46-a0e2-451766bb3297" containerID="41601dcf153b1e00253366b960707c5217ee9e19312ec6eb9f65a38e45cbcdd7" exitCode=2 Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.082119 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-th9dq" event={"ID":"85f05eb3-df7f-4c46-a0e2-451766bb3297","Type":"ContainerDied","Data":"41601dcf153b1e00253366b960707c5217ee9e19312ec6eb9f65a38e45cbcdd7"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.082146 4997 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ba8817aab2ae020cc998813edb099030cc3c3c6eb11e8f175e3cfd82b18686b2"} Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.082682 4997 scope.go:117] "RemoveContainer" containerID="41601dcf153b1e00253366b960707c5217ee9e19312ec6eb9f65a38e45cbcdd7" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.100686 4997 scope.go:117] "RemoveContainer" containerID="efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.109700 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-87ct9"] Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.109999 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="sbdb" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110012 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="sbdb" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.110023 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovnkube-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110029 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovnkube-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.110037 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovnkube-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110043 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovnkube-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.110050 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79a0f59e-c3c1-4252-94d3-d5a34f12ef31" containerName="extract-content" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110056 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="79a0f59e-c3c1-4252-94d3-d5a34f12ef31" containerName="extract-content" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.110064 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovnkube-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110070 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovnkube-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.110078 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovn-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110084 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovn-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.110092 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="nbdb" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110097 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="nbdb" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.110106 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="kubecfg-setup" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110111 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="kubecfg-setup" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.110118 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="kube-rbac-proxy-node" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110125 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="kube-rbac-proxy-node" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.110134 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79a0f59e-c3c1-4252-94d3-d5a34f12ef31" containerName="registry-server" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110140 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="79a0f59e-c3c1-4252-94d3-d5a34f12ef31" containerName="registry-server" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.110149 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovn-acl-logging" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110155 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovn-acl-logging" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.110161 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110169 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.110179 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="northd" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110184 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="northd" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.110193 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79a0f59e-c3c1-4252-94d3-d5a34f12ef31" containerName="extract-utilities" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110199 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="79a0f59e-c3c1-4252-94d3-d5a34f12ef31" containerName="extract-utilities" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110297 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="northd" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110306 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovnkube-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110312 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovnkube-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110319 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="79a0f59e-c3c1-4252-94d3-d5a34f12ef31" containerName="registry-server" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110325 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovn-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110333 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="nbdb" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110342 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110351 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="kube-rbac-proxy-node" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110358 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovnkube-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110366 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovn-acl-logging" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110374 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="sbdb" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.110515 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovnkube-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110523 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovnkube-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.110534 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovnkube-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110564 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovnkube-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110680 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovnkube-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.110689 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" containerName="ovnkube-controller" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.114030 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.135777 4997 scope.go:117] "RemoveContainer" containerID="742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.168677 4997 scope.go:117] "RemoveContainer" containerID="8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183086 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vmkd\" (UniqueName: \"kubernetes.io/projected/28f65c70-f59b-4393-b9c9-8163bb596f6c-kube-api-access-2vmkd\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183139 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-node-log\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183172 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-var-lib-openvswitch\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183195 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-systemd-units\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183211 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-run-netns\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183232 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovnkube-script-lib\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183250 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-kubelet\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183273 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-slash\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183288 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-openvswitch\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183321 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-env-overrides\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183386 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-ovn\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183409 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-systemd\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183437 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-etc-openvswitch\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183458 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovnkube-config\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183472 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-log-socket\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183486 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-cni-netd\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183511 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-cni-bin\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183538 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovn-node-metrics-cert\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183561 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-run-ovn-kubernetes\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.183578 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"28f65c70-f59b-4393-b9c9-8163bb596f6c\" (UID: \"28f65c70-f59b-4393-b9c9-8163bb596f6c\") " Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.184346 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.184433 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-node-log" (OuterVolumeSpecName: "node-log") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.184456 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.184477 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.184495 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.184786 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.184990 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.185026 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.185047 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-slash" (OuterVolumeSpecName: "host-slash") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.185069 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.185411 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.185783 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.186155 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.186191 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-log-socket" (OuterVolumeSpecName: "log-socket") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.186213 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.186273 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.186356 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.193810 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28f65c70-f59b-4393-b9c9-8163bb596f6c-kube-api-access-2vmkd" (OuterVolumeSpecName: "kube-api-access-2vmkd") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "kube-api-access-2vmkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.201803 4997 scope.go:117] "RemoveContainer" containerID="ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.208122 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.211869 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "28f65c70-f59b-4393-b9c9-8163bb596f6c" (UID: "28f65c70-f59b-4393-b9c9-8163bb596f6c"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.236914 4997 scope.go:117] "RemoveContainer" containerID="efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.261395 4997 scope.go:117] "RemoveContainer" containerID="46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.277967 4997 scope.go:117] "RemoveContainer" containerID="cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.285580 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-ovn-node-metrics-cert\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.285655 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-run-ovn\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.285733 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-systemd-units\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.285768 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-run-ovn-kubernetes\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.285938 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-cni-bin\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286154 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-etc-openvswitch\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286223 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-var-lib-openvswitch\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286250 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-run-openvswitch\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286293 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-run-netns\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286414 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-ovnkube-script-lib\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286442 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-log-socket\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286459 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-slash\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286476 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-ovnkube-config\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286496 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-node-log\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286522 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-env-overrides\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286556 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4tvh\" (UniqueName: \"kubernetes.io/projected/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-kube-api-access-c4tvh\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286573 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-run-systemd\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286601 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-kubelet\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286641 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286666 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-cni-netd\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286745 4997 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286767 4997 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.286783 4997 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.287094 4997 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.287107 4997 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-log-socket\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.287118 4997 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.287128 4997 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.287136 4997 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.287147 4997 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.287156 4997 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.287166 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vmkd\" (UniqueName: \"kubernetes.io/projected/28f65c70-f59b-4393-b9c9-8163bb596f6c-kube-api-access-2vmkd\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.287175 4997 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-node-log\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.287185 4997 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.287194 4997 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.287202 4997 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.287219 4997 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.287231 4997 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.287240 4997 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-host-slash\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.287250 4997 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28f65c70-f59b-4393-b9c9-8163bb596f6c-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.287258 4997 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28f65c70-f59b-4393-b9c9-8163bb596f6c-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.292486 4997 scope.go:117] "RemoveContainer" containerID="eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.307070 4997 scope.go:117] "RemoveContainer" containerID="a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.321139 4997 scope.go:117] "RemoveContainer" containerID="9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.321748 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa\": container with ID starting with 9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa not found: ID does not exist" containerID="9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.321779 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa"} err="failed to get container status \"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa\": rpc error: code = NotFound desc = could not find container \"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa\": container with ID starting with 9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.321835 4997 scope.go:117] "RemoveContainer" containerID="efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.322878 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57\": container with ID starting with efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57 not found: ID does not exist" containerID="efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.322899 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57"} err="failed to get container status \"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57\": rpc error: code = NotFound desc = could not find container \"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57\": container with ID starting with efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57 not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.322912 4997 scope.go:117] "RemoveContainer" containerID="742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.323417 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\": container with ID starting with 742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c not found: ID does not exist" containerID="742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.323435 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c"} err="failed to get container status \"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\": rpc error: code = NotFound desc = could not find container \"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\": container with ID starting with 742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.323448 4997 scope.go:117] "RemoveContainer" containerID="8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.323727 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\": container with ID starting with 8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b not found: ID does not exist" containerID="8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.323750 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b"} err="failed to get container status \"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\": rpc error: code = NotFound desc = could not find container \"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\": container with ID starting with 8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.323762 4997 scope.go:117] "RemoveContainer" containerID="ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.323993 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\": container with ID starting with ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34 not found: ID does not exist" containerID="ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.324010 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34"} err="failed to get container status \"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\": rpc error: code = NotFound desc = could not find container \"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\": container with ID starting with ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34 not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.324022 4997 scope.go:117] "RemoveContainer" containerID="efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.324244 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\": container with ID starting with efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb not found: ID does not exist" containerID="efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.324260 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb"} err="failed to get container status \"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\": rpc error: code = NotFound desc = could not find container \"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\": container with ID starting with efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.324275 4997 scope.go:117] "RemoveContainer" containerID="46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.324546 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\": container with ID starting with 46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf not found: ID does not exist" containerID="46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.324564 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf"} err="failed to get container status \"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\": rpc error: code = NotFound desc = could not find container \"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\": container with ID starting with 46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.324577 4997 scope.go:117] "RemoveContainer" containerID="cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.324839 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\": container with ID starting with cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c not found: ID does not exist" containerID="cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.324858 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c"} err="failed to get container status \"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\": rpc error: code = NotFound desc = could not find container \"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\": container with ID starting with cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.324869 4997 scope.go:117] "RemoveContainer" containerID="eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.325096 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\": container with ID starting with eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798 not found: ID does not exist" containerID="eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.325115 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798"} err="failed to get container status \"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\": rpc error: code = NotFound desc = could not find container \"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\": container with ID starting with eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798 not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.325130 4997 scope.go:117] "RemoveContainer" containerID="a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2" Dec 05 07:10:42 crc kubenswrapper[4997]: E1205 07:10:42.325350 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\": container with ID starting with a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2 not found: ID does not exist" containerID="a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.325367 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2"} err="failed to get container status \"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\": rpc error: code = NotFound desc = could not find container \"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\": container with ID starting with a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2 not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.325380 4997 scope.go:117] "RemoveContainer" containerID="9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.325782 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa"} err="failed to get container status \"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa\": rpc error: code = NotFound desc = could not find container \"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa\": container with ID starting with 9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.325802 4997 scope.go:117] "RemoveContainer" containerID="efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.326158 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57"} err="failed to get container status \"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57\": rpc error: code = NotFound desc = could not find container \"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57\": container with ID starting with efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57 not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.326181 4997 scope.go:117] "RemoveContainer" containerID="742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.326463 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c"} err="failed to get container status \"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\": rpc error: code = NotFound desc = could not find container \"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\": container with ID starting with 742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.326483 4997 scope.go:117] "RemoveContainer" containerID="8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.326728 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b"} err="failed to get container status \"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\": rpc error: code = NotFound desc = could not find container \"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\": container with ID starting with 8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.326749 4997 scope.go:117] "RemoveContainer" containerID="ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.326947 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34"} err="failed to get container status \"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\": rpc error: code = NotFound desc = could not find container \"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\": container with ID starting with ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34 not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.326962 4997 scope.go:117] "RemoveContainer" containerID="efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.327149 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb"} err="failed to get container status \"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\": rpc error: code = NotFound desc = could not find container \"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\": container with ID starting with efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.327164 4997 scope.go:117] "RemoveContainer" containerID="46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.327349 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf"} err="failed to get container status \"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\": rpc error: code = NotFound desc = could not find container \"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\": container with ID starting with 46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.327364 4997 scope.go:117] "RemoveContainer" containerID="cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.327721 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c"} err="failed to get container status \"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\": rpc error: code = NotFound desc = could not find container \"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\": container with ID starting with cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.327738 4997 scope.go:117] "RemoveContainer" containerID="eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.328088 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798"} err="failed to get container status \"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\": rpc error: code = NotFound desc = could not find container \"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\": container with ID starting with eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798 not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.328105 4997 scope.go:117] "RemoveContainer" containerID="a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.328381 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2"} err="failed to get container status \"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\": rpc error: code = NotFound desc = could not find container \"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\": container with ID starting with a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2 not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.328402 4997 scope.go:117] "RemoveContainer" containerID="9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.328639 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa"} err="failed to get container status \"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa\": rpc error: code = NotFound desc = could not find container \"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa\": container with ID starting with 9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.328662 4997 scope.go:117] "RemoveContainer" containerID="efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.328884 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57"} err="failed to get container status \"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57\": rpc error: code = NotFound desc = could not find container \"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57\": container with ID starting with efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57 not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.328903 4997 scope.go:117] "RemoveContainer" containerID="742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.329114 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c"} err="failed to get container status \"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\": rpc error: code = NotFound desc = could not find container \"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\": container with ID starting with 742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.329133 4997 scope.go:117] "RemoveContainer" containerID="8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.329316 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b"} err="failed to get container status \"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\": rpc error: code = NotFound desc = could not find container \"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\": container with ID starting with 8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.329333 4997 scope.go:117] "RemoveContainer" containerID="ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.329521 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34"} err="failed to get container status \"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\": rpc error: code = NotFound desc = could not find container \"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\": container with ID starting with ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34 not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.329537 4997 scope.go:117] "RemoveContainer" containerID="efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.329737 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb"} err="failed to get container status \"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\": rpc error: code = NotFound desc = could not find container \"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\": container with ID starting with efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.329753 4997 scope.go:117] "RemoveContainer" containerID="46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.329956 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf"} err="failed to get container status \"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\": rpc error: code = NotFound desc = could not find container \"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\": container with ID starting with 46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.329972 4997 scope.go:117] "RemoveContainer" containerID="cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.330153 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c"} err="failed to get container status \"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\": rpc error: code = NotFound desc = could not find container \"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\": container with ID starting with cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.330169 4997 scope.go:117] "RemoveContainer" containerID="eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.330338 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798"} err="failed to get container status \"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\": rpc error: code = NotFound desc = could not find container \"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\": container with ID starting with eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798 not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.330462 4997 scope.go:117] "RemoveContainer" containerID="a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.330697 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2"} err="failed to get container status \"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\": rpc error: code = NotFound desc = could not find container \"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\": container with ID starting with a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2 not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.330720 4997 scope.go:117] "RemoveContainer" containerID="9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.330958 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa"} err="failed to get container status \"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa\": rpc error: code = NotFound desc = could not find container \"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa\": container with ID starting with 9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.330973 4997 scope.go:117] "RemoveContainer" containerID="efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.331154 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57"} err="failed to get container status \"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57\": rpc error: code = NotFound desc = could not find container \"efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57\": container with ID starting with efb3f4f06cc18a13cfbd405592c799b89af7922328598c7ec28151b391791e57 not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.331170 4997 scope.go:117] "RemoveContainer" containerID="742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.331339 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c"} err="failed to get container status \"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\": rpc error: code = NotFound desc = could not find container \"742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c\": container with ID starting with 742a550c94baab17be9aa076987ab14b38639816654be9da34078e3179086d9c not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.331354 4997 scope.go:117] "RemoveContainer" containerID="8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.331528 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b"} err="failed to get container status \"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\": rpc error: code = NotFound desc = could not find container \"8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b\": container with ID starting with 8cd15b13efe85ec954de866139196a099dd096589de4f8324be9ccfbcc64418b not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.331543 4997 scope.go:117] "RemoveContainer" containerID="ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.331971 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34"} err="failed to get container status \"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\": rpc error: code = NotFound desc = could not find container \"ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34\": container with ID starting with ad46256d1f2359567ca713c4fca1ff3ffb3ff2aeec8321d785e53c7e039b6d34 not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.331987 4997 scope.go:117] "RemoveContainer" containerID="efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.332161 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb"} err="failed to get container status \"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\": rpc error: code = NotFound desc = could not find container \"efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb\": container with ID starting with efe0060a054679b7891f26a1670fed50e6d07b8eb0cba9a95601249355593edb not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.332178 4997 scope.go:117] "RemoveContainer" containerID="46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.332357 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf"} err="failed to get container status \"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\": rpc error: code = NotFound desc = could not find container \"46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf\": container with ID starting with 46c454d3308deac6125f8ddff6cf179944e0813642d64dd37cd6c3235cde50bf not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.332373 4997 scope.go:117] "RemoveContainer" containerID="cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.332550 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c"} err="failed to get container status \"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\": rpc error: code = NotFound desc = could not find container \"cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c\": container with ID starting with cd067bf80098a8a543bde53279859c7651f5119c97fa0c42a162c62de594990c not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.332566 4997 scope.go:117] "RemoveContainer" containerID="eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.332750 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798"} err="failed to get container status \"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\": rpc error: code = NotFound desc = could not find container \"eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798\": container with ID starting with eaca76ffc2f725c723786c4645837d0924054800746c017e41862cbc91dd3798 not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.332767 4997 scope.go:117] "RemoveContainer" containerID="a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.332964 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2"} err="failed to get container status \"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\": rpc error: code = NotFound desc = could not find container \"a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2\": container with ID starting with a0946924143a456c861289182abce84491eafa91ab16870032e78367896935a2 not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.332982 4997 scope.go:117] "RemoveContainer" containerID="9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.333182 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa"} err="failed to get container status \"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa\": rpc error: code = NotFound desc = could not find container \"9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa\": container with ID starting with 9acd8569246906b992c15763ba9a644fa92c3090f3be37ff645968406f4e94fa not found: ID does not exist" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.389105 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-ovnkube-script-lib\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.389219 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-log-socket\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.389297 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-slash\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.389363 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-ovnkube-config\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.389421 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-slash\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.389382 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-log-socket\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.389524 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-node-log\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.389593 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-env-overrides\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.389690 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4tvh\" (UniqueName: \"kubernetes.io/projected/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-kube-api-access-c4tvh\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.389773 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-run-systemd\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.389838 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-kubelet\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.389905 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.389984 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-cni-netd\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390057 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-ovn-node-metrics-cert\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390125 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-run-ovn\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390194 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-systemd-units\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390260 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-run-systemd\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390233 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390265 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-run-ovn-kubernetes\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390598 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-cni-bin\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.389593 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-node-log\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390665 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-etc-openvswitch\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390698 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-etc-openvswitch\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390742 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-kubelet\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390745 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-var-lib-openvswitch\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390776 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-cni-bin\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390778 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-run-openvswitch\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390806 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-run-openvswitch\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390846 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-var-lib-openvswitch\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390859 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-run-netns\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390893 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-run-ovn\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390927 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-cni-netd\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.391042 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-run-netns\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.391075 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-systemd-units\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.390203 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-env-overrides\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.391331 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-host-run-ovn-kubernetes\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.392636 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-ovnkube-config\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.393085 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-ovnkube-script-lib\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.395869 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-ovn-node-metrics-cert\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.413181 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4tvh\" (UniqueName: \"kubernetes.io/projected/4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6-kube-api-access-c4tvh\") pod \"ovnkube-node-87ct9\" (UID: \"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.417656 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bchpk"] Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.422577 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bchpk"] Dec 05 07:10:42 crc kubenswrapper[4997]: I1205 07:10:42.480852 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:42 crc kubenswrapper[4997]: W1205 07:10:42.507845 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4fef4ba1_6bc1_457f_8ff9_ceaf50b5f7e6.slice/crio-3db3a2717bceb4c1baa8a93db8face6069da0d00b765414dc3ea8eca64bde4da WatchSource:0}: Error finding container 3db3a2717bceb4c1baa8a93db8face6069da0d00b765414dc3ea8eca64bde4da: Status 404 returned error can't find the container with id 3db3a2717bceb4c1baa8a93db8face6069da0d00b765414dc3ea8eca64bde4da Dec 05 07:10:43 crc kubenswrapper[4997]: I1205 07:10:43.090214 4997 generic.go:334] "Generic (PLEG): container finished" podID="4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6" containerID="3d644591391717e171eb23a91bc96fa78b9ae86fa1b93e1c68296ae2ecf78816" exitCode=0 Dec 05 07:10:43 crc kubenswrapper[4997]: I1205 07:10:43.090317 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" event={"ID":"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6","Type":"ContainerDied","Data":"3d644591391717e171eb23a91bc96fa78b9ae86fa1b93e1c68296ae2ecf78816"} Dec 05 07:10:43 crc kubenswrapper[4997]: I1205 07:10:43.090607 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" event={"ID":"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6","Type":"ContainerStarted","Data":"3db3a2717bceb4c1baa8a93db8face6069da0d00b765414dc3ea8eca64bde4da"} Dec 05 07:10:43 crc kubenswrapper[4997]: I1205 07:10:43.096315 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-th9dq_85f05eb3-df7f-4c46-a0e2-451766bb3297/kube-multus/2.log" Dec 05 07:10:43 crc kubenswrapper[4997]: I1205 07:10:43.096839 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-th9dq_85f05eb3-df7f-4c46-a0e2-451766bb3297/kube-multus/1.log" Dec 05 07:10:43 crc kubenswrapper[4997]: I1205 07:10:43.096876 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-th9dq" event={"ID":"85f05eb3-df7f-4c46-a0e2-451766bb3297","Type":"ContainerStarted","Data":"5c5906f3fa22151f2ccf971c62cf38ae0a1fdb0eafbae5bd7b5c53d51fa5580b"} Dec 05 07:10:43 crc kubenswrapper[4997]: I1205 07:10:43.756934 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28f65c70-f59b-4393-b9c9-8163bb596f6c" path="/var/lib/kubelet/pods/28f65c70-f59b-4393-b9c9-8163bb596f6c/volumes" Dec 05 07:10:44 crc kubenswrapper[4997]: I1205 07:10:44.106660 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" event={"ID":"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6","Type":"ContainerStarted","Data":"435946d1d3d7580a52a2353ca22c1f8c3e2abff17efbacadab5a5f4dd1f7b8d6"} Dec 05 07:10:44 crc kubenswrapper[4997]: I1205 07:10:44.106733 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" event={"ID":"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6","Type":"ContainerStarted","Data":"6a6a865f81f39e871b6bb0ada0a0fc8f6aa0296f0e872bbb0f13bc83a66e4870"} Dec 05 07:10:44 crc kubenswrapper[4997]: I1205 07:10:44.106749 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" event={"ID":"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6","Type":"ContainerStarted","Data":"5e8f52939c00f1e290309e7988eaf80559e2402447b88e86721160e812c6485a"} Dec 05 07:10:44 crc kubenswrapper[4997]: I1205 07:10:44.106760 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" event={"ID":"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6","Type":"ContainerStarted","Data":"3ef9f8894a90bc5ec8417128eb33742dad74079b26e811a5ee213774f3dbe0ad"} Dec 05 07:10:44 crc kubenswrapper[4997]: I1205 07:10:44.106769 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" event={"ID":"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6","Type":"ContainerStarted","Data":"41cfb0849d54e5e291207bc004632e91c222cb61d9a1d188dba610c8734f337a"} Dec 05 07:10:44 crc kubenswrapper[4997]: I1205 07:10:44.106778 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" event={"ID":"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6","Type":"ContainerStarted","Data":"e352263c2821c535fd0066aee49b24985d215a36eb23951d1bc686c8b0e211a3"} Dec 05 07:10:44 crc kubenswrapper[4997]: I1205 07:10:44.761687 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-clw95"] Dec 05 07:10:44 crc kubenswrapper[4997]: I1205 07:10:44.763044 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:10:44 crc kubenswrapper[4997]: I1205 07:10:44.828768 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxm7t\" (UniqueName: \"kubernetes.io/projected/b52f20c9-c735-454a-b086-886643ea82b0-kube-api-access-bxm7t\") pod \"redhat-operators-clw95\" (UID: \"b52f20c9-c735-454a-b086-886643ea82b0\") " pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:10:44 crc kubenswrapper[4997]: I1205 07:10:44.828817 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b52f20c9-c735-454a-b086-886643ea82b0-catalog-content\") pod \"redhat-operators-clw95\" (UID: \"b52f20c9-c735-454a-b086-886643ea82b0\") " pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:10:44 crc kubenswrapper[4997]: I1205 07:10:44.828851 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b52f20c9-c735-454a-b086-886643ea82b0-utilities\") pod \"redhat-operators-clw95\" (UID: \"b52f20c9-c735-454a-b086-886643ea82b0\") " pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:10:44 crc kubenswrapper[4997]: I1205 07:10:44.930331 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxm7t\" (UniqueName: \"kubernetes.io/projected/b52f20c9-c735-454a-b086-886643ea82b0-kube-api-access-bxm7t\") pod \"redhat-operators-clw95\" (UID: \"b52f20c9-c735-454a-b086-886643ea82b0\") " pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:10:44 crc kubenswrapper[4997]: I1205 07:10:44.930900 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b52f20c9-c735-454a-b086-886643ea82b0-catalog-content\") pod \"redhat-operators-clw95\" (UID: \"b52f20c9-c735-454a-b086-886643ea82b0\") " pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:10:44 crc kubenswrapper[4997]: I1205 07:10:44.931117 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b52f20c9-c735-454a-b086-886643ea82b0-utilities\") pod \"redhat-operators-clw95\" (UID: \"b52f20c9-c735-454a-b086-886643ea82b0\") " pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:10:44 crc kubenswrapper[4997]: I1205 07:10:44.931412 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b52f20c9-c735-454a-b086-886643ea82b0-catalog-content\") pod \"redhat-operators-clw95\" (UID: \"b52f20c9-c735-454a-b086-886643ea82b0\") " pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:10:44 crc kubenswrapper[4997]: I1205 07:10:44.931653 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b52f20c9-c735-454a-b086-886643ea82b0-utilities\") pod \"redhat-operators-clw95\" (UID: \"b52f20c9-c735-454a-b086-886643ea82b0\") " pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:10:44 crc kubenswrapper[4997]: I1205 07:10:44.953012 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxm7t\" (UniqueName: \"kubernetes.io/projected/b52f20c9-c735-454a-b086-886643ea82b0-kube-api-access-bxm7t\") pod \"redhat-operators-clw95\" (UID: \"b52f20c9-c735-454a-b086-886643ea82b0\") " pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:10:45 crc kubenswrapper[4997]: I1205 07:10:45.092887 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:10:45 crc kubenswrapper[4997]: E1205 07:10:45.116242 4997 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-operators-clw95_openshift-marketplace_b52f20c9-c735-454a-b086-886643ea82b0_0(392c68fb2efc12fd90b96f1333a262151ccd161ffecb5bcab53ad589cefedab9): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 07:10:45 crc kubenswrapper[4997]: E1205 07:10:45.116320 4997 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-operators-clw95_openshift-marketplace_b52f20c9-c735-454a-b086-886643ea82b0_0(392c68fb2efc12fd90b96f1333a262151ccd161ffecb5bcab53ad589cefedab9): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:10:45 crc kubenswrapper[4997]: E1205 07:10:45.116359 4997 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-operators-clw95_openshift-marketplace_b52f20c9-c735-454a-b086-886643ea82b0_0(392c68fb2efc12fd90b96f1333a262151ccd161ffecb5bcab53ad589cefedab9): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:10:45 crc kubenswrapper[4997]: E1205 07:10:45.116407 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"redhat-operators-clw95_openshift-marketplace(b52f20c9-c735-454a-b086-886643ea82b0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"redhat-operators-clw95_openshift-marketplace(b52f20c9-c735-454a-b086-886643ea82b0)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-operators-clw95_openshift-marketplace_b52f20c9-c735-454a-b086-886643ea82b0_0(392c68fb2efc12fd90b96f1333a262151ccd161ffecb5bcab53ad589cefedab9): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/redhat-operators-clw95" podUID="b52f20c9-c735-454a-b086-886643ea82b0" Dec 05 07:10:46 crc kubenswrapper[4997]: I1205 07:10:46.119405 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" event={"ID":"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6","Type":"ContainerStarted","Data":"d84129f92bc9ed24fce5ccd638f5f5719f4b6690dabbd5f84f3e0b3883d2f8e3"} Dec 05 07:10:49 crc kubenswrapper[4997]: I1205 07:10:49.144429 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" event={"ID":"4fef4ba1-6bc1-457f-8ff9-ceaf50b5f7e6","Type":"ContainerStarted","Data":"8311d0cdd7a67de43a227e3facaf219e2320f9dff1715eeacda7e733d3751fa0"} Dec 05 07:10:49 crc kubenswrapper[4997]: I1205 07:10:49.145037 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:49 crc kubenswrapper[4997]: I1205 07:10:49.175358 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" podStartSLOduration=7.175338147 podStartE2EDuration="7.175338147s" podCreationTimestamp="2025-12-05 07:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:10:49.171941518 +0000 UTC m=+949.700848789" watchObservedRunningTime="2025-12-05 07:10:49.175338147 +0000 UTC m=+949.704245408" Dec 05 07:10:49 crc kubenswrapper[4997]: I1205 07:10:49.180151 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:50 crc kubenswrapper[4997]: I1205 07:10:50.149677 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:50 crc kubenswrapper[4997]: I1205 07:10:50.150730 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:50 crc kubenswrapper[4997]: I1205 07:10:50.171307 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:10:50 crc kubenswrapper[4997]: I1205 07:10:50.938661 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-clw95"] Dec 05 07:10:50 crc kubenswrapper[4997]: I1205 07:10:50.938831 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:10:50 crc kubenswrapper[4997]: I1205 07:10:50.939342 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:10:50 crc kubenswrapper[4997]: E1205 07:10:50.969062 4997 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-operators-clw95_openshift-marketplace_b52f20c9-c735-454a-b086-886643ea82b0_0(3e49ae003489b713f5a0bc1d28f1db1fa59ec0559eb6b10312301c330ecf8e37): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 07:10:50 crc kubenswrapper[4997]: E1205 07:10:50.969173 4997 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-operators-clw95_openshift-marketplace_b52f20c9-c735-454a-b086-886643ea82b0_0(3e49ae003489b713f5a0bc1d28f1db1fa59ec0559eb6b10312301c330ecf8e37): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:10:50 crc kubenswrapper[4997]: E1205 07:10:50.969203 4997 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-operators-clw95_openshift-marketplace_b52f20c9-c735-454a-b086-886643ea82b0_0(3e49ae003489b713f5a0bc1d28f1db1fa59ec0559eb6b10312301c330ecf8e37): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:10:50 crc kubenswrapper[4997]: E1205 07:10:50.969282 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"redhat-operators-clw95_openshift-marketplace(b52f20c9-c735-454a-b086-886643ea82b0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"redhat-operators-clw95_openshift-marketplace(b52f20c9-c735-454a-b086-886643ea82b0)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-operators-clw95_openshift-marketplace_b52f20c9-c735-454a-b086-886643ea82b0_0(3e49ae003489b713f5a0bc1d28f1db1fa59ec0559eb6b10312301c330ecf8e37): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/redhat-operators-clw95" podUID="b52f20c9-c735-454a-b086-886643ea82b0" Dec 05 07:10:51 crc kubenswrapper[4997]: I1205 07:10:51.392094 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-5qdxt"] Dec 05 07:10:51 crc kubenswrapper[4997]: I1205 07:10:51.393415 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:10:51 crc kubenswrapper[4997]: I1205 07:10:51.395829 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 05 07:10:51 crc kubenswrapper[4997]: I1205 07:10:51.396102 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 05 07:10:51 crc kubenswrapper[4997]: I1205 07:10:51.397438 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 05 07:10:51 crc kubenswrapper[4997]: I1205 07:10:51.397443 4997 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-dtsv6" Dec 05 07:10:51 crc kubenswrapper[4997]: I1205 07:10:51.399293 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-5qdxt"] Dec 05 07:10:51 crc kubenswrapper[4997]: I1205 07:10:51.516415 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtj26\" (UniqueName: \"kubernetes.io/projected/bf6e1555-1042-4e81-b239-1289b9ed77cd-kube-api-access-dtj26\") pod \"crc-storage-crc-5qdxt\" (UID: \"bf6e1555-1042-4e81-b239-1289b9ed77cd\") " pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:10:51 crc kubenswrapper[4997]: I1205 07:10:51.516506 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/bf6e1555-1042-4e81-b239-1289b9ed77cd-node-mnt\") pod \"crc-storage-crc-5qdxt\" (UID: \"bf6e1555-1042-4e81-b239-1289b9ed77cd\") " pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:10:51 crc kubenswrapper[4997]: I1205 07:10:51.516548 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/bf6e1555-1042-4e81-b239-1289b9ed77cd-crc-storage\") pod \"crc-storage-crc-5qdxt\" (UID: \"bf6e1555-1042-4e81-b239-1289b9ed77cd\") " pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:10:51 crc kubenswrapper[4997]: I1205 07:10:51.617847 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/bf6e1555-1042-4e81-b239-1289b9ed77cd-crc-storage\") pod \"crc-storage-crc-5qdxt\" (UID: \"bf6e1555-1042-4e81-b239-1289b9ed77cd\") " pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:10:51 crc kubenswrapper[4997]: I1205 07:10:51.618308 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtj26\" (UniqueName: \"kubernetes.io/projected/bf6e1555-1042-4e81-b239-1289b9ed77cd-kube-api-access-dtj26\") pod \"crc-storage-crc-5qdxt\" (UID: \"bf6e1555-1042-4e81-b239-1289b9ed77cd\") " pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:10:51 crc kubenswrapper[4997]: I1205 07:10:51.618456 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/bf6e1555-1042-4e81-b239-1289b9ed77cd-node-mnt\") pod \"crc-storage-crc-5qdxt\" (UID: \"bf6e1555-1042-4e81-b239-1289b9ed77cd\") " pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:10:51 crc kubenswrapper[4997]: I1205 07:10:51.618810 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/bf6e1555-1042-4e81-b239-1289b9ed77cd-node-mnt\") pod \"crc-storage-crc-5qdxt\" (UID: \"bf6e1555-1042-4e81-b239-1289b9ed77cd\") " pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:10:51 crc kubenswrapper[4997]: I1205 07:10:51.618904 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/bf6e1555-1042-4e81-b239-1289b9ed77cd-crc-storage\") pod \"crc-storage-crc-5qdxt\" (UID: \"bf6e1555-1042-4e81-b239-1289b9ed77cd\") " pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:10:51 crc kubenswrapper[4997]: I1205 07:10:51.644638 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtj26\" (UniqueName: \"kubernetes.io/projected/bf6e1555-1042-4e81-b239-1289b9ed77cd-kube-api-access-dtj26\") pod \"crc-storage-crc-5qdxt\" (UID: \"bf6e1555-1042-4e81-b239-1289b9ed77cd\") " pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:10:51 crc kubenswrapper[4997]: I1205 07:10:51.721895 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:10:51 crc kubenswrapper[4997]: E1205 07:10:51.744182 4997 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5qdxt_crc-storage_bf6e1555-1042-4e81-b239-1289b9ed77cd_0(b601fbb645191e638ac7678289a9920be727e996435cba637d72b9fe0e8c2f07): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 07:10:51 crc kubenswrapper[4997]: E1205 07:10:51.744281 4997 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5qdxt_crc-storage_bf6e1555-1042-4e81-b239-1289b9ed77cd_0(b601fbb645191e638ac7678289a9920be727e996435cba637d72b9fe0e8c2f07): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:10:51 crc kubenswrapper[4997]: E1205 07:10:51.744305 4997 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5qdxt_crc-storage_bf6e1555-1042-4e81-b239-1289b9ed77cd_0(b601fbb645191e638ac7678289a9920be727e996435cba637d72b9fe0e8c2f07): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:10:51 crc kubenswrapper[4997]: E1205 07:10:51.744354 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-5qdxt_crc-storage(bf6e1555-1042-4e81-b239-1289b9ed77cd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-5qdxt_crc-storage(bf6e1555-1042-4e81-b239-1289b9ed77cd)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5qdxt_crc-storage_bf6e1555-1042-4e81-b239-1289b9ed77cd_0(b601fbb645191e638ac7678289a9920be727e996435cba637d72b9fe0e8c2f07): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-5qdxt" podUID="bf6e1555-1042-4e81-b239-1289b9ed77cd" Dec 05 07:10:52 crc kubenswrapper[4997]: I1205 07:10:52.158120 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:10:52 crc kubenswrapper[4997]: I1205 07:10:52.158732 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:10:52 crc kubenswrapper[4997]: E1205 07:10:52.181343 4997 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5qdxt_crc-storage_bf6e1555-1042-4e81-b239-1289b9ed77cd_0(424574cbdb9a29210e0eb190817cfee73d9ebeafea7aaa51af82f67f7b50e143): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 07:10:52 crc kubenswrapper[4997]: E1205 07:10:52.181514 4997 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5qdxt_crc-storage_bf6e1555-1042-4e81-b239-1289b9ed77cd_0(424574cbdb9a29210e0eb190817cfee73d9ebeafea7aaa51af82f67f7b50e143): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:10:52 crc kubenswrapper[4997]: E1205 07:10:52.181599 4997 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5qdxt_crc-storage_bf6e1555-1042-4e81-b239-1289b9ed77cd_0(424574cbdb9a29210e0eb190817cfee73d9ebeafea7aaa51af82f67f7b50e143): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:10:52 crc kubenswrapper[4997]: E1205 07:10:52.181755 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-5qdxt_crc-storage(bf6e1555-1042-4e81-b239-1289b9ed77cd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-5qdxt_crc-storage(bf6e1555-1042-4e81-b239-1289b9ed77cd)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5qdxt_crc-storage_bf6e1555-1042-4e81-b239-1289b9ed77cd_0(424574cbdb9a29210e0eb190817cfee73d9ebeafea7aaa51af82f67f7b50e143): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-5qdxt" podUID="bf6e1555-1042-4e81-b239-1289b9ed77cd" Dec 05 07:11:00 crc kubenswrapper[4997]: I1205 07:11:00.064573 4997 scope.go:117] "RemoveContainer" containerID="ba8817aab2ae020cc998813edb099030cc3c3c6eb11e8f175e3cfd82b18686b2" Dec 05 07:11:00 crc kubenswrapper[4997]: I1205 07:11:00.196602 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-th9dq_85f05eb3-df7f-4c46-a0e2-451766bb3297/kube-multus/2.log" Dec 05 07:11:01 crc kubenswrapper[4997]: I1205 07:11:01.748847 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:11:01 crc kubenswrapper[4997]: I1205 07:11:01.750174 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:11:02 crc kubenswrapper[4997]: I1205 07:11:02.000201 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-clw95"] Dec 05 07:11:02 crc kubenswrapper[4997]: W1205 07:11:02.013605 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb52f20c9_c735_454a_b086_886643ea82b0.slice/crio-9dd4fe4fd7ae0d94649d6a0020228b489c93bc0370717d85b11695a859f73b3a WatchSource:0}: Error finding container 9dd4fe4fd7ae0d94649d6a0020228b489c93bc0370717d85b11695a859f73b3a: Status 404 returned error can't find the container with id 9dd4fe4fd7ae0d94649d6a0020228b489c93bc0370717d85b11695a859f73b3a Dec 05 07:11:02 crc kubenswrapper[4997]: I1205 07:11:02.210891 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clw95" event={"ID":"b52f20c9-c735-454a-b086-886643ea82b0","Type":"ContainerStarted","Data":"626619ca4a05db089daf721167c798167c165504369a9ff207deaeba1735b369"} Dec 05 07:11:02 crc kubenswrapper[4997]: I1205 07:11:02.211446 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clw95" event={"ID":"b52f20c9-c735-454a-b086-886643ea82b0","Type":"ContainerStarted","Data":"9dd4fe4fd7ae0d94649d6a0020228b489c93bc0370717d85b11695a859f73b3a"} Dec 05 07:11:02 crc kubenswrapper[4997]: I1205 07:11:02.748372 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:11:02 crc kubenswrapper[4997]: I1205 07:11:02.749441 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:11:02 crc kubenswrapper[4997]: I1205 07:11:02.957061 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-5qdxt"] Dec 05 07:11:03 crc kubenswrapper[4997]: I1205 07:11:03.222482 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-82mpd"] Dec 05 07:11:03 crc kubenswrapper[4997]: I1205 07:11:03.224343 4997 generic.go:334] "Generic (PLEG): container finished" podID="b52f20c9-c735-454a-b086-886643ea82b0" containerID="626619ca4a05db089daf721167c798167c165504369a9ff207deaeba1735b369" exitCode=0 Dec 05 07:11:03 crc kubenswrapper[4997]: I1205 07:11:03.228661 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-5qdxt" event={"ID":"bf6e1555-1042-4e81-b239-1289b9ed77cd","Type":"ContainerStarted","Data":"8f760f280228fcd23f024b70187fc342d68e520ab5b35595a563f0cb0a89c3d3"} Dec 05 07:11:03 crc kubenswrapper[4997]: I1205 07:11:03.228726 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clw95" event={"ID":"b52f20c9-c735-454a-b086-886643ea82b0","Type":"ContainerDied","Data":"626619ca4a05db089daf721167c798167c165504369a9ff207deaeba1735b369"} Dec 05 07:11:03 crc kubenswrapper[4997]: I1205 07:11:03.228843 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-82mpd" Dec 05 07:11:03 crc kubenswrapper[4997]: I1205 07:11:03.230879 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-82mpd"] Dec 05 07:11:03 crc kubenswrapper[4997]: I1205 07:11:03.278210 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8kw9\" (UniqueName: \"kubernetes.io/projected/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-kube-api-access-l8kw9\") pod \"community-operators-82mpd\" (UID: \"144aba01-6e9c-4c56-b7ec-e11dd2855bf3\") " pod="openshift-marketplace/community-operators-82mpd" Dec 05 07:11:03 crc kubenswrapper[4997]: I1205 07:11:03.278260 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-catalog-content\") pod \"community-operators-82mpd\" (UID: \"144aba01-6e9c-4c56-b7ec-e11dd2855bf3\") " pod="openshift-marketplace/community-operators-82mpd" Dec 05 07:11:03 crc kubenswrapper[4997]: I1205 07:11:03.278339 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-utilities\") pod \"community-operators-82mpd\" (UID: \"144aba01-6e9c-4c56-b7ec-e11dd2855bf3\") " pod="openshift-marketplace/community-operators-82mpd" Dec 05 07:11:03 crc kubenswrapper[4997]: I1205 07:11:03.380180 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-utilities\") pod \"community-operators-82mpd\" (UID: \"144aba01-6e9c-4c56-b7ec-e11dd2855bf3\") " pod="openshift-marketplace/community-operators-82mpd" Dec 05 07:11:03 crc kubenswrapper[4997]: I1205 07:11:03.380296 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8kw9\" (UniqueName: \"kubernetes.io/projected/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-kube-api-access-l8kw9\") pod \"community-operators-82mpd\" (UID: \"144aba01-6e9c-4c56-b7ec-e11dd2855bf3\") " pod="openshift-marketplace/community-operators-82mpd" Dec 05 07:11:03 crc kubenswrapper[4997]: I1205 07:11:03.380323 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-catalog-content\") pod \"community-operators-82mpd\" (UID: \"144aba01-6e9c-4c56-b7ec-e11dd2855bf3\") " pod="openshift-marketplace/community-operators-82mpd" Dec 05 07:11:03 crc kubenswrapper[4997]: I1205 07:11:03.380782 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-utilities\") pod \"community-operators-82mpd\" (UID: \"144aba01-6e9c-4c56-b7ec-e11dd2855bf3\") " pod="openshift-marketplace/community-operators-82mpd" Dec 05 07:11:03 crc kubenswrapper[4997]: I1205 07:11:03.380811 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-catalog-content\") pod \"community-operators-82mpd\" (UID: \"144aba01-6e9c-4c56-b7ec-e11dd2855bf3\") " pod="openshift-marketplace/community-operators-82mpd" Dec 05 07:11:03 crc kubenswrapper[4997]: I1205 07:11:03.402372 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8kw9\" (UniqueName: \"kubernetes.io/projected/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-kube-api-access-l8kw9\") pod \"community-operators-82mpd\" (UID: \"144aba01-6e9c-4c56-b7ec-e11dd2855bf3\") " pod="openshift-marketplace/community-operators-82mpd" Dec 05 07:11:03 crc kubenswrapper[4997]: I1205 07:11:03.568272 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-82mpd" Dec 05 07:11:03 crc kubenswrapper[4997]: I1205 07:11:03.890471 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-82mpd"] Dec 05 07:11:04 crc kubenswrapper[4997]: I1205 07:11:04.238508 4997 generic.go:334] "Generic (PLEG): container finished" podID="144aba01-6e9c-4c56-b7ec-e11dd2855bf3" containerID="4c083a32c89ad64b4ad5b027fe8cac8a4ba0b56748e8c76949a3e9c849ba171d" exitCode=0 Dec 05 07:11:04 crc kubenswrapper[4997]: I1205 07:11:04.239201 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82mpd" event={"ID":"144aba01-6e9c-4c56-b7ec-e11dd2855bf3","Type":"ContainerDied","Data":"4c083a32c89ad64b4ad5b027fe8cac8a4ba0b56748e8c76949a3e9c849ba171d"} Dec 05 07:11:04 crc kubenswrapper[4997]: I1205 07:11:04.239248 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82mpd" event={"ID":"144aba01-6e9c-4c56-b7ec-e11dd2855bf3","Type":"ContainerStarted","Data":"5c9ab07012f4393e86bbd60652cfd48c72dfe8acadd325fa53bc7efdb8b6e78f"} Dec 05 07:11:04 crc kubenswrapper[4997]: I1205 07:11:04.248953 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clw95" event={"ID":"b52f20c9-c735-454a-b086-886643ea82b0","Type":"ContainerStarted","Data":"456baa907c02760c7fb5a60ffa907123a94f8c1ee597b2b46943f1a3d64e4444"} Dec 05 07:11:05 crc kubenswrapper[4997]: I1205 07:11:05.267271 4997 generic.go:334] "Generic (PLEG): container finished" podID="b52f20c9-c735-454a-b086-886643ea82b0" containerID="456baa907c02760c7fb5a60ffa907123a94f8c1ee597b2b46943f1a3d64e4444" exitCode=0 Dec 05 07:11:05 crc kubenswrapper[4997]: I1205 07:11:05.267363 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clw95" event={"ID":"b52f20c9-c735-454a-b086-886643ea82b0","Type":"ContainerDied","Data":"456baa907c02760c7fb5a60ffa907123a94f8c1ee597b2b46943f1a3d64e4444"} Dec 05 07:11:05 crc kubenswrapper[4997]: I1205 07:11:05.270369 4997 generic.go:334] "Generic (PLEG): container finished" podID="bf6e1555-1042-4e81-b239-1289b9ed77cd" containerID="b609fe619a1e20f6758f3c4f05f2e54a4c673bbf3098d4459662bf0c7cd96193" exitCode=0 Dec 05 07:11:05 crc kubenswrapper[4997]: I1205 07:11:05.270451 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-5qdxt" event={"ID":"bf6e1555-1042-4e81-b239-1289b9ed77cd","Type":"ContainerDied","Data":"b609fe619a1e20f6758f3c4f05f2e54a4c673bbf3098d4459662bf0c7cd96193"} Dec 05 07:11:05 crc kubenswrapper[4997]: I1205 07:11:05.294743 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82mpd" event={"ID":"144aba01-6e9c-4c56-b7ec-e11dd2855bf3","Type":"ContainerStarted","Data":"fa2dd55190a2b5b0ae83603049210c478c36a853b896f8e7124df3b3276191ab"} Dec 05 07:11:06 crc kubenswrapper[4997]: I1205 07:11:06.301667 4997 generic.go:334] "Generic (PLEG): container finished" podID="144aba01-6e9c-4c56-b7ec-e11dd2855bf3" containerID="fa2dd55190a2b5b0ae83603049210c478c36a853b896f8e7124df3b3276191ab" exitCode=0 Dec 05 07:11:06 crc kubenswrapper[4997]: I1205 07:11:06.301773 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82mpd" event={"ID":"144aba01-6e9c-4c56-b7ec-e11dd2855bf3","Type":"ContainerDied","Data":"fa2dd55190a2b5b0ae83603049210c478c36a853b896f8e7124df3b3276191ab"} Dec 05 07:11:06 crc kubenswrapper[4997]: I1205 07:11:06.305114 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clw95" event={"ID":"b52f20c9-c735-454a-b086-886643ea82b0","Type":"ContainerStarted","Data":"81209b95927209dd7b101b913adcc0bc56961e816833cd8f9acc054752bcd207"} Dec 05 07:11:06 crc kubenswrapper[4997]: I1205 07:11:06.955560 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:11:06 crc kubenswrapper[4997]: I1205 07:11:06.971845 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-clw95" podStartSLOduration=20.415703094 podStartE2EDuration="22.971823109s" podCreationTimestamp="2025-12-05 07:10:44 +0000 UTC" firstStartedPulling="2025-12-05 07:11:03.233208762 +0000 UTC m=+963.762116023" lastFinishedPulling="2025-12-05 07:11:05.789328777 +0000 UTC m=+966.318236038" observedRunningTime="2025-12-05 07:11:06.352749354 +0000 UTC m=+966.881656615" watchObservedRunningTime="2025-12-05 07:11:06.971823109 +0000 UTC m=+967.500730370" Dec 05 07:11:07 crc kubenswrapper[4997]: I1205 07:11:07.129039 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/bf6e1555-1042-4e81-b239-1289b9ed77cd-crc-storage\") pod \"bf6e1555-1042-4e81-b239-1289b9ed77cd\" (UID: \"bf6e1555-1042-4e81-b239-1289b9ed77cd\") " Dec 05 07:11:07 crc kubenswrapper[4997]: I1205 07:11:07.130021 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf6e1555-1042-4e81-b239-1289b9ed77cd-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "bf6e1555-1042-4e81-b239-1289b9ed77cd" (UID: "bf6e1555-1042-4e81-b239-1289b9ed77cd"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:11:07 crc kubenswrapper[4997]: I1205 07:11:07.130058 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/bf6e1555-1042-4e81-b239-1289b9ed77cd-node-mnt\") pod \"bf6e1555-1042-4e81-b239-1289b9ed77cd\" (UID: \"bf6e1555-1042-4e81-b239-1289b9ed77cd\") " Dec 05 07:11:07 crc kubenswrapper[4997]: I1205 07:11:07.130595 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtj26\" (UniqueName: \"kubernetes.io/projected/bf6e1555-1042-4e81-b239-1289b9ed77cd-kube-api-access-dtj26\") pod \"bf6e1555-1042-4e81-b239-1289b9ed77cd\" (UID: \"bf6e1555-1042-4e81-b239-1289b9ed77cd\") " Dec 05 07:11:07 crc kubenswrapper[4997]: I1205 07:11:07.130828 4997 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/bf6e1555-1042-4e81-b239-1289b9ed77cd-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 05 07:11:07 crc kubenswrapper[4997]: I1205 07:11:07.134798 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf6e1555-1042-4e81-b239-1289b9ed77cd-kube-api-access-dtj26" (OuterVolumeSpecName: "kube-api-access-dtj26") pod "bf6e1555-1042-4e81-b239-1289b9ed77cd" (UID: "bf6e1555-1042-4e81-b239-1289b9ed77cd"). InnerVolumeSpecName "kube-api-access-dtj26". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:11:07 crc kubenswrapper[4997]: I1205 07:11:07.143397 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf6e1555-1042-4e81-b239-1289b9ed77cd-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "bf6e1555-1042-4e81-b239-1289b9ed77cd" (UID: "bf6e1555-1042-4e81-b239-1289b9ed77cd"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:11:07 crc kubenswrapper[4997]: I1205 07:11:07.231808 4997 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/bf6e1555-1042-4e81-b239-1289b9ed77cd-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 05 07:11:07 crc kubenswrapper[4997]: I1205 07:11:07.232292 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtj26\" (UniqueName: \"kubernetes.io/projected/bf6e1555-1042-4e81-b239-1289b9ed77cd-kube-api-access-dtj26\") on node \"crc\" DevicePath \"\"" Dec 05 07:11:07 crc kubenswrapper[4997]: I1205 07:11:07.315381 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82mpd" event={"ID":"144aba01-6e9c-4c56-b7ec-e11dd2855bf3","Type":"ContainerStarted","Data":"48315b20317dd0c336ea1ece790aa8d0124814e195936c990f07e1c3d41809e4"} Dec 05 07:11:07 crc kubenswrapper[4997]: I1205 07:11:07.317283 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5qdxt" Dec 05 07:11:07 crc kubenswrapper[4997]: I1205 07:11:07.317308 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-5qdxt" event={"ID":"bf6e1555-1042-4e81-b239-1289b9ed77cd","Type":"ContainerDied","Data":"8f760f280228fcd23f024b70187fc342d68e520ab5b35595a563f0cb0a89c3d3"} Dec 05 07:11:07 crc kubenswrapper[4997]: I1205 07:11:07.317366 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f760f280228fcd23f024b70187fc342d68e520ab5b35595a563f0cb0a89c3d3" Dec 05 07:11:07 crc kubenswrapper[4997]: I1205 07:11:07.341341 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-82mpd" podStartSLOduration=1.915676529 podStartE2EDuration="4.341314649s" podCreationTimestamp="2025-12-05 07:11:03 +0000 UTC" firstStartedPulling="2025-12-05 07:11:04.416084793 +0000 UTC m=+964.944992054" lastFinishedPulling="2025-12-05 07:11:06.841722913 +0000 UTC m=+967.370630174" observedRunningTime="2025-12-05 07:11:07.338995359 +0000 UTC m=+967.867902630" watchObservedRunningTime="2025-12-05 07:11:07.341314649 +0000 UTC m=+967.870221910" Dec 05 07:11:12 crc kubenswrapper[4997]: I1205 07:11:12.511002 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-87ct9" Dec 05 07:11:13 crc kubenswrapper[4997]: I1205 07:11:13.568801 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-82mpd" Dec 05 07:11:13 crc kubenswrapper[4997]: I1205 07:11:13.568875 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-82mpd" Dec 05 07:11:13 crc kubenswrapper[4997]: I1205 07:11:13.615562 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-82mpd" Dec 05 07:11:14 crc kubenswrapper[4997]: I1205 07:11:14.412322 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-82mpd" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.093407 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.094237 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.237553 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.422540 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.608718 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44"] Dec 05 07:11:15 crc kubenswrapper[4997]: E1205 07:11:15.609113 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf6e1555-1042-4e81-b239-1289b9ed77cd" containerName="storage" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.609134 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf6e1555-1042-4e81-b239-1289b9ed77cd" containerName="storage" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.609319 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf6e1555-1042-4e81-b239-1289b9ed77cd" containerName="storage" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.610735 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.615048 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.617391 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44"] Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.741696 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-82mpd"] Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.790039 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6b7e3f5-ed40-4170-a421-63547bd2fb65-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44\" (UID: \"e6b7e3f5-ed40-4170-a421-63547bd2fb65\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.790355 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dl96\" (UniqueName: \"kubernetes.io/projected/e6b7e3f5-ed40-4170-a421-63547bd2fb65-kube-api-access-9dl96\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44\" (UID: \"e6b7e3f5-ed40-4170-a421-63547bd2fb65\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.790474 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6b7e3f5-ed40-4170-a421-63547bd2fb65-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44\" (UID: \"e6b7e3f5-ed40-4170-a421-63547bd2fb65\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.892320 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dl96\" (UniqueName: \"kubernetes.io/projected/e6b7e3f5-ed40-4170-a421-63547bd2fb65-kube-api-access-9dl96\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44\" (UID: \"e6b7e3f5-ed40-4170-a421-63547bd2fb65\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.892483 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6b7e3f5-ed40-4170-a421-63547bd2fb65-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44\" (UID: \"e6b7e3f5-ed40-4170-a421-63547bd2fb65\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.892541 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6b7e3f5-ed40-4170-a421-63547bd2fb65-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44\" (UID: \"e6b7e3f5-ed40-4170-a421-63547bd2fb65\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.893271 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6b7e3f5-ed40-4170-a421-63547bd2fb65-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44\" (UID: \"e6b7e3f5-ed40-4170-a421-63547bd2fb65\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.893584 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6b7e3f5-ed40-4170-a421-63547bd2fb65-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44\" (UID: \"e6b7e3f5-ed40-4170-a421-63547bd2fb65\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.920572 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dl96\" (UniqueName: \"kubernetes.io/projected/e6b7e3f5-ed40-4170-a421-63547bd2fb65-kube-api-access-9dl96\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44\" (UID: \"e6b7e3f5-ed40-4170-a421-63547bd2fb65\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" Dec 05 07:11:15 crc kubenswrapper[4997]: I1205 07:11:15.946443 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" Dec 05 07:11:16 crc kubenswrapper[4997]: I1205 07:11:16.480778 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44"] Dec 05 07:11:17 crc kubenswrapper[4997]: I1205 07:11:17.381086 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" event={"ID":"e6b7e3f5-ed40-4170-a421-63547bd2fb65","Type":"ContainerStarted","Data":"08fa07ead80532d568d477feb2f1cdfe5a342ec230f373cf611c8b34d60a0d25"} Dec 05 07:11:17 crc kubenswrapper[4997]: I1205 07:11:17.381429 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-82mpd" podUID="144aba01-6e9c-4c56-b7ec-e11dd2855bf3" containerName="registry-server" containerID="cri-o://48315b20317dd0c336ea1ece790aa8d0124814e195936c990f07e1c3d41809e4" gracePeriod=2 Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.349200 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-clw95"] Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.387958 4997 generic.go:334] "Generic (PLEG): container finished" podID="e6b7e3f5-ed40-4170-a421-63547bd2fb65" containerID="e0c37b19b28e422396f0b373bad6e8edc864de447e04c7e96656bb7b6839e33e" exitCode=0 Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.388058 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" event={"ID":"e6b7e3f5-ed40-4170-a421-63547bd2fb65","Type":"ContainerDied","Data":"e0c37b19b28e422396f0b373bad6e8edc864de447e04c7e96656bb7b6839e33e"} Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.394456 4997 generic.go:334] "Generic (PLEG): container finished" podID="144aba01-6e9c-4c56-b7ec-e11dd2855bf3" containerID="48315b20317dd0c336ea1ece790aa8d0124814e195936c990f07e1c3d41809e4" exitCode=0 Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.394992 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-clw95" podUID="b52f20c9-c735-454a-b086-886643ea82b0" containerName="registry-server" containerID="cri-o://81209b95927209dd7b101b913adcc0bc56961e816833cd8f9acc054752bcd207" gracePeriod=2 Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.394530 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82mpd" event={"ID":"144aba01-6e9c-4c56-b7ec-e11dd2855bf3","Type":"ContainerDied","Data":"48315b20317dd0c336ea1ece790aa8d0124814e195936c990f07e1c3d41809e4"} Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.395089 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82mpd" event={"ID":"144aba01-6e9c-4c56-b7ec-e11dd2855bf3","Type":"ContainerDied","Data":"5c9ab07012f4393e86bbd60652cfd48c72dfe8acadd325fa53bc7efdb8b6e78f"} Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.395105 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c9ab07012f4393e86bbd60652cfd48c72dfe8acadd325fa53bc7efdb8b6e78f" Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.442845 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-82mpd" Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.553515 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-utilities\") pod \"144aba01-6e9c-4c56-b7ec-e11dd2855bf3\" (UID: \"144aba01-6e9c-4c56-b7ec-e11dd2855bf3\") " Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.553610 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8kw9\" (UniqueName: \"kubernetes.io/projected/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-kube-api-access-l8kw9\") pod \"144aba01-6e9c-4c56-b7ec-e11dd2855bf3\" (UID: \"144aba01-6e9c-4c56-b7ec-e11dd2855bf3\") " Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.553650 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-catalog-content\") pod \"144aba01-6e9c-4c56-b7ec-e11dd2855bf3\" (UID: \"144aba01-6e9c-4c56-b7ec-e11dd2855bf3\") " Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.554527 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-utilities" (OuterVolumeSpecName: "utilities") pod "144aba01-6e9c-4c56-b7ec-e11dd2855bf3" (UID: "144aba01-6e9c-4c56-b7ec-e11dd2855bf3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.569268 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-kube-api-access-l8kw9" (OuterVolumeSpecName: "kube-api-access-l8kw9") pod "144aba01-6e9c-4c56-b7ec-e11dd2855bf3" (UID: "144aba01-6e9c-4c56-b7ec-e11dd2855bf3"). InnerVolumeSpecName "kube-api-access-l8kw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.609776 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "144aba01-6e9c-4c56-b7ec-e11dd2855bf3" (UID: "144aba01-6e9c-4c56-b7ec-e11dd2855bf3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.658556 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.658594 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8kw9\" (UniqueName: \"kubernetes.io/projected/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-kube-api-access-l8kw9\") on node \"crc\" DevicePath \"\"" Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.658605 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/144aba01-6e9c-4c56-b7ec-e11dd2855bf3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.752300 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.861724 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b52f20c9-c735-454a-b086-886643ea82b0-catalog-content\") pod \"b52f20c9-c735-454a-b086-886643ea82b0\" (UID: \"b52f20c9-c735-454a-b086-886643ea82b0\") " Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.861828 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b52f20c9-c735-454a-b086-886643ea82b0-utilities\") pod \"b52f20c9-c735-454a-b086-886643ea82b0\" (UID: \"b52f20c9-c735-454a-b086-886643ea82b0\") " Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.861869 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxm7t\" (UniqueName: \"kubernetes.io/projected/b52f20c9-c735-454a-b086-886643ea82b0-kube-api-access-bxm7t\") pod \"b52f20c9-c735-454a-b086-886643ea82b0\" (UID: \"b52f20c9-c735-454a-b086-886643ea82b0\") " Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.862924 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b52f20c9-c735-454a-b086-886643ea82b0-utilities" (OuterVolumeSpecName: "utilities") pod "b52f20c9-c735-454a-b086-886643ea82b0" (UID: "b52f20c9-c735-454a-b086-886643ea82b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.867829 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b52f20c9-c735-454a-b086-886643ea82b0-kube-api-access-bxm7t" (OuterVolumeSpecName: "kube-api-access-bxm7t") pod "b52f20c9-c735-454a-b086-886643ea82b0" (UID: "b52f20c9-c735-454a-b086-886643ea82b0"). InnerVolumeSpecName "kube-api-access-bxm7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.964316 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxm7t\" (UniqueName: \"kubernetes.io/projected/b52f20c9-c735-454a-b086-886643ea82b0-kube-api-access-bxm7t\") on node \"crc\" DevicePath \"\"" Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.964388 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b52f20c9-c735-454a-b086-886643ea82b0-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:11:18 crc kubenswrapper[4997]: I1205 07:11:18.980439 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b52f20c9-c735-454a-b086-886643ea82b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b52f20c9-c735-454a-b086-886643ea82b0" (UID: "b52f20c9-c735-454a-b086-886643ea82b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.066169 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b52f20c9-c735-454a-b086-886643ea82b0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.405394 4997 generic.go:334] "Generic (PLEG): container finished" podID="b52f20c9-c735-454a-b086-886643ea82b0" containerID="81209b95927209dd7b101b913adcc0bc56961e816833cd8f9acc054752bcd207" exitCode=0 Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.405531 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-82mpd" Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.405558 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clw95" event={"ID":"b52f20c9-c735-454a-b086-886643ea82b0","Type":"ContainerDied","Data":"81209b95927209dd7b101b913adcc0bc56961e816833cd8f9acc054752bcd207"} Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.405685 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clw95" event={"ID":"b52f20c9-c735-454a-b086-886643ea82b0","Type":"ContainerDied","Data":"9dd4fe4fd7ae0d94649d6a0020228b489c93bc0370717d85b11695a859f73b3a"} Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.405723 4997 scope.go:117] "RemoveContainer" containerID="81209b95927209dd7b101b913adcc0bc56961e816833cd8f9acc054752bcd207" Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.405531 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-clw95" Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.434495 4997 scope.go:117] "RemoveContainer" containerID="456baa907c02760c7fb5a60ffa907123a94f8c1ee597b2b46943f1a3d64e4444" Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.454004 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-82mpd"] Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.461856 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-82mpd"] Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.481994 4997 scope.go:117] "RemoveContainer" containerID="626619ca4a05db089daf721167c798167c165504369a9ff207deaeba1735b369" Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.482651 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-clw95"] Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.489222 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-clw95"] Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.500601 4997 scope.go:117] "RemoveContainer" containerID="81209b95927209dd7b101b913adcc0bc56961e816833cd8f9acc054752bcd207" Dec 05 07:11:19 crc kubenswrapper[4997]: E1205 07:11:19.501266 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81209b95927209dd7b101b913adcc0bc56961e816833cd8f9acc054752bcd207\": container with ID starting with 81209b95927209dd7b101b913adcc0bc56961e816833cd8f9acc054752bcd207 not found: ID does not exist" containerID="81209b95927209dd7b101b913adcc0bc56961e816833cd8f9acc054752bcd207" Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.501306 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81209b95927209dd7b101b913adcc0bc56961e816833cd8f9acc054752bcd207"} err="failed to get container status \"81209b95927209dd7b101b913adcc0bc56961e816833cd8f9acc054752bcd207\": rpc error: code = NotFound desc = could not find container \"81209b95927209dd7b101b913adcc0bc56961e816833cd8f9acc054752bcd207\": container with ID starting with 81209b95927209dd7b101b913adcc0bc56961e816833cd8f9acc054752bcd207 not found: ID does not exist" Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.501339 4997 scope.go:117] "RemoveContainer" containerID="456baa907c02760c7fb5a60ffa907123a94f8c1ee597b2b46943f1a3d64e4444" Dec 05 07:11:19 crc kubenswrapper[4997]: E1205 07:11:19.502001 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"456baa907c02760c7fb5a60ffa907123a94f8c1ee597b2b46943f1a3d64e4444\": container with ID starting with 456baa907c02760c7fb5a60ffa907123a94f8c1ee597b2b46943f1a3d64e4444 not found: ID does not exist" containerID="456baa907c02760c7fb5a60ffa907123a94f8c1ee597b2b46943f1a3d64e4444" Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.502064 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"456baa907c02760c7fb5a60ffa907123a94f8c1ee597b2b46943f1a3d64e4444"} err="failed to get container status \"456baa907c02760c7fb5a60ffa907123a94f8c1ee597b2b46943f1a3d64e4444\": rpc error: code = NotFound desc = could not find container \"456baa907c02760c7fb5a60ffa907123a94f8c1ee597b2b46943f1a3d64e4444\": container with ID starting with 456baa907c02760c7fb5a60ffa907123a94f8c1ee597b2b46943f1a3d64e4444 not found: ID does not exist" Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.502107 4997 scope.go:117] "RemoveContainer" containerID="626619ca4a05db089daf721167c798167c165504369a9ff207deaeba1735b369" Dec 05 07:11:19 crc kubenswrapper[4997]: E1205 07:11:19.502896 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"626619ca4a05db089daf721167c798167c165504369a9ff207deaeba1735b369\": container with ID starting with 626619ca4a05db089daf721167c798167c165504369a9ff207deaeba1735b369 not found: ID does not exist" containerID="626619ca4a05db089daf721167c798167c165504369a9ff207deaeba1735b369" Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.502951 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"626619ca4a05db089daf721167c798167c165504369a9ff207deaeba1735b369"} err="failed to get container status \"626619ca4a05db089daf721167c798167c165504369a9ff207deaeba1735b369\": rpc error: code = NotFound desc = could not find container \"626619ca4a05db089daf721167c798167c165504369a9ff207deaeba1735b369\": container with ID starting with 626619ca4a05db089daf721167c798167c165504369a9ff207deaeba1735b369 not found: ID does not exist" Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.757881 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="144aba01-6e9c-4c56-b7ec-e11dd2855bf3" path="/var/lib/kubelet/pods/144aba01-6e9c-4c56-b7ec-e11dd2855bf3/volumes" Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.759351 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b52f20c9-c735-454a-b086-886643ea82b0" path="/var/lib/kubelet/pods/b52f20c9-c735-454a-b086-886643ea82b0/volumes" Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.769870 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:11:19 crc kubenswrapper[4997]: I1205 07:11:19.769959 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:11:20 crc kubenswrapper[4997]: I1205 07:11:20.415094 4997 generic.go:334] "Generic (PLEG): container finished" podID="e6b7e3f5-ed40-4170-a421-63547bd2fb65" containerID="8e82fd77ef47519a2075c22d2a9149394c3fcebd03a9f630ad15d04f6752fe60" exitCode=0 Dec 05 07:11:20 crc kubenswrapper[4997]: I1205 07:11:20.415230 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" event={"ID":"e6b7e3f5-ed40-4170-a421-63547bd2fb65","Type":"ContainerDied","Data":"8e82fd77ef47519a2075c22d2a9149394c3fcebd03a9f630ad15d04f6752fe60"} Dec 05 07:11:21 crc kubenswrapper[4997]: I1205 07:11:21.431993 4997 generic.go:334] "Generic (PLEG): container finished" podID="e6b7e3f5-ed40-4170-a421-63547bd2fb65" containerID="7921c18777c73f1db674fddaac4c42c15361156784a67ef2ad7d8b1e2a51181b" exitCode=0 Dec 05 07:11:21 crc kubenswrapper[4997]: I1205 07:11:21.432149 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" event={"ID":"e6b7e3f5-ed40-4170-a421-63547bd2fb65","Type":"ContainerDied","Data":"7921c18777c73f1db674fddaac4c42c15361156784a67ef2ad7d8b1e2a51181b"} Dec 05 07:11:22 crc kubenswrapper[4997]: I1205 07:11:22.639889 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" Dec 05 07:11:22 crc kubenswrapper[4997]: I1205 07:11:22.718500 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6b7e3f5-ed40-4170-a421-63547bd2fb65-bundle\") pod \"e6b7e3f5-ed40-4170-a421-63547bd2fb65\" (UID: \"e6b7e3f5-ed40-4170-a421-63547bd2fb65\") " Dec 05 07:11:22 crc kubenswrapper[4997]: I1205 07:11:22.718597 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6b7e3f5-ed40-4170-a421-63547bd2fb65-util\") pod \"e6b7e3f5-ed40-4170-a421-63547bd2fb65\" (UID: \"e6b7e3f5-ed40-4170-a421-63547bd2fb65\") " Dec 05 07:11:22 crc kubenswrapper[4997]: I1205 07:11:22.718713 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dl96\" (UniqueName: \"kubernetes.io/projected/e6b7e3f5-ed40-4170-a421-63547bd2fb65-kube-api-access-9dl96\") pod \"e6b7e3f5-ed40-4170-a421-63547bd2fb65\" (UID: \"e6b7e3f5-ed40-4170-a421-63547bd2fb65\") " Dec 05 07:11:22 crc kubenswrapper[4997]: I1205 07:11:22.719321 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6b7e3f5-ed40-4170-a421-63547bd2fb65-bundle" (OuterVolumeSpecName: "bundle") pod "e6b7e3f5-ed40-4170-a421-63547bd2fb65" (UID: "e6b7e3f5-ed40-4170-a421-63547bd2fb65"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:11:22 crc kubenswrapper[4997]: I1205 07:11:22.723836 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6b7e3f5-ed40-4170-a421-63547bd2fb65-kube-api-access-9dl96" (OuterVolumeSpecName: "kube-api-access-9dl96") pod "e6b7e3f5-ed40-4170-a421-63547bd2fb65" (UID: "e6b7e3f5-ed40-4170-a421-63547bd2fb65"). InnerVolumeSpecName "kube-api-access-9dl96". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:11:22 crc kubenswrapper[4997]: I1205 07:11:22.734697 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6b7e3f5-ed40-4170-a421-63547bd2fb65-util" (OuterVolumeSpecName: "util") pod "e6b7e3f5-ed40-4170-a421-63547bd2fb65" (UID: "e6b7e3f5-ed40-4170-a421-63547bd2fb65"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:11:22 crc kubenswrapper[4997]: I1205 07:11:22.820629 4997 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6b7e3f5-ed40-4170-a421-63547bd2fb65-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:11:22 crc kubenswrapper[4997]: I1205 07:11:22.820663 4997 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6b7e3f5-ed40-4170-a421-63547bd2fb65-util\") on node \"crc\" DevicePath \"\"" Dec 05 07:11:22 crc kubenswrapper[4997]: I1205 07:11:22.820673 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dl96\" (UniqueName: \"kubernetes.io/projected/e6b7e3f5-ed40-4170-a421-63547bd2fb65-kube-api-access-9dl96\") on node \"crc\" DevicePath \"\"" Dec 05 07:11:23 crc kubenswrapper[4997]: I1205 07:11:23.446191 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" event={"ID":"e6b7e3f5-ed40-4170-a421-63547bd2fb65","Type":"ContainerDied","Data":"08fa07ead80532d568d477feb2f1cdfe5a342ec230f373cf611c8b34d60a0d25"} Dec 05 07:11:23 crc kubenswrapper[4997]: I1205 07:11:23.446527 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08fa07ead80532d568d477feb2f1cdfe5a342ec230f373cf611c8b34d60a0d25" Dec 05 07:11:23 crc kubenswrapper[4997]: I1205 07:11:23.446300 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.467347 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-r64sz"] Dec 05 07:11:25 crc kubenswrapper[4997]: E1205 07:11:25.467569 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b52f20c9-c735-454a-b086-886643ea82b0" containerName="extract-utilities" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.467581 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b52f20c9-c735-454a-b086-886643ea82b0" containerName="extract-utilities" Dec 05 07:11:25 crc kubenswrapper[4997]: E1205 07:11:25.467591 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b52f20c9-c735-454a-b086-886643ea82b0" containerName="registry-server" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.467598 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b52f20c9-c735-454a-b086-886643ea82b0" containerName="registry-server" Dec 05 07:11:25 crc kubenswrapper[4997]: E1205 07:11:25.467606 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="144aba01-6e9c-4c56-b7ec-e11dd2855bf3" containerName="extract-utilities" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.467629 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="144aba01-6e9c-4c56-b7ec-e11dd2855bf3" containerName="extract-utilities" Dec 05 07:11:25 crc kubenswrapper[4997]: E1205 07:11:25.467639 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b52f20c9-c735-454a-b086-886643ea82b0" containerName="extract-content" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.467644 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b52f20c9-c735-454a-b086-886643ea82b0" containerName="extract-content" Dec 05 07:11:25 crc kubenswrapper[4997]: E1205 07:11:25.467651 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6b7e3f5-ed40-4170-a421-63547bd2fb65" containerName="extract" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.467656 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6b7e3f5-ed40-4170-a421-63547bd2fb65" containerName="extract" Dec 05 07:11:25 crc kubenswrapper[4997]: E1205 07:11:25.467665 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6b7e3f5-ed40-4170-a421-63547bd2fb65" containerName="util" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.467671 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6b7e3f5-ed40-4170-a421-63547bd2fb65" containerName="util" Dec 05 07:11:25 crc kubenswrapper[4997]: E1205 07:11:25.467684 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6b7e3f5-ed40-4170-a421-63547bd2fb65" containerName="pull" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.467689 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6b7e3f5-ed40-4170-a421-63547bd2fb65" containerName="pull" Dec 05 07:11:25 crc kubenswrapper[4997]: E1205 07:11:25.467700 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="144aba01-6e9c-4c56-b7ec-e11dd2855bf3" containerName="extract-content" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.467706 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="144aba01-6e9c-4c56-b7ec-e11dd2855bf3" containerName="extract-content" Dec 05 07:11:25 crc kubenswrapper[4997]: E1205 07:11:25.467714 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="144aba01-6e9c-4c56-b7ec-e11dd2855bf3" containerName="registry-server" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.467720 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="144aba01-6e9c-4c56-b7ec-e11dd2855bf3" containerName="registry-server" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.467812 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6b7e3f5-ed40-4170-a421-63547bd2fb65" containerName="extract" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.467827 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="b52f20c9-c735-454a-b086-886643ea82b0" containerName="registry-server" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.467834 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="144aba01-6e9c-4c56-b7ec-e11dd2855bf3" containerName="registry-server" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.468213 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-r64sz" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.470735 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.471033 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.471355 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-zzn82" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.483659 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-r64sz"] Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.557745 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxw29\" (UniqueName: \"kubernetes.io/projected/147e0782-52e6-4511-aa4a-27fe62b112dd-kube-api-access-lxw29\") pod \"nmstate-operator-5b5b58f5c8-r64sz\" (UID: \"147e0782-52e6-4511-aa4a-27fe62b112dd\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-r64sz" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.659433 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxw29\" (UniqueName: \"kubernetes.io/projected/147e0782-52e6-4511-aa4a-27fe62b112dd-kube-api-access-lxw29\") pod \"nmstate-operator-5b5b58f5c8-r64sz\" (UID: \"147e0782-52e6-4511-aa4a-27fe62b112dd\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-r64sz" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.678069 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxw29\" (UniqueName: \"kubernetes.io/projected/147e0782-52e6-4511-aa4a-27fe62b112dd-kube-api-access-lxw29\") pod \"nmstate-operator-5b5b58f5c8-r64sz\" (UID: \"147e0782-52e6-4511-aa4a-27fe62b112dd\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-r64sz" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.783273 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-r64sz" Dec 05 07:11:25 crc kubenswrapper[4997]: I1205 07:11:25.989495 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-r64sz"] Dec 05 07:11:26 crc kubenswrapper[4997]: I1205 07:11:26.463994 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-r64sz" event={"ID":"147e0782-52e6-4511-aa4a-27fe62b112dd","Type":"ContainerStarted","Data":"eae651e90e2e6b0cf96fb77443d2069d19638d0371268bd8861de6fba27acd9f"} Dec 05 07:11:28 crc kubenswrapper[4997]: I1205 07:11:28.478587 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-r64sz" event={"ID":"147e0782-52e6-4511-aa4a-27fe62b112dd","Type":"ContainerStarted","Data":"891ecbb957b05a571d22a6c025f4c419fad2c2e3d1482e584b5a8e5d1b322d80"} Dec 05 07:11:28 crc kubenswrapper[4997]: I1205 07:11:28.499678 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-r64sz" podStartSLOduration=1.440821602 podStartE2EDuration="3.499649371s" podCreationTimestamp="2025-12-05 07:11:25 +0000 UTC" firstStartedPulling="2025-12-05 07:11:25.999701263 +0000 UTC m=+986.528608524" lastFinishedPulling="2025-12-05 07:11:28.058529032 +0000 UTC m=+988.587436293" observedRunningTime="2025-12-05 07:11:28.495970603 +0000 UTC m=+989.024877934" watchObservedRunningTime="2025-12-05 07:11:28.499649371 +0000 UTC m=+989.028556632" Dec 05 07:11:34 crc kubenswrapper[4997]: I1205 07:11:34.867535 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-vdrwf"] Dec 05 07:11:34 crc kubenswrapper[4997]: I1205 07:11:34.868994 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vdrwf" Dec 05 07:11:34 crc kubenswrapper[4997]: I1205 07:11:34.871475 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-4mpg8" Dec 05 07:11:34 crc kubenswrapper[4997]: I1205 07:11:34.884083 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth"] Dec 05 07:11:34 crc kubenswrapper[4997]: I1205 07:11:34.884962 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth" Dec 05 07:11:34 crc kubenswrapper[4997]: I1205 07:11:34.887904 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 05 07:11:34 crc kubenswrapper[4997]: I1205 07:11:34.889006 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-vdrwf"] Dec 05 07:11:34 crc kubenswrapper[4997]: I1205 07:11:34.896565 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth"] Dec 05 07:11:34 crc kubenswrapper[4997]: I1205 07:11:34.909821 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-j74g9"] Dec 05 07:11:34 crc kubenswrapper[4997]: I1205 07:11:34.910581 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-j74g9" Dec 05 07:11:34 crc kubenswrapper[4997]: I1205 07:11:34.986170 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/90785160-89df-4c52-b2ed-05c8740f050d-dbus-socket\") pod \"nmstate-handler-j74g9\" (UID: \"90785160-89df-4c52-b2ed-05c8740f050d\") " pod="openshift-nmstate/nmstate-handler-j74g9" Dec 05 07:11:34 crc kubenswrapper[4997]: I1205 07:11:34.986226 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/90785160-89df-4c52-b2ed-05c8740f050d-ovs-socket\") pod \"nmstate-handler-j74g9\" (UID: \"90785160-89df-4c52-b2ed-05c8740f050d\") " pod="openshift-nmstate/nmstate-handler-j74g9" Dec 05 07:11:34 crc kubenswrapper[4997]: I1205 07:11:34.986249 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcrxb\" (UniqueName: \"kubernetes.io/projected/9dc3f953-29af-435f-ba80-fc7d2f3b0dc6-kube-api-access-dcrxb\") pod \"nmstate-webhook-5f6d4c5ccb-4cxth\" (UID: \"9dc3f953-29af-435f-ba80-fc7d2f3b0dc6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth" Dec 05 07:11:34 crc kubenswrapper[4997]: I1205 07:11:34.986281 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2hsd\" (UniqueName: \"kubernetes.io/projected/5607a1c5-5883-42d9-8d2f-9bb2194ee33d-kube-api-access-p2hsd\") pod \"nmstate-metrics-7f946cbc9-vdrwf\" (UID: \"5607a1c5-5883-42d9-8d2f-9bb2194ee33d\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vdrwf" Dec 05 07:11:34 crc kubenswrapper[4997]: I1205 07:11:34.986510 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/90785160-89df-4c52-b2ed-05c8740f050d-nmstate-lock\") pod \"nmstate-handler-j74g9\" (UID: \"90785160-89df-4c52-b2ed-05c8740f050d\") " pod="openshift-nmstate/nmstate-handler-j74g9" Dec 05 07:11:34 crc kubenswrapper[4997]: I1205 07:11:34.986676 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwfrn\" (UniqueName: \"kubernetes.io/projected/90785160-89df-4c52-b2ed-05c8740f050d-kube-api-access-kwfrn\") pod \"nmstate-handler-j74g9\" (UID: \"90785160-89df-4c52-b2ed-05c8740f050d\") " pod="openshift-nmstate/nmstate-handler-j74g9" Dec 05 07:11:34 crc kubenswrapper[4997]: I1205 07:11:34.986759 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9dc3f953-29af-435f-ba80-fc7d2f3b0dc6-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-4cxth\" (UID: \"9dc3f953-29af-435f-ba80-fc7d2f3b0dc6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.011998 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq"] Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.013088 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.015848 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-d58tj" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.015987 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.016129 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.041114 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq"] Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.087845 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/81f37659-4cca-4dc1-85cd-f755196a4ca6-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-gl7kq\" (UID: \"81f37659-4cca-4dc1-85cd-f755196a4ca6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.087904 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/90785160-89df-4c52-b2ed-05c8740f050d-dbus-socket\") pod \"nmstate-handler-j74g9\" (UID: \"90785160-89df-4c52-b2ed-05c8740f050d\") " pod="openshift-nmstate/nmstate-handler-j74g9" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.087927 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/90785160-89df-4c52-b2ed-05c8740f050d-ovs-socket\") pod \"nmstate-handler-j74g9\" (UID: \"90785160-89df-4c52-b2ed-05c8740f050d\") " pod="openshift-nmstate/nmstate-handler-j74g9" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.087952 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcrxb\" (UniqueName: \"kubernetes.io/projected/9dc3f953-29af-435f-ba80-fc7d2f3b0dc6-kube-api-access-dcrxb\") pod \"nmstate-webhook-5f6d4c5ccb-4cxth\" (UID: \"9dc3f953-29af-435f-ba80-fc7d2f3b0dc6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.087998 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2hsd\" (UniqueName: \"kubernetes.io/projected/5607a1c5-5883-42d9-8d2f-9bb2194ee33d-kube-api-access-p2hsd\") pod \"nmstate-metrics-7f946cbc9-vdrwf\" (UID: \"5607a1c5-5883-42d9-8d2f-9bb2194ee33d\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vdrwf" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.088016 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9lng\" (UniqueName: \"kubernetes.io/projected/81f37659-4cca-4dc1-85cd-f755196a4ca6-kube-api-access-b9lng\") pod \"nmstate-console-plugin-7fbb5f6569-gl7kq\" (UID: \"81f37659-4cca-4dc1-85cd-f755196a4ca6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.088037 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/90785160-89df-4c52-b2ed-05c8740f050d-nmstate-lock\") pod \"nmstate-handler-j74g9\" (UID: \"90785160-89df-4c52-b2ed-05c8740f050d\") " pod="openshift-nmstate/nmstate-handler-j74g9" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.088030 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/90785160-89df-4c52-b2ed-05c8740f050d-ovs-socket\") pod \"nmstate-handler-j74g9\" (UID: \"90785160-89df-4c52-b2ed-05c8740f050d\") " pod="openshift-nmstate/nmstate-handler-j74g9" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.088062 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/81f37659-4cca-4dc1-85cd-f755196a4ca6-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-gl7kq\" (UID: \"81f37659-4cca-4dc1-85cd-f755196a4ca6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.088159 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwfrn\" (UniqueName: \"kubernetes.io/projected/90785160-89df-4c52-b2ed-05c8740f050d-kube-api-access-kwfrn\") pod \"nmstate-handler-j74g9\" (UID: \"90785160-89df-4c52-b2ed-05c8740f050d\") " pod="openshift-nmstate/nmstate-handler-j74g9" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.088226 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9dc3f953-29af-435f-ba80-fc7d2f3b0dc6-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-4cxth\" (UID: \"9dc3f953-29af-435f-ba80-fc7d2f3b0dc6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.088251 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/90785160-89df-4c52-b2ed-05c8740f050d-dbus-socket\") pod \"nmstate-handler-j74g9\" (UID: \"90785160-89df-4c52-b2ed-05c8740f050d\") " pod="openshift-nmstate/nmstate-handler-j74g9" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.088354 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/90785160-89df-4c52-b2ed-05c8740f050d-nmstate-lock\") pod \"nmstate-handler-j74g9\" (UID: \"90785160-89df-4c52-b2ed-05c8740f050d\") " pod="openshift-nmstate/nmstate-handler-j74g9" Dec 05 07:11:35 crc kubenswrapper[4997]: E1205 07:11:35.088402 4997 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 05 07:11:35 crc kubenswrapper[4997]: E1205 07:11:35.088556 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9dc3f953-29af-435f-ba80-fc7d2f3b0dc6-tls-key-pair podName:9dc3f953-29af-435f-ba80-fc7d2f3b0dc6 nodeName:}" failed. No retries permitted until 2025-12-05 07:11:35.588538901 +0000 UTC m=+996.117446162 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/9dc3f953-29af-435f-ba80-fc7d2f3b0dc6-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-4cxth" (UID: "9dc3f953-29af-435f-ba80-fc7d2f3b0dc6") : secret "openshift-nmstate-webhook" not found Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.110489 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwfrn\" (UniqueName: \"kubernetes.io/projected/90785160-89df-4c52-b2ed-05c8740f050d-kube-api-access-kwfrn\") pod \"nmstate-handler-j74g9\" (UID: \"90785160-89df-4c52-b2ed-05c8740f050d\") " pod="openshift-nmstate/nmstate-handler-j74g9" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.110641 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2hsd\" (UniqueName: \"kubernetes.io/projected/5607a1c5-5883-42d9-8d2f-9bb2194ee33d-kube-api-access-p2hsd\") pod \"nmstate-metrics-7f946cbc9-vdrwf\" (UID: \"5607a1c5-5883-42d9-8d2f-9bb2194ee33d\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vdrwf" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.112129 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcrxb\" (UniqueName: \"kubernetes.io/projected/9dc3f953-29af-435f-ba80-fc7d2f3b0dc6-kube-api-access-dcrxb\") pod \"nmstate-webhook-5f6d4c5ccb-4cxth\" (UID: \"9dc3f953-29af-435f-ba80-fc7d2f3b0dc6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.186766 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vdrwf" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.189286 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/81f37659-4cca-4dc1-85cd-f755196a4ca6-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-gl7kq\" (UID: \"81f37659-4cca-4dc1-85cd-f755196a4ca6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.189336 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9lng\" (UniqueName: \"kubernetes.io/projected/81f37659-4cca-4dc1-85cd-f755196a4ca6-kube-api-access-b9lng\") pod \"nmstate-console-plugin-7fbb5f6569-gl7kq\" (UID: \"81f37659-4cca-4dc1-85cd-f755196a4ca6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.189402 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/81f37659-4cca-4dc1-85cd-f755196a4ca6-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-gl7kq\" (UID: \"81f37659-4cca-4dc1-85cd-f755196a4ca6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq" Dec 05 07:11:35 crc kubenswrapper[4997]: E1205 07:11:35.189519 4997 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 05 07:11:35 crc kubenswrapper[4997]: E1205 07:11:35.189594 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81f37659-4cca-4dc1-85cd-f755196a4ca6-plugin-serving-cert podName:81f37659-4cca-4dc1-85cd-f755196a4ca6 nodeName:}" failed. No retries permitted until 2025-12-05 07:11:35.689572602 +0000 UTC m=+996.218479863 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/81f37659-4cca-4dc1-85cd-f755196a4ca6-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-gl7kq" (UID: "81f37659-4cca-4dc1-85cd-f755196a4ca6") : secret "plugin-serving-cert" not found Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.190522 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/81f37659-4cca-4dc1-85cd-f755196a4ca6-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-gl7kq\" (UID: \"81f37659-4cca-4dc1-85cd-f755196a4ca6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.235755 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-9fb45fd54-wjww6"] Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.237011 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.265791 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9lng\" (UniqueName: \"kubernetes.io/projected/81f37659-4cca-4dc1-85cd-f755196a4ca6-kube-api-access-b9lng\") pod \"nmstate-console-plugin-7fbb5f6569-gl7kq\" (UID: \"81f37659-4cca-4dc1-85cd-f755196a4ca6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.266326 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-j74g9" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.271738 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-9fb45fd54-wjww6"] Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.291254 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-console-config\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.291321 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8ttf\" (UniqueName: \"kubernetes.io/projected/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-kube-api-access-r8ttf\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.291387 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-service-ca\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.291409 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-oauth-serving-cert\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.291455 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-trusted-ca-bundle\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.291510 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-console-serving-cert\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.291572 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-console-oauth-config\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: W1205 07:11:35.309050 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90785160_89df_4c52_b2ed_05c8740f050d.slice/crio-850cacf0a069b57d5460219565e231abf4d72cea3317f92e263dc714b5638c96 WatchSource:0}: Error finding container 850cacf0a069b57d5460219565e231abf4d72cea3317f92e263dc714b5638c96: Status 404 returned error can't find the container with id 850cacf0a069b57d5460219565e231abf4d72cea3317f92e263dc714b5638c96 Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.392867 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8ttf\" (UniqueName: \"kubernetes.io/projected/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-kube-api-access-r8ttf\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.392938 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-service-ca\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.392955 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-oauth-serving-cert\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.392979 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-trusted-ca-bundle\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.393022 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-console-serving-cert\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.393040 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-console-oauth-config\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.393075 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-console-config\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.394240 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-oauth-serving-cert\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.395133 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-service-ca\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.395976 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-trusted-ca-bundle\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.397135 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-console-config\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.398354 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-console-oauth-config\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.401241 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-console-serving-cert\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.415343 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8ttf\" (UniqueName: \"kubernetes.io/projected/7fd7a7c0-ea30-4711-9d97-6a93226be5c0-kube-api-access-r8ttf\") pod \"console-9fb45fd54-wjww6\" (UID: \"7fd7a7c0-ea30-4711-9d97-6a93226be5c0\") " pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.436711 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-vdrwf"] Dec 05 07:11:35 crc kubenswrapper[4997]: W1205 07:11:35.445271 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5607a1c5_5883_42d9_8d2f_9bb2194ee33d.slice/crio-3a2818e31a603e26192a9b2b6eed4f8f5efa3f1ded5b9c1fefb103547dd3b3ec WatchSource:0}: Error finding container 3a2818e31a603e26192a9b2b6eed4f8f5efa3f1ded5b9c1fefb103547dd3b3ec: Status 404 returned error can't find the container with id 3a2818e31a603e26192a9b2b6eed4f8f5efa3f1ded5b9c1fefb103547dd3b3ec Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.517807 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vdrwf" event={"ID":"5607a1c5-5883-42d9-8d2f-9bb2194ee33d","Type":"ContainerStarted","Data":"3a2818e31a603e26192a9b2b6eed4f8f5efa3f1ded5b9c1fefb103547dd3b3ec"} Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.520088 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-j74g9" event={"ID":"90785160-89df-4c52-b2ed-05c8740f050d","Type":"ContainerStarted","Data":"850cacf0a069b57d5460219565e231abf4d72cea3317f92e263dc714b5638c96"} Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.572965 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.595925 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9dc3f953-29af-435f-ba80-fc7d2f3b0dc6-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-4cxth\" (UID: \"9dc3f953-29af-435f-ba80-fc7d2f3b0dc6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.599975 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9dc3f953-29af-435f-ba80-fc7d2f3b0dc6-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-4cxth\" (UID: \"9dc3f953-29af-435f-ba80-fc7d2f3b0dc6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.697903 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/81f37659-4cca-4dc1-85cd-f755196a4ca6-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-gl7kq\" (UID: \"81f37659-4cca-4dc1-85cd-f755196a4ca6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.701263 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/81f37659-4cca-4dc1-85cd-f755196a4ca6-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-gl7kq\" (UID: \"81f37659-4cca-4dc1-85cd-f755196a4ca6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.851401 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.927215 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq" Dec 05 07:11:35 crc kubenswrapper[4997]: I1205 07:11:35.967461 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-9fb45fd54-wjww6"] Dec 05 07:11:36 crc kubenswrapper[4997]: I1205 07:11:36.062140 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth"] Dec 05 07:11:36 crc kubenswrapper[4997]: W1205 07:11:36.075707 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9dc3f953_29af_435f_ba80_fc7d2f3b0dc6.slice/crio-dc772761add3d8c6d79a3345ef61429d58c615eb2207d6f9371cf7b416231514 WatchSource:0}: Error finding container dc772761add3d8c6d79a3345ef61429d58c615eb2207d6f9371cf7b416231514: Status 404 returned error can't find the container with id dc772761add3d8c6d79a3345ef61429d58c615eb2207d6f9371cf7b416231514 Dec 05 07:11:36 crc kubenswrapper[4997]: I1205 07:11:36.135932 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq"] Dec 05 07:11:36 crc kubenswrapper[4997]: W1205 07:11:36.145847 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81f37659_4cca_4dc1_85cd_f755196a4ca6.slice/crio-a627b377a83654732f860c85153ff7452253bda887d54b8fc01703c8bca6ec99 WatchSource:0}: Error finding container a627b377a83654732f860c85153ff7452253bda887d54b8fc01703c8bca6ec99: Status 404 returned error can't find the container with id a627b377a83654732f860c85153ff7452253bda887d54b8fc01703c8bca6ec99 Dec 05 07:11:36 crc kubenswrapper[4997]: I1205 07:11:36.527272 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq" event={"ID":"81f37659-4cca-4dc1-85cd-f755196a4ca6","Type":"ContainerStarted","Data":"a627b377a83654732f860c85153ff7452253bda887d54b8fc01703c8bca6ec99"} Dec 05 07:11:36 crc kubenswrapper[4997]: I1205 07:11:36.530632 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-9fb45fd54-wjww6" event={"ID":"7fd7a7c0-ea30-4711-9d97-6a93226be5c0","Type":"ContainerStarted","Data":"ab5f15e17d09d7ce01887f6f3f7606588d48e558fc033081fabbd170d0d86b1c"} Dec 05 07:11:36 crc kubenswrapper[4997]: I1205 07:11:36.530676 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-9fb45fd54-wjww6" event={"ID":"7fd7a7c0-ea30-4711-9d97-6a93226be5c0","Type":"ContainerStarted","Data":"baefcbb4fd682a2ba9a61cfe9dffafc824dab15658b2cd9cb61b7366b791c7ef"} Dec 05 07:11:36 crc kubenswrapper[4997]: I1205 07:11:36.533816 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth" event={"ID":"9dc3f953-29af-435f-ba80-fc7d2f3b0dc6","Type":"ContainerStarted","Data":"dc772761add3d8c6d79a3345ef61429d58c615eb2207d6f9371cf7b416231514"} Dec 05 07:11:36 crc kubenswrapper[4997]: I1205 07:11:36.557085 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-9fb45fd54-wjww6" podStartSLOduration=1.557058016 podStartE2EDuration="1.557058016s" podCreationTimestamp="2025-12-05 07:11:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:11:36.552641119 +0000 UTC m=+997.081548400" watchObservedRunningTime="2025-12-05 07:11:36.557058016 +0000 UTC m=+997.085965287" Dec 05 07:11:38 crc kubenswrapper[4997]: I1205 07:11:38.551659 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vdrwf" event={"ID":"5607a1c5-5883-42d9-8d2f-9bb2194ee33d","Type":"ContainerStarted","Data":"f807420e3fc78090a493166bb687aec11c4e18260d964d93821f93ba92c2255b"} Dec 05 07:11:38 crc kubenswrapper[4997]: I1205 07:11:38.555563 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-j74g9" event={"ID":"90785160-89df-4c52-b2ed-05c8740f050d","Type":"ContainerStarted","Data":"f6d0dd8cebbca2bba7e823b0157e8e98625dcf00e3ad84280bc9e64a1f414fe2"} Dec 05 07:11:38 crc kubenswrapper[4997]: I1205 07:11:38.555737 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-j74g9" Dec 05 07:11:38 crc kubenswrapper[4997]: I1205 07:11:38.558112 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth" event={"ID":"9dc3f953-29af-435f-ba80-fc7d2f3b0dc6","Type":"ContainerStarted","Data":"805e2a91397e262c0a7056dbbefc57f1a0a7bc620951fecad8fe0728976133b5"} Dec 05 07:11:38 crc kubenswrapper[4997]: I1205 07:11:38.558331 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth" Dec 05 07:11:38 crc kubenswrapper[4997]: I1205 07:11:38.578476 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-j74g9" podStartSLOduration=2.040552678 podStartE2EDuration="4.578453847s" podCreationTimestamp="2025-12-05 07:11:34 +0000 UTC" firstStartedPulling="2025-12-05 07:11:35.311876619 +0000 UTC m=+995.840783880" lastFinishedPulling="2025-12-05 07:11:37.849777788 +0000 UTC m=+998.378685049" observedRunningTime="2025-12-05 07:11:38.572254182 +0000 UTC m=+999.101161463" watchObservedRunningTime="2025-12-05 07:11:38.578453847 +0000 UTC m=+999.107361108" Dec 05 07:11:38 crc kubenswrapper[4997]: I1205 07:11:38.603061 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth" podStartSLOduration=2.800791498 podStartE2EDuration="4.603028542s" podCreationTimestamp="2025-12-05 07:11:34 +0000 UTC" firstStartedPulling="2025-12-05 07:11:36.078027067 +0000 UTC m=+996.606934328" lastFinishedPulling="2025-12-05 07:11:37.880264101 +0000 UTC m=+998.409171372" observedRunningTime="2025-12-05 07:11:38.60298084 +0000 UTC m=+999.131888121" watchObservedRunningTime="2025-12-05 07:11:38.603028542 +0000 UTC m=+999.131935803" Dec 05 07:11:39 crc kubenswrapper[4997]: I1205 07:11:39.566225 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq" event={"ID":"81f37659-4cca-4dc1-85cd-f755196a4ca6","Type":"ContainerStarted","Data":"f254345fe1f4753c5399251fc975a25a5dd952a581ce6dcd8fe97faf430c8d4e"} Dec 05 07:11:39 crc kubenswrapper[4997]: I1205 07:11:39.770094 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-gl7kq" podStartSLOduration=3.02950262 podStartE2EDuration="5.770072656s" podCreationTimestamp="2025-12-05 07:11:34 +0000 UTC" firstStartedPulling="2025-12-05 07:11:36.147994701 +0000 UTC m=+996.676901962" lastFinishedPulling="2025-12-05 07:11:38.888564737 +0000 UTC m=+999.417471998" observedRunningTime="2025-12-05 07:11:39.584161105 +0000 UTC m=+1000.113068386" watchObservedRunningTime="2025-12-05 07:11:39.770072656 +0000 UTC m=+1000.298979917" Dec 05 07:11:40 crc kubenswrapper[4997]: I1205 07:11:40.575311 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vdrwf" event={"ID":"5607a1c5-5883-42d9-8d2f-9bb2194ee33d","Type":"ContainerStarted","Data":"dac836a433c7ad3614a6f33316a1902cbc0d0112cf1491dfdeb44ff615bcaba1"} Dec 05 07:11:40 crc kubenswrapper[4997]: I1205 07:11:40.596797 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vdrwf" podStartSLOduration=1.792723297 podStartE2EDuration="6.596763727s" podCreationTimestamp="2025-12-05 07:11:34 +0000 UTC" firstStartedPulling="2025-12-05 07:11:35.447284076 +0000 UTC m=+995.976191337" lastFinishedPulling="2025-12-05 07:11:40.251324506 +0000 UTC m=+1000.780231767" observedRunningTime="2025-12-05 07:11:40.593799678 +0000 UTC m=+1001.122706969" watchObservedRunningTime="2025-12-05 07:11:40.596763727 +0000 UTC m=+1001.125671018" Dec 05 07:11:45 crc kubenswrapper[4997]: I1205 07:11:45.287077 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-j74g9" Dec 05 07:11:45 crc kubenswrapper[4997]: I1205 07:11:45.574004 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:45 crc kubenswrapper[4997]: I1205 07:11:45.574268 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:45 crc kubenswrapper[4997]: I1205 07:11:45.583516 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:45 crc kubenswrapper[4997]: I1205 07:11:45.613588 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-9fb45fd54-wjww6" Dec 05 07:11:45 crc kubenswrapper[4997]: I1205 07:11:45.676202 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-lskhh"] Dec 05 07:11:49 crc kubenswrapper[4997]: I1205 07:11:49.770027 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:11:49 crc kubenswrapper[4997]: I1205 07:11:49.770406 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:11:55 crc kubenswrapper[4997]: I1205 07:11:55.858470 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth" Dec 05 07:12:08 crc kubenswrapper[4997]: I1205 07:12:08.268695 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z"] Dec 05 07:12:08 crc kubenswrapper[4997]: I1205 07:12:08.270367 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" Dec 05 07:12:08 crc kubenswrapper[4997]: I1205 07:12:08.272698 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 05 07:12:08 crc kubenswrapper[4997]: I1205 07:12:08.279741 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z"] Dec 05 07:12:08 crc kubenswrapper[4997]: I1205 07:12:08.381426 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3e947fc-fe79-4011-b548-b9accaefd1b1-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z\" (UID: \"e3e947fc-fe79-4011-b548-b9accaefd1b1\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" Dec 05 07:12:08 crc kubenswrapper[4997]: I1205 07:12:08.381524 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb566\" (UniqueName: \"kubernetes.io/projected/e3e947fc-fe79-4011-b548-b9accaefd1b1-kube-api-access-pb566\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z\" (UID: \"e3e947fc-fe79-4011-b548-b9accaefd1b1\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" Dec 05 07:12:08 crc kubenswrapper[4997]: I1205 07:12:08.381555 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3e947fc-fe79-4011-b548-b9accaefd1b1-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z\" (UID: \"e3e947fc-fe79-4011-b548-b9accaefd1b1\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" Dec 05 07:12:08 crc kubenswrapper[4997]: I1205 07:12:08.483825 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb566\" (UniqueName: \"kubernetes.io/projected/e3e947fc-fe79-4011-b548-b9accaefd1b1-kube-api-access-pb566\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z\" (UID: \"e3e947fc-fe79-4011-b548-b9accaefd1b1\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" Dec 05 07:12:08 crc kubenswrapper[4997]: I1205 07:12:08.483907 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3e947fc-fe79-4011-b548-b9accaefd1b1-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z\" (UID: \"e3e947fc-fe79-4011-b548-b9accaefd1b1\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" Dec 05 07:12:08 crc kubenswrapper[4997]: I1205 07:12:08.483975 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3e947fc-fe79-4011-b548-b9accaefd1b1-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z\" (UID: \"e3e947fc-fe79-4011-b548-b9accaefd1b1\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" Dec 05 07:12:08 crc kubenswrapper[4997]: I1205 07:12:08.485282 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3e947fc-fe79-4011-b548-b9accaefd1b1-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z\" (UID: \"e3e947fc-fe79-4011-b548-b9accaefd1b1\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" Dec 05 07:12:08 crc kubenswrapper[4997]: I1205 07:12:08.485291 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3e947fc-fe79-4011-b548-b9accaefd1b1-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z\" (UID: \"e3e947fc-fe79-4011-b548-b9accaefd1b1\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" Dec 05 07:12:08 crc kubenswrapper[4997]: I1205 07:12:08.522899 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb566\" (UniqueName: \"kubernetes.io/projected/e3e947fc-fe79-4011-b548-b9accaefd1b1-kube-api-access-pb566\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z\" (UID: \"e3e947fc-fe79-4011-b548-b9accaefd1b1\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" Dec 05 07:12:08 crc kubenswrapper[4997]: I1205 07:12:08.592569 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" Dec 05 07:12:09 crc kubenswrapper[4997]: I1205 07:12:09.008963 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z"] Dec 05 07:12:09 crc kubenswrapper[4997]: I1205 07:12:09.767734 4997 generic.go:334] "Generic (PLEG): container finished" podID="e3e947fc-fe79-4011-b548-b9accaefd1b1" containerID="0931d9dfc3030beba53c739aa84479770d2f50cf701d41d281c18d8af9487e98" exitCode=0 Dec 05 07:12:09 crc kubenswrapper[4997]: I1205 07:12:09.767977 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" event={"ID":"e3e947fc-fe79-4011-b548-b9accaefd1b1","Type":"ContainerDied","Data":"0931d9dfc3030beba53c739aa84479770d2f50cf701d41d281c18d8af9487e98"} Dec 05 07:12:09 crc kubenswrapper[4997]: I1205 07:12:09.768147 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" event={"ID":"e3e947fc-fe79-4011-b548-b9accaefd1b1","Type":"ContainerStarted","Data":"257656b4efbdcc5170a2cf08971cb91b534c0e68a8295c82821b5991eb137805"} Dec 05 07:12:10 crc kubenswrapper[4997]: I1205 07:12:10.728305 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-lskhh" podUID="0160154e-35ae-471c-b848-214b9f33ec93" containerName="console" containerID="cri-o://beb6b2ddc048d904d5272115b463f6cb94839a1053b7f034b77aaa415d342db0" gracePeriod=15 Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.126552 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-lskhh_0160154e-35ae-471c-b848-214b9f33ec93/console/0.log" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.127133 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-lskhh" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.232864 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0160154e-35ae-471c-b848-214b9f33ec93-console-serving-cert\") pod \"0160154e-35ae-471c-b848-214b9f33ec93\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.232997 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-oauth-serving-cert\") pod \"0160154e-35ae-471c-b848-214b9f33ec93\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.233050 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-service-ca\") pod \"0160154e-35ae-471c-b848-214b9f33ec93\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.233088 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-console-config\") pod \"0160154e-35ae-471c-b848-214b9f33ec93\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.233130 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0160154e-35ae-471c-b848-214b9f33ec93-console-oauth-config\") pod \"0160154e-35ae-471c-b848-214b9f33ec93\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.233182 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-trusted-ca-bundle\") pod \"0160154e-35ae-471c-b848-214b9f33ec93\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.233244 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vs7j\" (UniqueName: \"kubernetes.io/projected/0160154e-35ae-471c-b848-214b9f33ec93-kube-api-access-6vs7j\") pod \"0160154e-35ae-471c-b848-214b9f33ec93\" (UID: \"0160154e-35ae-471c-b848-214b9f33ec93\") " Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.235086 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-service-ca" (OuterVolumeSpecName: "service-ca") pod "0160154e-35ae-471c-b848-214b9f33ec93" (UID: "0160154e-35ae-471c-b848-214b9f33ec93"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.235102 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "0160154e-35ae-471c-b848-214b9f33ec93" (UID: "0160154e-35ae-471c-b848-214b9f33ec93"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.235273 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-console-config" (OuterVolumeSpecName: "console-config") pod "0160154e-35ae-471c-b848-214b9f33ec93" (UID: "0160154e-35ae-471c-b848-214b9f33ec93"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.235345 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "0160154e-35ae-471c-b848-214b9f33ec93" (UID: "0160154e-35ae-471c-b848-214b9f33ec93"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.241078 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0160154e-35ae-471c-b848-214b9f33ec93-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "0160154e-35ae-471c-b848-214b9f33ec93" (UID: "0160154e-35ae-471c-b848-214b9f33ec93"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.241256 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0160154e-35ae-471c-b848-214b9f33ec93-kube-api-access-6vs7j" (OuterVolumeSpecName: "kube-api-access-6vs7j") pod "0160154e-35ae-471c-b848-214b9f33ec93" (UID: "0160154e-35ae-471c-b848-214b9f33ec93"). InnerVolumeSpecName "kube-api-access-6vs7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.241473 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0160154e-35ae-471c-b848-214b9f33ec93-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "0160154e-35ae-471c-b848-214b9f33ec93" (UID: "0160154e-35ae-471c-b848-214b9f33ec93"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.334724 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vs7j\" (UniqueName: \"kubernetes.io/projected/0160154e-35ae-471c-b848-214b9f33ec93-kube-api-access-6vs7j\") on node \"crc\" DevicePath \"\"" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.334792 4997 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0160154e-35ae-471c-b848-214b9f33ec93-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.334812 4997 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.334832 4997 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.334850 4997 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-console-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.334869 4997 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0160154e-35ae-471c-b848-214b9f33ec93-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.334887 4997 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0160154e-35ae-471c-b848-214b9f33ec93-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.788908 4997 generic.go:334] "Generic (PLEG): container finished" podID="e3e947fc-fe79-4011-b548-b9accaefd1b1" containerID="abbf6520ca742e89be5f95b5c3a2d23341e55eefe518190353a22f447cb8b9c9" exitCode=0 Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.789087 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" event={"ID":"e3e947fc-fe79-4011-b548-b9accaefd1b1","Type":"ContainerDied","Data":"abbf6520ca742e89be5f95b5c3a2d23341e55eefe518190353a22f447cb8b9c9"} Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.792657 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-lskhh_0160154e-35ae-471c-b848-214b9f33ec93/console/0.log" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.792746 4997 generic.go:334] "Generic (PLEG): container finished" podID="0160154e-35ae-471c-b848-214b9f33ec93" containerID="beb6b2ddc048d904d5272115b463f6cb94839a1053b7f034b77aaa415d342db0" exitCode=2 Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.792790 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-lskhh" event={"ID":"0160154e-35ae-471c-b848-214b9f33ec93","Type":"ContainerDied","Data":"beb6b2ddc048d904d5272115b463f6cb94839a1053b7f034b77aaa415d342db0"} Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.792831 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-lskhh" event={"ID":"0160154e-35ae-471c-b848-214b9f33ec93","Type":"ContainerDied","Data":"dca35c09a300c42e2a141902bd5ba976fceec3765d300b85974ab61bc38f1783"} Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.792875 4997 scope.go:117] "RemoveContainer" containerID="beb6b2ddc048d904d5272115b463f6cb94839a1053b7f034b77aaa415d342db0" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.793052 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-lskhh" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.832754 4997 scope.go:117] "RemoveContainer" containerID="beb6b2ddc048d904d5272115b463f6cb94839a1053b7f034b77aaa415d342db0" Dec 05 07:12:11 crc kubenswrapper[4997]: E1205 07:12:11.834157 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"beb6b2ddc048d904d5272115b463f6cb94839a1053b7f034b77aaa415d342db0\": container with ID starting with beb6b2ddc048d904d5272115b463f6cb94839a1053b7f034b77aaa415d342db0 not found: ID does not exist" containerID="beb6b2ddc048d904d5272115b463f6cb94839a1053b7f034b77aaa415d342db0" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.834237 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beb6b2ddc048d904d5272115b463f6cb94839a1053b7f034b77aaa415d342db0"} err="failed to get container status \"beb6b2ddc048d904d5272115b463f6cb94839a1053b7f034b77aaa415d342db0\": rpc error: code = NotFound desc = could not find container \"beb6b2ddc048d904d5272115b463f6cb94839a1053b7f034b77aaa415d342db0\": container with ID starting with beb6b2ddc048d904d5272115b463f6cb94839a1053b7f034b77aaa415d342db0 not found: ID does not exist" Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.868933 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-lskhh"] Dec 05 07:12:11 crc kubenswrapper[4997]: I1205 07:12:11.873267 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-lskhh"] Dec 05 07:12:12 crc kubenswrapper[4997]: I1205 07:12:12.805560 4997 generic.go:334] "Generic (PLEG): container finished" podID="e3e947fc-fe79-4011-b548-b9accaefd1b1" containerID="988be18529c199053788f160d9e670238367b7f6aa815e76c3ef2c47c3812f36" exitCode=0 Dec 05 07:12:12 crc kubenswrapper[4997]: I1205 07:12:12.805671 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" event={"ID":"e3e947fc-fe79-4011-b548-b9accaefd1b1","Type":"ContainerDied","Data":"988be18529c199053788f160d9e670238367b7f6aa815e76c3ef2c47c3812f36"} Dec 05 07:12:13 crc kubenswrapper[4997]: I1205 07:12:13.757907 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0160154e-35ae-471c-b848-214b9f33ec93" path="/var/lib/kubelet/pods/0160154e-35ae-471c-b848-214b9f33ec93/volumes" Dec 05 07:12:14 crc kubenswrapper[4997]: I1205 07:12:14.139855 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" Dec 05 07:12:14 crc kubenswrapper[4997]: I1205 07:12:14.179726 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3e947fc-fe79-4011-b548-b9accaefd1b1-bundle\") pod \"e3e947fc-fe79-4011-b548-b9accaefd1b1\" (UID: \"e3e947fc-fe79-4011-b548-b9accaefd1b1\") " Dec 05 07:12:14 crc kubenswrapper[4997]: I1205 07:12:14.179885 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3e947fc-fe79-4011-b548-b9accaefd1b1-util\") pod \"e3e947fc-fe79-4011-b548-b9accaefd1b1\" (UID: \"e3e947fc-fe79-4011-b548-b9accaefd1b1\") " Dec 05 07:12:14 crc kubenswrapper[4997]: I1205 07:12:14.180035 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb566\" (UniqueName: \"kubernetes.io/projected/e3e947fc-fe79-4011-b548-b9accaefd1b1-kube-api-access-pb566\") pod \"e3e947fc-fe79-4011-b548-b9accaefd1b1\" (UID: \"e3e947fc-fe79-4011-b548-b9accaefd1b1\") " Dec 05 07:12:14 crc kubenswrapper[4997]: I1205 07:12:14.180966 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3e947fc-fe79-4011-b548-b9accaefd1b1-bundle" (OuterVolumeSpecName: "bundle") pod "e3e947fc-fe79-4011-b548-b9accaefd1b1" (UID: "e3e947fc-fe79-4011-b548-b9accaefd1b1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:12:14 crc kubenswrapper[4997]: I1205 07:12:14.186768 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3e947fc-fe79-4011-b548-b9accaefd1b1-kube-api-access-pb566" (OuterVolumeSpecName: "kube-api-access-pb566") pod "e3e947fc-fe79-4011-b548-b9accaefd1b1" (UID: "e3e947fc-fe79-4011-b548-b9accaefd1b1"). InnerVolumeSpecName "kube-api-access-pb566". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:12:14 crc kubenswrapper[4997]: I1205 07:12:14.200148 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3e947fc-fe79-4011-b548-b9accaefd1b1-util" (OuterVolumeSpecName: "util") pod "e3e947fc-fe79-4011-b548-b9accaefd1b1" (UID: "e3e947fc-fe79-4011-b548-b9accaefd1b1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:12:14 crc kubenswrapper[4997]: I1205 07:12:14.282347 4997 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3e947fc-fe79-4011-b548-b9accaefd1b1-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:12:14 crc kubenswrapper[4997]: I1205 07:12:14.282405 4997 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3e947fc-fe79-4011-b548-b9accaefd1b1-util\") on node \"crc\" DevicePath \"\"" Dec 05 07:12:14 crc kubenswrapper[4997]: I1205 07:12:14.282425 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb566\" (UniqueName: \"kubernetes.io/projected/e3e947fc-fe79-4011-b548-b9accaefd1b1-kube-api-access-pb566\") on node \"crc\" DevicePath \"\"" Dec 05 07:12:14 crc kubenswrapper[4997]: I1205 07:12:14.827485 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" event={"ID":"e3e947fc-fe79-4011-b548-b9accaefd1b1","Type":"ContainerDied","Data":"257656b4efbdcc5170a2cf08971cb91b534c0e68a8295c82821b5991eb137805"} Dec 05 07:12:14 crc kubenswrapper[4997]: I1205 07:12:14.827561 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="257656b4efbdcc5170a2cf08971cb91b534c0e68a8295c82821b5991eb137805" Dec 05 07:12:14 crc kubenswrapper[4997]: I1205 07:12:14.827657 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z" Dec 05 07:12:19 crc kubenswrapper[4997]: I1205 07:12:19.770465 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:12:19 crc kubenswrapper[4997]: I1205 07:12:19.771306 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:12:19 crc kubenswrapper[4997]: I1205 07:12:19.771374 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 07:12:19 crc kubenswrapper[4997]: I1205 07:12:19.772211 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"046abc9c3b110cde88610f054e681c42bd85fac9c9e032fa744fe8940f595f0c"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 07:12:19 crc kubenswrapper[4997]: I1205 07:12:19.772292 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://046abc9c3b110cde88610f054e681c42bd85fac9c9e032fa744fe8940f595f0c" gracePeriod=600 Dec 05 07:12:20 crc kubenswrapper[4997]: I1205 07:12:20.866703 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="046abc9c3b110cde88610f054e681c42bd85fac9c9e032fa744fe8940f595f0c" exitCode=0 Dec 05 07:12:20 crc kubenswrapper[4997]: I1205 07:12:20.866785 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"046abc9c3b110cde88610f054e681c42bd85fac9c9e032fa744fe8940f595f0c"} Dec 05 07:12:20 crc kubenswrapper[4997]: I1205 07:12:20.867325 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"6cb7be4205d2b6d3befa7784153570d8d5c45d7400acdc190e0c1608d52679b2"} Dec 05 07:12:20 crc kubenswrapper[4997]: I1205 07:12:20.867369 4997 scope.go:117] "RemoveContainer" containerID="1ff8d1fdf0f1aff6bb5ff2695a5046a42e548eeb16bf2838702262c451fcb00b" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.417921 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h54rr"] Dec 05 07:12:21 crc kubenswrapper[4997]: E1205 07:12:21.418668 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e947fc-fe79-4011-b548-b9accaefd1b1" containerName="extract" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.418689 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e947fc-fe79-4011-b548-b9accaefd1b1" containerName="extract" Dec 05 07:12:21 crc kubenswrapper[4997]: E1205 07:12:21.418702 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e947fc-fe79-4011-b548-b9accaefd1b1" containerName="util" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.418709 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e947fc-fe79-4011-b548-b9accaefd1b1" containerName="util" Dec 05 07:12:21 crc kubenswrapper[4997]: E1205 07:12:21.418718 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0160154e-35ae-471c-b848-214b9f33ec93" containerName="console" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.418724 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0160154e-35ae-471c-b848-214b9f33ec93" containerName="console" Dec 05 07:12:21 crc kubenswrapper[4997]: E1205 07:12:21.418740 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e947fc-fe79-4011-b548-b9accaefd1b1" containerName="pull" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.418747 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e947fc-fe79-4011-b548-b9accaefd1b1" containerName="pull" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.418867 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0160154e-35ae-471c-b848-214b9f33ec93" containerName="console" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.418886 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3e947fc-fe79-4011-b548-b9accaefd1b1" containerName="extract" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.419905 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h54rr" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.434833 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h54rr"] Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.491567 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdncz\" (UniqueName: \"kubernetes.io/projected/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-kube-api-access-wdncz\") pod \"certified-operators-h54rr\" (UID: \"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782\") " pod="openshift-marketplace/certified-operators-h54rr" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.491733 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-utilities\") pod \"certified-operators-h54rr\" (UID: \"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782\") " pod="openshift-marketplace/certified-operators-h54rr" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.491778 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-catalog-content\") pod \"certified-operators-h54rr\" (UID: \"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782\") " pod="openshift-marketplace/certified-operators-h54rr" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.593013 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-utilities\") pod \"certified-operators-h54rr\" (UID: \"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782\") " pod="openshift-marketplace/certified-operators-h54rr" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.593089 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-catalog-content\") pod \"certified-operators-h54rr\" (UID: \"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782\") " pod="openshift-marketplace/certified-operators-h54rr" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.593118 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdncz\" (UniqueName: \"kubernetes.io/projected/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-kube-api-access-wdncz\") pod \"certified-operators-h54rr\" (UID: \"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782\") " pod="openshift-marketplace/certified-operators-h54rr" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.593523 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-utilities\") pod \"certified-operators-h54rr\" (UID: \"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782\") " pod="openshift-marketplace/certified-operators-h54rr" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.593576 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-catalog-content\") pod \"certified-operators-h54rr\" (UID: \"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782\") " pod="openshift-marketplace/certified-operators-h54rr" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.623571 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdncz\" (UniqueName: \"kubernetes.io/projected/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-kube-api-access-wdncz\") pod \"certified-operators-h54rr\" (UID: \"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782\") " pod="openshift-marketplace/certified-operators-h54rr" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.739499 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h54rr" Dec 05 07:12:21 crc kubenswrapper[4997]: I1205 07:12:21.970439 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h54rr"] Dec 05 07:12:22 crc kubenswrapper[4997]: I1205 07:12:22.892983 4997 generic.go:334] "Generic (PLEG): container finished" podID="3fcb9a01-6fb0-44c4-9bbe-cabf57b35782" containerID="f61682d1b42afc3d3992d8ea0f285ca36fcd1f04913ad9f7113da00305c4f2e3" exitCode=0 Dec 05 07:12:22 crc kubenswrapper[4997]: I1205 07:12:22.893073 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h54rr" event={"ID":"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782","Type":"ContainerDied","Data":"f61682d1b42afc3d3992d8ea0f285ca36fcd1f04913ad9f7113da00305c4f2e3"} Dec 05 07:12:22 crc kubenswrapper[4997]: I1205 07:12:22.893393 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h54rr" event={"ID":"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782","Type":"ContainerStarted","Data":"4e70d86cf63381b55a4eb2a5572a6c86bdc3a42ac66dd47f3cb22e366ab6c404"} Dec 05 07:12:23 crc kubenswrapper[4997]: I1205 07:12:23.904062 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h54rr" event={"ID":"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782","Type":"ContainerStarted","Data":"6d1d5a3339d8b1dc64f49834c41e3bb34d86d460feab4d4cecb54b63bcc794f6"} Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.437222 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n"] Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.438576 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.445359 4997 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.445548 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.445573 4997 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-lgq5d" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.445715 4997 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.445780 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.455883 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n"] Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.561293 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0f68e314-6e6f-40e9-b439-3f91124f9150-apiservice-cert\") pod \"metallb-operator-controller-manager-5b6d8f454b-9vl4n\" (UID: \"0f68e314-6e6f-40e9-b439-3f91124f9150\") " pod="metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.561413 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mjbd\" (UniqueName: \"kubernetes.io/projected/0f68e314-6e6f-40e9-b439-3f91124f9150-kube-api-access-2mjbd\") pod \"metallb-operator-controller-manager-5b6d8f454b-9vl4n\" (UID: \"0f68e314-6e6f-40e9-b439-3f91124f9150\") " pod="metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.561457 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0f68e314-6e6f-40e9-b439-3f91124f9150-webhook-cert\") pod \"metallb-operator-controller-manager-5b6d8f454b-9vl4n\" (UID: \"0f68e314-6e6f-40e9-b439-3f91124f9150\") " pod="metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.662711 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0f68e314-6e6f-40e9-b439-3f91124f9150-apiservice-cert\") pod \"metallb-operator-controller-manager-5b6d8f454b-9vl4n\" (UID: \"0f68e314-6e6f-40e9-b439-3f91124f9150\") " pod="metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.662832 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mjbd\" (UniqueName: \"kubernetes.io/projected/0f68e314-6e6f-40e9-b439-3f91124f9150-kube-api-access-2mjbd\") pod \"metallb-operator-controller-manager-5b6d8f454b-9vl4n\" (UID: \"0f68e314-6e6f-40e9-b439-3f91124f9150\") " pod="metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.662884 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0f68e314-6e6f-40e9-b439-3f91124f9150-webhook-cert\") pod \"metallb-operator-controller-manager-5b6d8f454b-9vl4n\" (UID: \"0f68e314-6e6f-40e9-b439-3f91124f9150\") " pod="metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.669862 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0f68e314-6e6f-40e9-b439-3f91124f9150-webhook-cert\") pod \"metallb-operator-controller-manager-5b6d8f454b-9vl4n\" (UID: \"0f68e314-6e6f-40e9-b439-3f91124f9150\") " pod="metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.678687 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0f68e314-6e6f-40e9-b439-3f91124f9150-apiservice-cert\") pod \"metallb-operator-controller-manager-5b6d8f454b-9vl4n\" (UID: \"0f68e314-6e6f-40e9-b439-3f91124f9150\") " pod="metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.681318 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mjbd\" (UniqueName: \"kubernetes.io/projected/0f68e314-6e6f-40e9-b439-3f91124f9150-kube-api-access-2mjbd\") pod \"metallb-operator-controller-manager-5b6d8f454b-9vl4n\" (UID: \"0f68e314-6e6f-40e9-b439-3f91124f9150\") " pod="metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.757408 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.772570 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm"] Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.773412 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.775906 4997 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.776132 4997 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.776559 4997 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-rpt87" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.861898 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm"] Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.912749 4997 generic.go:334] "Generic (PLEG): container finished" podID="3fcb9a01-6fb0-44c4-9bbe-cabf57b35782" containerID="6d1d5a3339d8b1dc64f49834c41e3bb34d86d460feab4d4cecb54b63bcc794f6" exitCode=0 Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.912827 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h54rr" event={"ID":"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782","Type":"ContainerDied","Data":"6d1d5a3339d8b1dc64f49834c41e3bb34d86d460feab4d4cecb54b63bcc794f6"} Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.971128 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/211e2b11-88f1-4660-9a8b-b43dc7ceb1ec-webhook-cert\") pod \"metallb-operator-webhook-server-68c5d7fff9-9gxmm\" (UID: \"211e2b11-88f1-4660-9a8b-b43dc7ceb1ec\") " pod="metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.971248 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/211e2b11-88f1-4660-9a8b-b43dc7ceb1ec-apiservice-cert\") pod \"metallb-operator-webhook-server-68c5d7fff9-9gxmm\" (UID: \"211e2b11-88f1-4660-9a8b-b43dc7ceb1ec\") " pod="metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm" Dec 05 07:12:24 crc kubenswrapper[4997]: I1205 07:12:24.971562 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgk75\" (UniqueName: \"kubernetes.io/projected/211e2b11-88f1-4660-9a8b-b43dc7ceb1ec-kube-api-access-fgk75\") pod \"metallb-operator-webhook-server-68c5d7fff9-9gxmm\" (UID: \"211e2b11-88f1-4660-9a8b-b43dc7ceb1ec\") " pod="metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm" Dec 05 07:12:25 crc kubenswrapper[4997]: I1205 07:12:25.073734 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/211e2b11-88f1-4660-9a8b-b43dc7ceb1ec-webhook-cert\") pod \"metallb-operator-webhook-server-68c5d7fff9-9gxmm\" (UID: \"211e2b11-88f1-4660-9a8b-b43dc7ceb1ec\") " pod="metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm" Dec 05 07:12:25 crc kubenswrapper[4997]: I1205 07:12:25.073820 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/211e2b11-88f1-4660-9a8b-b43dc7ceb1ec-apiservice-cert\") pod \"metallb-operator-webhook-server-68c5d7fff9-9gxmm\" (UID: \"211e2b11-88f1-4660-9a8b-b43dc7ceb1ec\") " pod="metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm" Dec 05 07:12:25 crc kubenswrapper[4997]: I1205 07:12:25.073922 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgk75\" (UniqueName: \"kubernetes.io/projected/211e2b11-88f1-4660-9a8b-b43dc7ceb1ec-kube-api-access-fgk75\") pod \"metallb-operator-webhook-server-68c5d7fff9-9gxmm\" (UID: \"211e2b11-88f1-4660-9a8b-b43dc7ceb1ec\") " pod="metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm" Dec 05 07:12:25 crc kubenswrapper[4997]: I1205 07:12:25.102159 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/211e2b11-88f1-4660-9a8b-b43dc7ceb1ec-webhook-cert\") pod \"metallb-operator-webhook-server-68c5d7fff9-9gxmm\" (UID: \"211e2b11-88f1-4660-9a8b-b43dc7ceb1ec\") " pod="metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm" Dec 05 07:12:25 crc kubenswrapper[4997]: I1205 07:12:25.103485 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/211e2b11-88f1-4660-9a8b-b43dc7ceb1ec-apiservice-cert\") pod \"metallb-operator-webhook-server-68c5d7fff9-9gxmm\" (UID: \"211e2b11-88f1-4660-9a8b-b43dc7ceb1ec\") " pod="metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm" Dec 05 07:12:25 crc kubenswrapper[4997]: I1205 07:12:25.126199 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgk75\" (UniqueName: \"kubernetes.io/projected/211e2b11-88f1-4660-9a8b-b43dc7ceb1ec-kube-api-access-fgk75\") pod \"metallb-operator-webhook-server-68c5d7fff9-9gxmm\" (UID: \"211e2b11-88f1-4660-9a8b-b43dc7ceb1ec\") " pod="metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm" Dec 05 07:12:25 crc kubenswrapper[4997]: I1205 07:12:25.141998 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm" Dec 05 07:12:25 crc kubenswrapper[4997]: I1205 07:12:25.215033 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n"] Dec 05 07:12:25 crc kubenswrapper[4997]: W1205 07:12:25.220924 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f68e314_6e6f_40e9_b439_3f91124f9150.slice/crio-805f1c9890a869e9c7a2f7e1095e3369e0a6bb22e9cd3bd4212f99abe41d5c0e WatchSource:0}: Error finding container 805f1c9890a869e9c7a2f7e1095e3369e0a6bb22e9cd3bd4212f99abe41d5c0e: Status 404 returned error can't find the container with id 805f1c9890a869e9c7a2f7e1095e3369e0a6bb22e9cd3bd4212f99abe41d5c0e Dec 05 07:12:25 crc kubenswrapper[4997]: I1205 07:12:25.482013 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm"] Dec 05 07:12:25 crc kubenswrapper[4997]: W1205 07:12:25.490283 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod211e2b11_88f1_4660_9a8b_b43dc7ceb1ec.slice/crio-3f2649159ef56640e0cae3f04a1db642971aa4541ddf085c25c3b7e34354c501 WatchSource:0}: Error finding container 3f2649159ef56640e0cae3f04a1db642971aa4541ddf085c25c3b7e34354c501: Status 404 returned error can't find the container with id 3f2649159ef56640e0cae3f04a1db642971aa4541ddf085c25c3b7e34354c501 Dec 05 07:12:25 crc kubenswrapper[4997]: I1205 07:12:25.921011 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n" event={"ID":"0f68e314-6e6f-40e9-b439-3f91124f9150","Type":"ContainerStarted","Data":"805f1c9890a869e9c7a2f7e1095e3369e0a6bb22e9cd3bd4212f99abe41d5c0e"} Dec 05 07:12:25 crc kubenswrapper[4997]: I1205 07:12:25.923308 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h54rr" event={"ID":"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782","Type":"ContainerStarted","Data":"20c6a54615120458c01a11cf011ccf1a15bb826fea4be9774fee8002a6781521"} Dec 05 07:12:25 crc kubenswrapper[4997]: I1205 07:12:25.929075 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm" event={"ID":"211e2b11-88f1-4660-9a8b-b43dc7ceb1ec","Type":"ContainerStarted","Data":"3f2649159ef56640e0cae3f04a1db642971aa4541ddf085c25c3b7e34354c501"} Dec 05 07:12:25 crc kubenswrapper[4997]: I1205 07:12:25.945846 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h54rr" podStartSLOduration=2.494560658 podStartE2EDuration="4.945824469s" podCreationTimestamp="2025-12-05 07:12:21 +0000 UTC" firstStartedPulling="2025-12-05 07:12:22.895098727 +0000 UTC m=+1043.424005988" lastFinishedPulling="2025-12-05 07:12:25.346362538 +0000 UTC m=+1045.875269799" observedRunningTime="2025-12-05 07:12:25.943889587 +0000 UTC m=+1046.472796868" watchObservedRunningTime="2025-12-05 07:12:25.945824469 +0000 UTC m=+1046.474731730" Dec 05 07:12:31 crc kubenswrapper[4997]: I1205 07:12:31.740294 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h54rr" Dec 05 07:12:31 crc kubenswrapper[4997]: I1205 07:12:31.741100 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h54rr" Dec 05 07:12:31 crc kubenswrapper[4997]: I1205 07:12:31.809081 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h54rr" Dec 05 07:12:32 crc kubenswrapper[4997]: I1205 07:12:32.108447 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h54rr" Dec 05 07:12:32 crc kubenswrapper[4997]: I1205 07:12:32.155875 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h54rr"] Dec 05 07:12:33 crc kubenswrapper[4997]: I1205 07:12:33.019955 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n" event={"ID":"0f68e314-6e6f-40e9-b439-3f91124f9150","Type":"ContainerStarted","Data":"03bd8373d597e7c70cc230fa056ce8267347a33ee0f5ab910940039fd9d2f1fe"} Dec 05 07:12:33 crc kubenswrapper[4997]: I1205 07:12:33.021807 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n" Dec 05 07:12:33 crc kubenswrapper[4997]: I1205 07:12:33.023772 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm" event={"ID":"211e2b11-88f1-4660-9a8b-b43dc7ceb1ec","Type":"ContainerStarted","Data":"81ddedfa768a6f3c5996bbc3b173e037d65925523df30d043cd115b3c59a9d09"} Dec 05 07:12:33 crc kubenswrapper[4997]: I1205 07:12:33.045683 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n" podStartSLOduration=1.984273704 podStartE2EDuration="9.045663194s" podCreationTimestamp="2025-12-05 07:12:24 +0000 UTC" firstStartedPulling="2025-12-05 07:12:25.226095384 +0000 UTC m=+1045.755002645" lastFinishedPulling="2025-12-05 07:12:32.287484874 +0000 UTC m=+1052.816392135" observedRunningTime="2025-12-05 07:12:33.044080391 +0000 UTC m=+1053.572987672" watchObservedRunningTime="2025-12-05 07:12:33.045663194 +0000 UTC m=+1053.574570475" Dec 05 07:12:33 crc kubenswrapper[4997]: I1205 07:12:33.066951 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm" podStartSLOduration=2.258460587 podStartE2EDuration="9.066928596s" podCreationTimestamp="2025-12-05 07:12:24 +0000 UTC" firstStartedPulling="2025-12-05 07:12:25.494528172 +0000 UTC m=+1046.023435433" lastFinishedPulling="2025-12-05 07:12:32.302996181 +0000 UTC m=+1052.831903442" observedRunningTime="2025-12-05 07:12:33.064298085 +0000 UTC m=+1053.593205356" watchObservedRunningTime="2025-12-05 07:12:33.066928596 +0000 UTC m=+1053.595835857" Dec 05 07:12:34 crc kubenswrapper[4997]: I1205 07:12:34.029395 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h54rr" podUID="3fcb9a01-6fb0-44c4-9bbe-cabf57b35782" containerName="registry-server" containerID="cri-o://20c6a54615120458c01a11cf011ccf1a15bb826fea4be9774fee8002a6781521" gracePeriod=2 Dec 05 07:12:34 crc kubenswrapper[4997]: I1205 07:12:34.029632 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm" Dec 05 07:12:35 crc kubenswrapper[4997]: I1205 07:12:35.039890 4997 generic.go:334] "Generic (PLEG): container finished" podID="3fcb9a01-6fb0-44c4-9bbe-cabf57b35782" containerID="20c6a54615120458c01a11cf011ccf1a15bb826fea4be9774fee8002a6781521" exitCode=0 Dec 05 07:12:35 crc kubenswrapper[4997]: I1205 07:12:35.039966 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h54rr" event={"ID":"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782","Type":"ContainerDied","Data":"20c6a54615120458c01a11cf011ccf1a15bb826fea4be9774fee8002a6781521"} Dec 05 07:12:35 crc kubenswrapper[4997]: I1205 07:12:35.569971 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h54rr" Dec 05 07:12:35 crc kubenswrapper[4997]: I1205 07:12:35.741738 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdncz\" (UniqueName: \"kubernetes.io/projected/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-kube-api-access-wdncz\") pod \"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782\" (UID: \"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782\") " Dec 05 07:12:35 crc kubenswrapper[4997]: I1205 07:12:35.741840 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-catalog-content\") pod \"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782\" (UID: \"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782\") " Dec 05 07:12:35 crc kubenswrapper[4997]: I1205 07:12:35.741912 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-utilities\") pod \"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782\" (UID: \"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782\") " Dec 05 07:12:35 crc kubenswrapper[4997]: I1205 07:12:35.743746 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-utilities" (OuterVolumeSpecName: "utilities") pod "3fcb9a01-6fb0-44c4-9bbe-cabf57b35782" (UID: "3fcb9a01-6fb0-44c4-9bbe-cabf57b35782"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:12:35 crc kubenswrapper[4997]: I1205 07:12:35.751017 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-kube-api-access-wdncz" (OuterVolumeSpecName: "kube-api-access-wdncz") pod "3fcb9a01-6fb0-44c4-9bbe-cabf57b35782" (UID: "3fcb9a01-6fb0-44c4-9bbe-cabf57b35782"). InnerVolumeSpecName "kube-api-access-wdncz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:12:35 crc kubenswrapper[4997]: I1205 07:12:35.801638 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3fcb9a01-6fb0-44c4-9bbe-cabf57b35782" (UID: "3fcb9a01-6fb0-44c4-9bbe-cabf57b35782"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:12:35 crc kubenswrapper[4997]: I1205 07:12:35.843761 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdncz\" (UniqueName: \"kubernetes.io/projected/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-kube-api-access-wdncz\") on node \"crc\" DevicePath \"\"" Dec 05 07:12:35 crc kubenswrapper[4997]: I1205 07:12:35.843800 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:12:35 crc kubenswrapper[4997]: I1205 07:12:35.843811 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:12:36 crc kubenswrapper[4997]: I1205 07:12:36.048897 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h54rr" event={"ID":"3fcb9a01-6fb0-44c4-9bbe-cabf57b35782","Type":"ContainerDied","Data":"4e70d86cf63381b55a4eb2a5572a6c86bdc3a42ac66dd47f3cb22e366ab6c404"} Dec 05 07:12:36 crc kubenswrapper[4997]: I1205 07:12:36.048958 4997 scope.go:117] "RemoveContainer" containerID="20c6a54615120458c01a11cf011ccf1a15bb826fea4be9774fee8002a6781521" Dec 05 07:12:36 crc kubenswrapper[4997]: I1205 07:12:36.048960 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h54rr" Dec 05 07:12:36 crc kubenswrapper[4997]: I1205 07:12:36.068481 4997 scope.go:117] "RemoveContainer" containerID="6d1d5a3339d8b1dc64f49834c41e3bb34d86d460feab4d4cecb54b63bcc794f6" Dec 05 07:12:36 crc kubenswrapper[4997]: I1205 07:12:36.084668 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h54rr"] Dec 05 07:12:36 crc kubenswrapper[4997]: I1205 07:12:36.088686 4997 scope.go:117] "RemoveContainer" containerID="f61682d1b42afc3d3992d8ea0f285ca36fcd1f04913ad9f7113da00305c4f2e3" Dec 05 07:12:36 crc kubenswrapper[4997]: I1205 07:12:36.088840 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h54rr"] Dec 05 07:12:37 crc kubenswrapper[4997]: I1205 07:12:37.757002 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fcb9a01-6fb0-44c4-9bbe-cabf57b35782" path="/var/lib/kubelet/pods/3fcb9a01-6fb0-44c4-9bbe-cabf57b35782/volumes" Dec 05 07:12:45 crc kubenswrapper[4997]: I1205 07:12:45.152573 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-68c5d7fff9-9gxmm" Dec 05 07:13:04 crc kubenswrapper[4997]: I1205 07:13:04.761467 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5b6d8f454b-9vl4n" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.456228 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6"] Dec 05 07:13:05 crc kubenswrapper[4997]: E1205 07:13:05.456828 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fcb9a01-6fb0-44c4-9bbe-cabf57b35782" containerName="extract-content" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.456843 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fcb9a01-6fb0-44c4-9bbe-cabf57b35782" containerName="extract-content" Dec 05 07:13:05 crc kubenswrapper[4997]: E1205 07:13:05.456857 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fcb9a01-6fb0-44c4-9bbe-cabf57b35782" containerName="registry-server" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.456863 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fcb9a01-6fb0-44c4-9bbe-cabf57b35782" containerName="registry-server" Dec 05 07:13:05 crc kubenswrapper[4997]: E1205 07:13:05.456874 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fcb9a01-6fb0-44c4-9bbe-cabf57b35782" containerName="extract-utilities" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.456881 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fcb9a01-6fb0-44c4-9bbe-cabf57b35782" containerName="extract-utilities" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.456986 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fcb9a01-6fb0-44c4-9bbe-cabf57b35782" containerName="registry-server" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.457468 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.467531 4997 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.467819 4997 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-smcsh" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.472098 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-2plrj"] Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.473594 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khxx8\" (UniqueName: \"kubernetes.io/projected/79868cb5-be74-451c-8260-04bca75ded8c-kube-api-access-khxx8\") pod \"frr-k8s-webhook-server-7fcb986d4-c9zt6\" (UID: \"79868cb5-be74-451c-8260-04bca75ded8c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.473935 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79868cb5-be74-451c-8260-04bca75ded8c-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-c9zt6\" (UID: \"79868cb5-be74-451c-8260-04bca75ded8c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.474964 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.481726 4997 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.482573 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.486715 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6"] Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.574820 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/24697cfc-2c63-4b44-b4db-88a361476c5f-metrics\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.575026 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khxx8\" (UniqueName: \"kubernetes.io/projected/79868cb5-be74-451c-8260-04bca75ded8c-kube-api-access-khxx8\") pod \"frr-k8s-webhook-server-7fcb986d4-c9zt6\" (UID: \"79868cb5-be74-451c-8260-04bca75ded8c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.575563 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/24697cfc-2c63-4b44-b4db-88a361476c5f-reloader\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.575718 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/24697cfc-2c63-4b44-b4db-88a361476c5f-frr-startup\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.575816 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/24697cfc-2c63-4b44-b4db-88a361476c5f-frr-conf\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.575898 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhjs9\" (UniqueName: \"kubernetes.io/projected/24697cfc-2c63-4b44-b4db-88a361476c5f-kube-api-access-qhjs9\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.575989 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/24697cfc-2c63-4b44-b4db-88a361476c5f-frr-sockets\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.576093 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79868cb5-be74-451c-8260-04bca75ded8c-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-c9zt6\" (UID: \"79868cb5-be74-451c-8260-04bca75ded8c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.576194 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/24697cfc-2c63-4b44-b4db-88a361476c5f-metrics-certs\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: E1205 07:13:05.576216 4997 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 05 07:13:05 crc kubenswrapper[4997]: E1205 07:13:05.576463 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79868cb5-be74-451c-8260-04bca75ded8c-cert podName:79868cb5-be74-451c-8260-04bca75ded8c nodeName:}" failed. No retries permitted until 2025-12-05 07:13:06.076439498 +0000 UTC m=+1086.605346759 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/79868cb5-be74-451c-8260-04bca75ded8c-cert") pod "frr-k8s-webhook-server-7fcb986d4-c9zt6" (UID: "79868cb5-be74-451c-8260-04bca75ded8c") : secret "frr-k8s-webhook-server-cert" not found Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.589519 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-sdv46"] Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.590641 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-sdv46" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.593468 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.593926 4997 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.594420 4997 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-ffx87" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.595856 4997 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.615117 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-4qdnj"] Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.621506 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-4qdnj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.623180 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khxx8\" (UniqueName: \"kubernetes.io/projected/79868cb5-be74-451c-8260-04bca75ded8c-kube-api-access-khxx8\") pod \"frr-k8s-webhook-server-7fcb986d4-c9zt6\" (UID: \"79868cb5-be74-451c-8260-04bca75ded8c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.627004 4997 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.642454 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-4qdnj"] Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.678224 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pvqv\" (UniqueName: \"kubernetes.io/projected/5e7b1770-04d8-41e3-bdfa-cafc699d03e1-kube-api-access-2pvqv\") pod \"controller-f8648f98b-4qdnj\" (UID: \"5e7b1770-04d8-41e3-bdfa-cafc699d03e1\") " pod="metallb-system/controller-f8648f98b-4qdnj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.678459 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/24697cfc-2c63-4b44-b4db-88a361476c5f-metrics-certs\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.678536 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c0d286d5-b144-44ef-9696-21777bfad09b-metallb-excludel2\") pod \"speaker-sdv46\" (UID: \"c0d286d5-b144-44ef-9696-21777bfad09b\") " pod="metallb-system/speaker-sdv46" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.678578 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k57l\" (UniqueName: \"kubernetes.io/projected/c0d286d5-b144-44ef-9696-21777bfad09b-kube-api-access-7k57l\") pod \"speaker-sdv46\" (UID: \"c0d286d5-b144-44ef-9696-21777bfad09b\") " pod="metallb-system/speaker-sdv46" Dec 05 07:13:05 crc kubenswrapper[4997]: E1205 07:13:05.678648 4997 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.678666 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c0d286d5-b144-44ef-9696-21777bfad09b-memberlist\") pod \"speaker-sdv46\" (UID: \"c0d286d5-b144-44ef-9696-21777bfad09b\") " pod="metallb-system/speaker-sdv46" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.678695 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7b1770-04d8-41e3-bdfa-cafc699d03e1-metrics-certs\") pod \"controller-f8648f98b-4qdnj\" (UID: \"5e7b1770-04d8-41e3-bdfa-cafc699d03e1\") " pod="metallb-system/controller-f8648f98b-4qdnj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.678745 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/24697cfc-2c63-4b44-b4db-88a361476c5f-metrics\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: E1205 07:13:05.678818 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/24697cfc-2c63-4b44-b4db-88a361476c5f-metrics-certs podName:24697cfc-2c63-4b44-b4db-88a361476c5f nodeName:}" failed. No retries permitted until 2025-12-05 07:13:06.178793912 +0000 UTC m=+1086.707701173 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/24697cfc-2c63-4b44-b4db-88a361476c5f-metrics-certs") pod "frr-k8s-2plrj" (UID: "24697cfc-2c63-4b44-b4db-88a361476c5f") : secret "frr-k8s-certs-secret" not found Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.678894 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/24697cfc-2c63-4b44-b4db-88a361476c5f-reloader\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.679109 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/24697cfc-2c63-4b44-b4db-88a361476c5f-frr-startup\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.679153 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5e7b1770-04d8-41e3-bdfa-cafc699d03e1-cert\") pod \"controller-f8648f98b-4qdnj\" (UID: \"5e7b1770-04d8-41e3-bdfa-cafc699d03e1\") " pod="metallb-system/controller-f8648f98b-4qdnj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.679184 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/24697cfc-2c63-4b44-b4db-88a361476c5f-frr-conf\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.679218 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhjs9\" (UniqueName: \"kubernetes.io/projected/24697cfc-2c63-4b44-b4db-88a361476c5f-kube-api-access-qhjs9\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.679250 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/24697cfc-2c63-4b44-b4db-88a361476c5f-frr-sockets\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.679289 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c0d286d5-b144-44ef-9696-21777bfad09b-metrics-certs\") pod \"speaker-sdv46\" (UID: \"c0d286d5-b144-44ef-9696-21777bfad09b\") " pod="metallb-system/speaker-sdv46" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.680117 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/24697cfc-2c63-4b44-b4db-88a361476c5f-metrics\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.680530 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/24697cfc-2c63-4b44-b4db-88a361476c5f-frr-sockets\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.680727 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/24697cfc-2c63-4b44-b4db-88a361476c5f-frr-conf\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.681090 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/24697cfc-2c63-4b44-b4db-88a361476c5f-reloader\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.681480 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/24697cfc-2c63-4b44-b4db-88a361476c5f-frr-startup\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.700113 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhjs9\" (UniqueName: \"kubernetes.io/projected/24697cfc-2c63-4b44-b4db-88a361476c5f-kube-api-access-qhjs9\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.780600 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pvqv\" (UniqueName: \"kubernetes.io/projected/5e7b1770-04d8-41e3-bdfa-cafc699d03e1-kube-api-access-2pvqv\") pod \"controller-f8648f98b-4qdnj\" (UID: \"5e7b1770-04d8-41e3-bdfa-cafc699d03e1\") " pod="metallb-system/controller-f8648f98b-4qdnj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.780702 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c0d286d5-b144-44ef-9696-21777bfad09b-metallb-excludel2\") pod \"speaker-sdv46\" (UID: \"c0d286d5-b144-44ef-9696-21777bfad09b\") " pod="metallb-system/speaker-sdv46" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.780727 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k57l\" (UniqueName: \"kubernetes.io/projected/c0d286d5-b144-44ef-9696-21777bfad09b-kube-api-access-7k57l\") pod \"speaker-sdv46\" (UID: \"c0d286d5-b144-44ef-9696-21777bfad09b\") " pod="metallb-system/speaker-sdv46" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.780756 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c0d286d5-b144-44ef-9696-21777bfad09b-memberlist\") pod \"speaker-sdv46\" (UID: \"c0d286d5-b144-44ef-9696-21777bfad09b\") " pod="metallb-system/speaker-sdv46" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.780773 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7b1770-04d8-41e3-bdfa-cafc699d03e1-metrics-certs\") pod \"controller-f8648f98b-4qdnj\" (UID: \"5e7b1770-04d8-41e3-bdfa-cafc699d03e1\") " pod="metallb-system/controller-f8648f98b-4qdnj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.780809 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5e7b1770-04d8-41e3-bdfa-cafc699d03e1-cert\") pod \"controller-f8648f98b-4qdnj\" (UID: \"5e7b1770-04d8-41e3-bdfa-cafc699d03e1\") " pod="metallb-system/controller-f8648f98b-4qdnj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.780860 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c0d286d5-b144-44ef-9696-21777bfad09b-metrics-certs\") pod \"speaker-sdv46\" (UID: \"c0d286d5-b144-44ef-9696-21777bfad09b\") " pod="metallb-system/speaker-sdv46" Dec 05 07:13:05 crc kubenswrapper[4997]: E1205 07:13:05.780942 4997 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 05 07:13:05 crc kubenswrapper[4997]: E1205 07:13:05.780987 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c0d286d5-b144-44ef-9696-21777bfad09b-metrics-certs podName:c0d286d5-b144-44ef-9696-21777bfad09b nodeName:}" failed. No retries permitted until 2025-12-05 07:13:06.280972079 +0000 UTC m=+1086.809879340 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c0d286d5-b144-44ef-9696-21777bfad09b-metrics-certs") pod "speaker-sdv46" (UID: "c0d286d5-b144-44ef-9696-21777bfad09b") : secret "speaker-certs-secret" not found Dec 05 07:13:05 crc kubenswrapper[4997]: E1205 07:13:05.781088 4997 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 05 07:13:05 crc kubenswrapper[4997]: E1205 07:13:05.781106 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c0d286d5-b144-44ef-9696-21777bfad09b-memberlist podName:c0d286d5-b144-44ef-9696-21777bfad09b nodeName:}" failed. No retries permitted until 2025-12-05 07:13:06.281100593 +0000 UTC m=+1086.810007854 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c0d286d5-b144-44ef-9696-21777bfad09b-memberlist") pod "speaker-sdv46" (UID: "c0d286d5-b144-44ef-9696-21777bfad09b") : secret "metallb-memberlist" not found Dec 05 07:13:05 crc kubenswrapper[4997]: E1205 07:13:05.781142 4997 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 05 07:13:05 crc kubenswrapper[4997]: E1205 07:13:05.781161 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e7b1770-04d8-41e3-bdfa-cafc699d03e1-metrics-certs podName:5e7b1770-04d8-41e3-bdfa-cafc699d03e1 nodeName:}" failed. No retries permitted until 2025-12-05 07:13:06.281155445 +0000 UTC m=+1086.810062706 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e7b1770-04d8-41e3-bdfa-cafc699d03e1-metrics-certs") pod "controller-f8648f98b-4qdnj" (UID: "5e7b1770-04d8-41e3-bdfa-cafc699d03e1") : secret "controller-certs-secret" not found Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.781799 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c0d286d5-b144-44ef-9696-21777bfad09b-metallb-excludel2\") pod \"speaker-sdv46\" (UID: \"c0d286d5-b144-44ef-9696-21777bfad09b\") " pod="metallb-system/speaker-sdv46" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.783283 4997 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.796132 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5e7b1770-04d8-41e3-bdfa-cafc699d03e1-cert\") pod \"controller-f8648f98b-4qdnj\" (UID: \"5e7b1770-04d8-41e3-bdfa-cafc699d03e1\") " pod="metallb-system/controller-f8648f98b-4qdnj" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.798235 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k57l\" (UniqueName: \"kubernetes.io/projected/c0d286d5-b144-44ef-9696-21777bfad09b-kube-api-access-7k57l\") pod \"speaker-sdv46\" (UID: \"c0d286d5-b144-44ef-9696-21777bfad09b\") " pod="metallb-system/speaker-sdv46" Dec 05 07:13:05 crc kubenswrapper[4997]: I1205 07:13:05.805206 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pvqv\" (UniqueName: \"kubernetes.io/projected/5e7b1770-04d8-41e3-bdfa-cafc699d03e1-kube-api-access-2pvqv\") pod \"controller-f8648f98b-4qdnj\" (UID: \"5e7b1770-04d8-41e3-bdfa-cafc699d03e1\") " pod="metallb-system/controller-f8648f98b-4qdnj" Dec 05 07:13:06 crc kubenswrapper[4997]: I1205 07:13:06.086630 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79868cb5-be74-451c-8260-04bca75ded8c-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-c9zt6\" (UID: \"79868cb5-be74-451c-8260-04bca75ded8c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6" Dec 05 07:13:06 crc kubenswrapper[4997]: I1205 07:13:06.091294 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79868cb5-be74-451c-8260-04bca75ded8c-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-c9zt6\" (UID: \"79868cb5-be74-451c-8260-04bca75ded8c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6" Dec 05 07:13:06 crc kubenswrapper[4997]: I1205 07:13:06.188588 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/24697cfc-2c63-4b44-b4db-88a361476c5f-metrics-certs\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:06 crc kubenswrapper[4997]: I1205 07:13:06.194327 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/24697cfc-2c63-4b44-b4db-88a361476c5f-metrics-certs\") pod \"frr-k8s-2plrj\" (UID: \"24697cfc-2c63-4b44-b4db-88a361476c5f\") " pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:06 crc kubenswrapper[4997]: I1205 07:13:06.290410 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c0d286d5-b144-44ef-9696-21777bfad09b-metrics-certs\") pod \"speaker-sdv46\" (UID: \"c0d286d5-b144-44ef-9696-21777bfad09b\") " pod="metallb-system/speaker-sdv46" Dec 05 07:13:06 crc kubenswrapper[4997]: I1205 07:13:06.290509 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c0d286d5-b144-44ef-9696-21777bfad09b-memberlist\") pod \"speaker-sdv46\" (UID: \"c0d286d5-b144-44ef-9696-21777bfad09b\") " pod="metallb-system/speaker-sdv46" Dec 05 07:13:06 crc kubenswrapper[4997]: I1205 07:13:06.290532 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7b1770-04d8-41e3-bdfa-cafc699d03e1-metrics-certs\") pod \"controller-f8648f98b-4qdnj\" (UID: \"5e7b1770-04d8-41e3-bdfa-cafc699d03e1\") " pod="metallb-system/controller-f8648f98b-4qdnj" Dec 05 07:13:06 crc kubenswrapper[4997]: E1205 07:13:06.290843 4997 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 05 07:13:06 crc kubenswrapper[4997]: E1205 07:13:06.290993 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c0d286d5-b144-44ef-9696-21777bfad09b-memberlist podName:c0d286d5-b144-44ef-9696-21777bfad09b nodeName:}" failed. No retries permitted until 2025-12-05 07:13:07.290972455 +0000 UTC m=+1087.819879906 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c0d286d5-b144-44ef-9696-21777bfad09b-memberlist") pod "speaker-sdv46" (UID: "c0d286d5-b144-44ef-9696-21777bfad09b") : secret "metallb-memberlist" not found Dec 05 07:13:06 crc kubenswrapper[4997]: I1205 07:13:06.293376 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c0d286d5-b144-44ef-9696-21777bfad09b-metrics-certs\") pod \"speaker-sdv46\" (UID: \"c0d286d5-b144-44ef-9696-21777bfad09b\") " pod="metallb-system/speaker-sdv46" Dec 05 07:13:06 crc kubenswrapper[4997]: I1205 07:13:06.295189 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7b1770-04d8-41e3-bdfa-cafc699d03e1-metrics-certs\") pod \"controller-f8648f98b-4qdnj\" (UID: \"5e7b1770-04d8-41e3-bdfa-cafc699d03e1\") " pod="metallb-system/controller-f8648f98b-4qdnj" Dec 05 07:13:06 crc kubenswrapper[4997]: I1205 07:13:06.391439 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6" Dec 05 07:13:06 crc kubenswrapper[4997]: I1205 07:13:06.402903 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:06 crc kubenswrapper[4997]: I1205 07:13:06.541696 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-4qdnj" Dec 05 07:13:06 crc kubenswrapper[4997]: I1205 07:13:06.685298 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6"] Dec 05 07:13:06 crc kubenswrapper[4997]: I1205 07:13:06.897426 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-4qdnj"] Dec 05 07:13:07 crc kubenswrapper[4997]: I1205 07:13:07.246946 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2plrj" event={"ID":"24697cfc-2c63-4b44-b4db-88a361476c5f","Type":"ContainerStarted","Data":"b8bac469a35bba5f0f9e0923bae2425c275aff11f70ebd305723153778c164ca"} Dec 05 07:13:07 crc kubenswrapper[4997]: I1205 07:13:07.251974 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-4qdnj" event={"ID":"5e7b1770-04d8-41e3-bdfa-cafc699d03e1","Type":"ContainerStarted","Data":"523b80efebec409daab4c19aa1e7a06f7e1540bb2c90792a9c156e6e89c390c6"} Dec 05 07:13:07 crc kubenswrapper[4997]: I1205 07:13:07.252039 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-4qdnj" event={"ID":"5e7b1770-04d8-41e3-bdfa-cafc699d03e1","Type":"ContainerStarted","Data":"e4df28a74d4e95e692d2fc079bd4a9deb57be76f556425f59b4ba8ebeab5d0d1"} Dec 05 07:13:07 crc kubenswrapper[4997]: I1205 07:13:07.254977 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6" event={"ID":"79868cb5-be74-451c-8260-04bca75ded8c","Type":"ContainerStarted","Data":"2e9e40c825fea3c6513f634bfbb89d2ce503dc4dc824ec45bfe55808a375e235"} Dec 05 07:13:07 crc kubenswrapper[4997]: I1205 07:13:07.389274 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c0d286d5-b144-44ef-9696-21777bfad09b-memberlist\") pod \"speaker-sdv46\" (UID: \"c0d286d5-b144-44ef-9696-21777bfad09b\") " pod="metallb-system/speaker-sdv46" Dec 05 07:13:07 crc kubenswrapper[4997]: I1205 07:13:07.396480 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c0d286d5-b144-44ef-9696-21777bfad09b-memberlist\") pod \"speaker-sdv46\" (UID: \"c0d286d5-b144-44ef-9696-21777bfad09b\") " pod="metallb-system/speaker-sdv46" Dec 05 07:13:07 crc kubenswrapper[4997]: I1205 07:13:07.458503 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-sdv46" Dec 05 07:13:08 crc kubenswrapper[4997]: I1205 07:13:08.297300 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-sdv46" event={"ID":"c0d286d5-b144-44ef-9696-21777bfad09b","Type":"ContainerStarted","Data":"ab142fc7d92701742a07fbe149b4f4ce629315ea29940c33840dd64dd2bffcce"} Dec 05 07:13:08 crc kubenswrapper[4997]: I1205 07:13:08.297675 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-sdv46" event={"ID":"c0d286d5-b144-44ef-9696-21777bfad09b","Type":"ContainerStarted","Data":"1e90be4e903476d5a1b32af5fc52aec2a0076961108260115da9e0531e2bb4d7"} Dec 05 07:13:08 crc kubenswrapper[4997]: I1205 07:13:08.302987 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-4qdnj" event={"ID":"5e7b1770-04d8-41e3-bdfa-cafc699d03e1","Type":"ContainerStarted","Data":"3b5edaa81ce39c5e046a156cc065284e6b524d1c32608d4792e41789007a3ebd"} Dec 05 07:13:08 crc kubenswrapper[4997]: I1205 07:13:08.303782 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-4qdnj" Dec 05 07:13:09 crc kubenswrapper[4997]: I1205 07:13:09.315915 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-sdv46" event={"ID":"c0d286d5-b144-44ef-9696-21777bfad09b","Type":"ContainerStarted","Data":"eb6ff3aeacf47779b22f7065027999572e6fd11d1f351485a2e311d2efbdf025"} Dec 05 07:13:09 crc kubenswrapper[4997]: I1205 07:13:09.337228 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-sdv46" podStartSLOduration=4.337210186 podStartE2EDuration="4.337210186s" podCreationTimestamp="2025-12-05 07:13:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:13:09.334822433 +0000 UTC m=+1089.863729714" watchObservedRunningTime="2025-12-05 07:13:09.337210186 +0000 UTC m=+1089.866117447" Dec 05 07:13:09 crc kubenswrapper[4997]: I1205 07:13:09.338177 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-4qdnj" podStartSLOduration=4.338172693 podStartE2EDuration="4.338172693s" podCreationTimestamp="2025-12-05 07:13:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:13:08.350317227 +0000 UTC m=+1088.879224498" watchObservedRunningTime="2025-12-05 07:13:09.338172693 +0000 UTC m=+1089.867079954" Dec 05 07:13:10 crc kubenswrapper[4997]: I1205 07:13:10.324170 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-sdv46" Dec 05 07:13:17 crc kubenswrapper[4997]: I1205 07:13:17.463816 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-sdv46" Dec 05 07:13:19 crc kubenswrapper[4997]: I1205 07:13:19.774857 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p"] Dec 05 07:13:19 crc kubenswrapper[4997]: I1205 07:13:19.776587 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" Dec 05 07:13:19 crc kubenswrapper[4997]: I1205 07:13:19.779783 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 05 07:13:19 crc kubenswrapper[4997]: I1205 07:13:19.790249 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p"] Dec 05 07:13:19 crc kubenswrapper[4997]: I1205 07:13:19.908222 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5bbd9d2d-29b4-424d-8ad2-46f497377d53-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p\" (UID: \"5bbd9d2d-29b4-424d-8ad2-46f497377d53\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" Dec 05 07:13:19 crc kubenswrapper[4997]: I1205 07:13:19.908283 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pcmp\" (UniqueName: \"kubernetes.io/projected/5bbd9d2d-29b4-424d-8ad2-46f497377d53-kube-api-access-5pcmp\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p\" (UID: \"5bbd9d2d-29b4-424d-8ad2-46f497377d53\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" Dec 05 07:13:19 crc kubenswrapper[4997]: I1205 07:13:19.908391 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5bbd9d2d-29b4-424d-8ad2-46f497377d53-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p\" (UID: \"5bbd9d2d-29b4-424d-8ad2-46f497377d53\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" Dec 05 07:13:20 crc kubenswrapper[4997]: I1205 07:13:20.027152 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5bbd9d2d-29b4-424d-8ad2-46f497377d53-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p\" (UID: \"5bbd9d2d-29b4-424d-8ad2-46f497377d53\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" Dec 05 07:13:20 crc kubenswrapper[4997]: I1205 07:13:20.027215 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pcmp\" (UniqueName: \"kubernetes.io/projected/5bbd9d2d-29b4-424d-8ad2-46f497377d53-kube-api-access-5pcmp\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p\" (UID: \"5bbd9d2d-29b4-424d-8ad2-46f497377d53\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" Dec 05 07:13:20 crc kubenswrapper[4997]: I1205 07:13:20.027257 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5bbd9d2d-29b4-424d-8ad2-46f497377d53-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p\" (UID: \"5bbd9d2d-29b4-424d-8ad2-46f497377d53\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" Dec 05 07:13:20 crc kubenswrapper[4997]: I1205 07:13:20.027791 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5bbd9d2d-29b4-424d-8ad2-46f497377d53-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p\" (UID: \"5bbd9d2d-29b4-424d-8ad2-46f497377d53\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" Dec 05 07:13:20 crc kubenswrapper[4997]: I1205 07:13:20.027960 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5bbd9d2d-29b4-424d-8ad2-46f497377d53-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p\" (UID: \"5bbd9d2d-29b4-424d-8ad2-46f497377d53\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" Dec 05 07:13:20 crc kubenswrapper[4997]: I1205 07:13:20.057936 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pcmp\" (UniqueName: \"kubernetes.io/projected/5bbd9d2d-29b4-424d-8ad2-46f497377d53-kube-api-access-5pcmp\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p\" (UID: \"5bbd9d2d-29b4-424d-8ad2-46f497377d53\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" Dec 05 07:13:20 crc kubenswrapper[4997]: I1205 07:13:20.108238 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" Dec 05 07:13:23 crc kubenswrapper[4997]: E1205 07:13:23.888577 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/frr-rhel9@sha256:e5c5e7ca4ed54c9edba5dfa1d504bbe58016c2abdc872ebb8b26a628958e5a2a" Dec 05 07:13:23 crc kubenswrapper[4997]: E1205 07:13:23.889954 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:cp-frr-files,Image:registry.redhat.io/openshift4/frr-rhel9@sha256:e5c5e7ca4ed54c9edba5dfa1d504bbe58016c2abdc872ebb8b26a628958e5a2a,Command:[/bin/sh -c cp -rLf /tmp/frr/* /etc/frr/],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:frr-startup,ReadOnly:false,MountPath:/tmp/frr,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:frr-conf,ReadOnly:false,MountPath:/etc/frr,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qhjs9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*100,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*101,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod frr-k8s-2plrj_metallb-system(24697cfc-2c63-4b44-b4db-88a361476c5f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 07:13:23 crc kubenswrapper[4997]: E1205 07:13:23.891154 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cp-frr-files\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="metallb-system/frr-k8s-2plrj" podUID="24697cfc-2c63-4b44-b4db-88a361476c5f" Dec 05 07:13:23 crc kubenswrapper[4997]: E1205 07:13:23.914041 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/frr-rhel9@sha256:e5c5e7ca4ed54c9edba5dfa1d504bbe58016c2abdc872ebb8b26a628958e5a2a" Dec 05 07:13:23 crc kubenswrapper[4997]: E1205 07:13:23.914253 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:frr-k8s-webhook-server,Image:registry.redhat.io/openshift4/frr-rhel9@sha256:e5c5e7ca4ed54c9edba5dfa1d504bbe58016c2abdc872ebb8b26a628958e5a2a,Command:[/frr-k8s],Args:[--log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=$(NAMESPACE) --metrics-bind-address=:7572],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:monitoring,HostPort:0,ContainerPort:7572,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-khxx8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/metrics,Port:{1 0 monitoring},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/metrics,Port:{1 0 monitoring},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000690000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod frr-k8s-webhook-server-7fcb986d4-c9zt6_metallb-system(79868cb5-be74-451c-8260-04bca75ded8c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 07:13:23 crc kubenswrapper[4997]: E1205 07:13:23.915504 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"frr-k8s-webhook-server\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6" podUID="79868cb5-be74-451c-8260-04bca75ded8c" Dec 05 07:13:24 crc kubenswrapper[4997]: I1205 07:13:24.202991 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p"] Dec 05 07:13:24 crc kubenswrapper[4997]: I1205 07:13:24.687592 4997 generic.go:334] "Generic (PLEG): container finished" podID="5bbd9d2d-29b4-424d-8ad2-46f497377d53" containerID="80988b2c1939397befaffdfc2c6a70f2cf96dd722ec2c07d59faa6d02003657d" exitCode=0 Dec 05 07:13:24 crc kubenswrapper[4997]: I1205 07:13:24.687725 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" event={"ID":"5bbd9d2d-29b4-424d-8ad2-46f497377d53","Type":"ContainerDied","Data":"80988b2c1939397befaffdfc2c6a70f2cf96dd722ec2c07d59faa6d02003657d"} Dec 05 07:13:24 crc kubenswrapper[4997]: I1205 07:13:24.688160 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" event={"ID":"5bbd9d2d-29b4-424d-8ad2-46f497377d53","Type":"ContainerStarted","Data":"6ce74a894b0f2ec4f34e6139127d446ec6c42eaf16eb7e4fba0a5b7baaf98260"} Dec 05 07:13:24 crc kubenswrapper[4997]: E1205 07:13:24.689535 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cp-frr-files\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/frr-rhel9@sha256:e5c5e7ca4ed54c9edba5dfa1d504bbe58016c2abdc872ebb8b26a628958e5a2a\\\"\"" pod="metallb-system/frr-k8s-2plrj" podUID="24697cfc-2c63-4b44-b4db-88a361476c5f" Dec 05 07:13:24 crc kubenswrapper[4997]: E1205 07:13:24.689594 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"frr-k8s-webhook-server\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/frr-rhel9@sha256:e5c5e7ca4ed54c9edba5dfa1d504bbe58016c2abdc872ebb8b26a628958e5a2a\\\"\"" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6" podUID="79868cb5-be74-451c-8260-04bca75ded8c" Dec 05 07:13:26 crc kubenswrapper[4997]: I1205 07:13:26.545859 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-4qdnj" Dec 05 07:13:28 crc kubenswrapper[4997]: I1205 07:13:28.715353 4997 generic.go:334] "Generic (PLEG): container finished" podID="5bbd9d2d-29b4-424d-8ad2-46f497377d53" containerID="41e47a10e754a3d6b11ece0680ed2c760621c9815c1a635915cb26cda708f551" exitCode=0 Dec 05 07:13:28 crc kubenswrapper[4997]: I1205 07:13:28.715722 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" event={"ID":"5bbd9d2d-29b4-424d-8ad2-46f497377d53","Type":"ContainerDied","Data":"41e47a10e754a3d6b11ece0680ed2c760621c9815c1a635915cb26cda708f551"} Dec 05 07:13:29 crc kubenswrapper[4997]: I1205 07:13:29.727480 4997 generic.go:334] "Generic (PLEG): container finished" podID="5bbd9d2d-29b4-424d-8ad2-46f497377d53" containerID="91913a182a300beb3f6a26c9e388f6d4cfa929107de9a5cdd0571ad020e33360" exitCode=0 Dec 05 07:13:29 crc kubenswrapper[4997]: I1205 07:13:29.727662 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" event={"ID":"5bbd9d2d-29b4-424d-8ad2-46f497377d53","Type":"ContainerDied","Data":"91913a182a300beb3f6a26c9e388f6d4cfa929107de9a5cdd0571ad020e33360"} Dec 05 07:13:31 crc kubenswrapper[4997]: I1205 07:13:31.079608 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" Dec 05 07:13:31 crc kubenswrapper[4997]: I1205 07:13:31.181560 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pcmp\" (UniqueName: \"kubernetes.io/projected/5bbd9d2d-29b4-424d-8ad2-46f497377d53-kube-api-access-5pcmp\") pod \"5bbd9d2d-29b4-424d-8ad2-46f497377d53\" (UID: \"5bbd9d2d-29b4-424d-8ad2-46f497377d53\") " Dec 05 07:13:31 crc kubenswrapper[4997]: I1205 07:13:31.181992 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5bbd9d2d-29b4-424d-8ad2-46f497377d53-util\") pod \"5bbd9d2d-29b4-424d-8ad2-46f497377d53\" (UID: \"5bbd9d2d-29b4-424d-8ad2-46f497377d53\") " Dec 05 07:13:31 crc kubenswrapper[4997]: I1205 07:13:31.182185 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5bbd9d2d-29b4-424d-8ad2-46f497377d53-bundle\") pod \"5bbd9d2d-29b4-424d-8ad2-46f497377d53\" (UID: \"5bbd9d2d-29b4-424d-8ad2-46f497377d53\") " Dec 05 07:13:31 crc kubenswrapper[4997]: I1205 07:13:31.183236 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bbd9d2d-29b4-424d-8ad2-46f497377d53-bundle" (OuterVolumeSpecName: "bundle") pod "5bbd9d2d-29b4-424d-8ad2-46f497377d53" (UID: "5bbd9d2d-29b4-424d-8ad2-46f497377d53"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:13:31 crc kubenswrapper[4997]: I1205 07:13:31.183916 4997 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5bbd9d2d-29b4-424d-8ad2-46f497377d53-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:13:31 crc kubenswrapper[4997]: I1205 07:13:31.190945 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bbd9d2d-29b4-424d-8ad2-46f497377d53-kube-api-access-5pcmp" (OuterVolumeSpecName: "kube-api-access-5pcmp") pod "5bbd9d2d-29b4-424d-8ad2-46f497377d53" (UID: "5bbd9d2d-29b4-424d-8ad2-46f497377d53"). InnerVolumeSpecName "kube-api-access-5pcmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:13:31 crc kubenswrapper[4997]: I1205 07:13:31.222840 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bbd9d2d-29b4-424d-8ad2-46f497377d53-util" (OuterVolumeSpecName: "util") pod "5bbd9d2d-29b4-424d-8ad2-46f497377d53" (UID: "5bbd9d2d-29b4-424d-8ad2-46f497377d53"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:13:31 crc kubenswrapper[4997]: I1205 07:13:31.285165 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pcmp\" (UniqueName: \"kubernetes.io/projected/5bbd9d2d-29b4-424d-8ad2-46f497377d53-kube-api-access-5pcmp\") on node \"crc\" DevicePath \"\"" Dec 05 07:13:31 crc kubenswrapper[4997]: I1205 07:13:31.285197 4997 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5bbd9d2d-29b4-424d-8ad2-46f497377d53-util\") on node \"crc\" DevicePath \"\"" Dec 05 07:13:31 crc kubenswrapper[4997]: I1205 07:13:31.743010 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" event={"ID":"5bbd9d2d-29b4-424d-8ad2-46f497377d53","Type":"ContainerDied","Data":"6ce74a894b0f2ec4f34e6139127d446ec6c42eaf16eb7e4fba0a5b7baaf98260"} Dec 05 07:13:31 crc kubenswrapper[4997]: I1205 07:13:31.743462 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ce74a894b0f2ec4f34e6139127d446ec6c42eaf16eb7e4fba0a5b7baaf98260" Dec 05 07:13:31 crc kubenswrapper[4997]: I1205 07:13:31.743104 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p" Dec 05 07:13:38 crc kubenswrapper[4997]: I1205 07:13:38.324211 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fvmkb"] Dec 05 07:13:38 crc kubenswrapper[4997]: E1205 07:13:38.325356 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bbd9d2d-29b4-424d-8ad2-46f497377d53" containerName="util" Dec 05 07:13:38 crc kubenswrapper[4997]: I1205 07:13:38.325374 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bbd9d2d-29b4-424d-8ad2-46f497377d53" containerName="util" Dec 05 07:13:38 crc kubenswrapper[4997]: E1205 07:13:38.325388 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bbd9d2d-29b4-424d-8ad2-46f497377d53" containerName="extract" Dec 05 07:13:38 crc kubenswrapper[4997]: I1205 07:13:38.325394 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bbd9d2d-29b4-424d-8ad2-46f497377d53" containerName="extract" Dec 05 07:13:38 crc kubenswrapper[4997]: E1205 07:13:38.325411 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bbd9d2d-29b4-424d-8ad2-46f497377d53" containerName="pull" Dec 05 07:13:38 crc kubenswrapper[4997]: I1205 07:13:38.325418 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bbd9d2d-29b4-424d-8ad2-46f497377d53" containerName="pull" Dec 05 07:13:38 crc kubenswrapper[4997]: I1205 07:13:38.325527 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bbd9d2d-29b4-424d-8ad2-46f497377d53" containerName="extract" Dec 05 07:13:38 crc kubenswrapper[4997]: I1205 07:13:38.326008 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fvmkb" Dec 05 07:13:38 crc kubenswrapper[4997]: I1205 07:13:38.329041 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Dec 05 07:13:38 crc kubenswrapper[4997]: I1205 07:13:38.329107 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Dec 05 07:13:38 crc kubenswrapper[4997]: I1205 07:13:38.329709 4997 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-9q2kq" Dec 05 07:13:38 crc kubenswrapper[4997]: I1205 07:13:38.329928 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ec59d9ce-18f1-4019-9ebf-e10dcd710ce1-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fvmkb\" (UID: \"ec59d9ce-18f1-4019-9ebf-e10dcd710ce1\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fvmkb" Dec 05 07:13:38 crc kubenswrapper[4997]: I1205 07:13:38.330008 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r5fn\" (UniqueName: \"kubernetes.io/projected/ec59d9ce-18f1-4019-9ebf-e10dcd710ce1-kube-api-access-2r5fn\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fvmkb\" (UID: \"ec59d9ce-18f1-4019-9ebf-e10dcd710ce1\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fvmkb" Dec 05 07:13:38 crc kubenswrapper[4997]: I1205 07:13:38.361724 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fvmkb"] Dec 05 07:13:38 crc kubenswrapper[4997]: I1205 07:13:38.431355 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ec59d9ce-18f1-4019-9ebf-e10dcd710ce1-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fvmkb\" (UID: \"ec59d9ce-18f1-4019-9ebf-e10dcd710ce1\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fvmkb" Dec 05 07:13:38 crc kubenswrapper[4997]: I1205 07:13:38.431428 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r5fn\" (UniqueName: \"kubernetes.io/projected/ec59d9ce-18f1-4019-9ebf-e10dcd710ce1-kube-api-access-2r5fn\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fvmkb\" (UID: \"ec59d9ce-18f1-4019-9ebf-e10dcd710ce1\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fvmkb" Dec 05 07:13:38 crc kubenswrapper[4997]: I1205 07:13:38.432472 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ec59d9ce-18f1-4019-9ebf-e10dcd710ce1-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fvmkb\" (UID: \"ec59d9ce-18f1-4019-9ebf-e10dcd710ce1\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fvmkb" Dec 05 07:13:38 crc kubenswrapper[4997]: I1205 07:13:38.483585 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r5fn\" (UniqueName: \"kubernetes.io/projected/ec59d9ce-18f1-4019-9ebf-e10dcd710ce1-kube-api-access-2r5fn\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fvmkb\" (UID: \"ec59d9ce-18f1-4019-9ebf-e10dcd710ce1\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fvmkb" Dec 05 07:13:38 crc kubenswrapper[4997]: I1205 07:13:38.644060 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fvmkb" Dec 05 07:13:39 crc kubenswrapper[4997]: I1205 07:13:39.147421 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fvmkb"] Dec 05 07:13:39 crc kubenswrapper[4997]: I1205 07:13:39.795247 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fvmkb" event={"ID":"ec59d9ce-18f1-4019-9ebf-e10dcd710ce1","Type":"ContainerStarted","Data":"7fa1426eadd482a88538219a6f7034c2b7be5e136cd42998c01ce569bc403e44"} Dec 05 07:13:42 crc kubenswrapper[4997]: I1205 07:13:42.816279 4997 generic.go:334] "Generic (PLEG): container finished" podID="24697cfc-2c63-4b44-b4db-88a361476c5f" containerID="5a878f76bf89c1370aeea68a3669b6a56f08f193caeb0d312833a08c7964b0ea" exitCode=0 Dec 05 07:13:42 crc kubenswrapper[4997]: I1205 07:13:42.816361 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2plrj" event={"ID":"24697cfc-2c63-4b44-b4db-88a361476c5f","Type":"ContainerDied","Data":"5a878f76bf89c1370aeea68a3669b6a56f08f193caeb0d312833a08c7964b0ea"} Dec 05 07:13:43 crc kubenswrapper[4997]: I1205 07:13:43.828927 4997 generic.go:334] "Generic (PLEG): container finished" podID="24697cfc-2c63-4b44-b4db-88a361476c5f" containerID="ef1bfd44754f6d3486bf36ffbde321a598b8f75e728530fcc1fbb623e9bfa713" exitCode=0 Dec 05 07:13:43 crc kubenswrapper[4997]: I1205 07:13:43.829014 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2plrj" event={"ID":"24697cfc-2c63-4b44-b4db-88a361476c5f","Type":"ContainerDied","Data":"ef1bfd44754f6d3486bf36ffbde321a598b8f75e728530fcc1fbb623e9bfa713"} Dec 05 07:13:43 crc kubenswrapper[4997]: I1205 07:13:43.843290 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6" event={"ID":"79868cb5-be74-451c-8260-04bca75ded8c","Type":"ContainerStarted","Data":"0378e801c616a646d8b014a1fff4644b2b5ba06f83dbaa0c605bc832d7644990"} Dec 05 07:13:43 crc kubenswrapper[4997]: I1205 07:13:43.843592 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6" Dec 05 07:13:43 crc kubenswrapper[4997]: I1205 07:13:43.870942 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6" podStartSLOduration=2.988012784 podStartE2EDuration="38.870896411s" podCreationTimestamp="2025-12-05 07:13:05 +0000 UTC" firstStartedPulling="2025-12-05 07:13:06.694429835 +0000 UTC m=+1087.223337096" lastFinishedPulling="2025-12-05 07:13:42.577313462 +0000 UTC m=+1123.106220723" observedRunningTime="2025-12-05 07:13:43.869594437 +0000 UTC m=+1124.398501708" watchObservedRunningTime="2025-12-05 07:13:43.870896411 +0000 UTC m=+1124.399803672" Dec 05 07:13:44 crc kubenswrapper[4997]: I1205 07:13:44.851894 4997 generic.go:334] "Generic (PLEG): container finished" podID="24697cfc-2c63-4b44-b4db-88a361476c5f" containerID="6fd2abe8c4289ce42facf75f82d13b19b24a2fdd4495d26289acabe5689e2dad" exitCode=0 Dec 05 07:13:44 crc kubenswrapper[4997]: I1205 07:13:44.851977 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2plrj" event={"ID":"24697cfc-2c63-4b44-b4db-88a361476c5f","Type":"ContainerDied","Data":"6fd2abe8c4289ce42facf75f82d13b19b24a2fdd4495d26289acabe5689e2dad"} Dec 05 07:13:45 crc kubenswrapper[4997]: I1205 07:13:45.868747 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2plrj" event={"ID":"24697cfc-2c63-4b44-b4db-88a361476c5f","Type":"ContainerStarted","Data":"7c86b3fa451fc1f553178ce1b6438c30ba0d93ba5ec8a1d88009da7954c8f91c"} Dec 05 07:13:45 crc kubenswrapper[4997]: I1205 07:13:45.869183 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2plrj" event={"ID":"24697cfc-2c63-4b44-b4db-88a361476c5f","Type":"ContainerStarted","Data":"4a8cf19348b8d00e53aee355396b531e7ed3cda76d39d5fa2ba672de6ddc1d87"} Dec 05 07:13:45 crc kubenswrapper[4997]: I1205 07:13:45.869196 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2plrj" event={"ID":"24697cfc-2c63-4b44-b4db-88a361476c5f","Type":"ContainerStarted","Data":"15559b85c3d99ad4e703338846fc7368d89b2807d09fb4300abf009c8d1e2a1d"} Dec 05 07:13:51 crc kubenswrapper[4997]: I1205 07:13:51.917192 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fvmkb" event={"ID":"ec59d9ce-18f1-4019-9ebf-e10dcd710ce1","Type":"ContainerStarted","Data":"335fce224b1aa0cc588bfe53a4e1a2650b4c2ac37dd25aed1ef9968dcbc3b881"} Dec 05 07:13:51 crc kubenswrapper[4997]: I1205 07:13:51.923123 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2plrj" event={"ID":"24697cfc-2c63-4b44-b4db-88a361476c5f","Type":"ContainerStarted","Data":"44200e6d50a9e64aad59b5efd382639f41deb7460f69c2f144371485b7a01d47"} Dec 05 07:13:51 crc kubenswrapper[4997]: I1205 07:13:51.923218 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2plrj" event={"ID":"24697cfc-2c63-4b44-b4db-88a361476c5f","Type":"ContainerStarted","Data":"a4c270300464364e175c2a919747004ee8d7dd5f1c5a5a80e8b56dca1d4919aa"} Dec 05 07:13:51 crc kubenswrapper[4997]: I1205 07:13:51.923249 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2plrj" event={"ID":"24697cfc-2c63-4b44-b4db-88a361476c5f","Type":"ContainerStarted","Data":"3ae0342c92843958744c93003ff417ee19335b539b60bdbf25d98df6e4d6f3c0"} Dec 05 07:13:51 crc kubenswrapper[4997]: I1205 07:13:51.923305 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:51 crc kubenswrapper[4997]: I1205 07:13:51.944384 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fvmkb" podStartSLOduration=2.017711512 podStartE2EDuration="13.944347607s" podCreationTimestamp="2025-12-05 07:13:38 +0000 UTC" firstStartedPulling="2025-12-05 07:13:39.167054813 +0000 UTC m=+1119.695962074" lastFinishedPulling="2025-12-05 07:13:51.093690908 +0000 UTC m=+1131.622598169" observedRunningTime="2025-12-05 07:13:51.939216069 +0000 UTC m=+1132.468123360" watchObservedRunningTime="2025-12-05 07:13:51.944347607 +0000 UTC m=+1132.473254868" Dec 05 07:13:54 crc kubenswrapper[4997]: I1205 07:13:54.774749 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-2plrj" podStartSLOduration=14.072661255 podStartE2EDuration="49.77473097s" podCreationTimestamp="2025-12-05 07:13:05 +0000 UTC" firstStartedPulling="2025-12-05 07:13:06.572501416 +0000 UTC m=+1087.101408677" lastFinishedPulling="2025-12-05 07:13:42.274571131 +0000 UTC m=+1122.803478392" observedRunningTime="2025-12-05 07:13:51.969593894 +0000 UTC m=+1132.498501175" watchObservedRunningTime="2025-12-05 07:13:54.77473097 +0000 UTC m=+1135.303638231" Dec 05 07:13:54 crc kubenswrapper[4997]: I1205 07:13:54.778865 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-8z9rx"] Dec 05 07:13:54 crc kubenswrapper[4997]: I1205 07:13:54.779733 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-8z9rx" Dec 05 07:13:54 crc kubenswrapper[4997]: I1205 07:13:54.781853 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 05 07:13:54 crc kubenswrapper[4997]: I1205 07:13:54.783162 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 05 07:13:54 crc kubenswrapper[4997]: I1205 07:13:54.787915 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-8z9rx"] Dec 05 07:13:54 crc kubenswrapper[4997]: I1205 07:13:54.789764 4997 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-cqw4v" Dec 05 07:13:54 crc kubenswrapper[4997]: I1205 07:13:54.957449 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/849d3f8f-a64b-4b10-a847-d9305ed4c13b-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-8z9rx\" (UID: \"849d3f8f-a64b-4b10-a847-d9305ed4c13b\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-8z9rx" Dec 05 07:13:54 crc kubenswrapper[4997]: I1205 07:13:54.957497 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgfsw\" (UniqueName: \"kubernetes.io/projected/849d3f8f-a64b-4b10-a847-d9305ed4c13b-kube-api-access-hgfsw\") pod \"cert-manager-webhook-f4fb5df64-8z9rx\" (UID: \"849d3f8f-a64b-4b10-a847-d9305ed4c13b\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-8z9rx" Dec 05 07:13:55 crc kubenswrapper[4997]: I1205 07:13:55.059022 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/849d3f8f-a64b-4b10-a847-d9305ed4c13b-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-8z9rx\" (UID: \"849d3f8f-a64b-4b10-a847-d9305ed4c13b\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-8z9rx" Dec 05 07:13:55 crc kubenswrapper[4997]: I1205 07:13:55.059083 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgfsw\" (UniqueName: \"kubernetes.io/projected/849d3f8f-a64b-4b10-a847-d9305ed4c13b-kube-api-access-hgfsw\") pod \"cert-manager-webhook-f4fb5df64-8z9rx\" (UID: \"849d3f8f-a64b-4b10-a847-d9305ed4c13b\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-8z9rx" Dec 05 07:13:55 crc kubenswrapper[4997]: I1205 07:13:55.088315 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/849d3f8f-a64b-4b10-a847-d9305ed4c13b-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-8z9rx\" (UID: \"849d3f8f-a64b-4b10-a847-d9305ed4c13b\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-8z9rx" Dec 05 07:13:55 crc kubenswrapper[4997]: I1205 07:13:55.088392 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgfsw\" (UniqueName: \"kubernetes.io/projected/849d3f8f-a64b-4b10-a847-d9305ed4c13b-kube-api-access-hgfsw\") pod \"cert-manager-webhook-f4fb5df64-8z9rx\" (UID: \"849d3f8f-a64b-4b10-a847-d9305ed4c13b\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-8z9rx" Dec 05 07:13:55 crc kubenswrapper[4997]: I1205 07:13:55.097513 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-8z9rx" Dec 05 07:13:55 crc kubenswrapper[4997]: I1205 07:13:55.588142 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-8z9rx"] Dec 05 07:13:55 crc kubenswrapper[4997]: W1205 07:13:55.600036 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod849d3f8f_a64b_4b10_a847_d9305ed4c13b.slice/crio-e281a96891687e0ada9ad55cf81e6786838bdd8f66ad04cdbb1d76c843883b94 WatchSource:0}: Error finding container e281a96891687e0ada9ad55cf81e6786838bdd8f66ad04cdbb1d76c843883b94: Status 404 returned error can't find the container with id e281a96891687e0ada9ad55cf81e6786838bdd8f66ad04cdbb1d76c843883b94 Dec 05 07:13:55 crc kubenswrapper[4997]: I1205 07:13:55.949473 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-8z9rx" event={"ID":"849d3f8f-a64b-4b10-a847-d9305ed4c13b","Type":"ContainerStarted","Data":"e281a96891687e0ada9ad55cf81e6786838bdd8f66ad04cdbb1d76c843883b94"} Dec 05 07:13:56 crc kubenswrapper[4997]: I1205 07:13:56.399470 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c9zt6" Dec 05 07:13:56 crc kubenswrapper[4997]: I1205 07:13:56.403337 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:56 crc kubenswrapper[4997]: I1205 07:13:56.406738 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:56 crc kubenswrapper[4997]: I1205 07:13:56.460278 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-2plrj" Dec 05 07:13:58 crc kubenswrapper[4997]: I1205 07:13:58.572256 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-qdq9n"] Dec 05 07:13:58 crc kubenswrapper[4997]: I1205 07:13:58.573524 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-qdq9n" Dec 05 07:13:58 crc kubenswrapper[4997]: I1205 07:13:58.575806 4997 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-ttvcl" Dec 05 07:13:58 crc kubenswrapper[4997]: I1205 07:13:58.580003 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-qdq9n"] Dec 05 07:13:58 crc kubenswrapper[4997]: I1205 07:13:58.622638 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d9e89c33-9251-4912-82c5-c37e1aec78bc-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-qdq9n\" (UID: \"d9e89c33-9251-4912-82c5-c37e1aec78bc\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-qdq9n" Dec 05 07:13:58 crc kubenswrapper[4997]: I1205 07:13:58.622726 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h95pv\" (UniqueName: \"kubernetes.io/projected/d9e89c33-9251-4912-82c5-c37e1aec78bc-kube-api-access-h95pv\") pod \"cert-manager-cainjector-855d9ccff4-qdq9n\" (UID: \"d9e89c33-9251-4912-82c5-c37e1aec78bc\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-qdq9n" Dec 05 07:13:58 crc kubenswrapper[4997]: I1205 07:13:58.724701 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h95pv\" (UniqueName: \"kubernetes.io/projected/d9e89c33-9251-4912-82c5-c37e1aec78bc-kube-api-access-h95pv\") pod \"cert-manager-cainjector-855d9ccff4-qdq9n\" (UID: \"d9e89c33-9251-4912-82c5-c37e1aec78bc\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-qdq9n" Dec 05 07:13:58 crc kubenswrapper[4997]: I1205 07:13:58.724807 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d9e89c33-9251-4912-82c5-c37e1aec78bc-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-qdq9n\" (UID: \"d9e89c33-9251-4912-82c5-c37e1aec78bc\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-qdq9n" Dec 05 07:13:58 crc kubenswrapper[4997]: I1205 07:13:58.745407 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d9e89c33-9251-4912-82c5-c37e1aec78bc-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-qdq9n\" (UID: \"d9e89c33-9251-4912-82c5-c37e1aec78bc\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-qdq9n" Dec 05 07:13:58 crc kubenswrapper[4997]: I1205 07:13:58.755137 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h95pv\" (UniqueName: \"kubernetes.io/projected/d9e89c33-9251-4912-82c5-c37e1aec78bc-kube-api-access-h95pv\") pod \"cert-manager-cainjector-855d9ccff4-qdq9n\" (UID: \"d9e89c33-9251-4912-82c5-c37e1aec78bc\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-qdq9n" Dec 05 07:13:58 crc kubenswrapper[4997]: I1205 07:13:58.893220 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-qdq9n" Dec 05 07:14:05 crc kubenswrapper[4997]: I1205 07:14:05.702630 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-p6wrp"] Dec 05 07:14:05 crc kubenswrapper[4997]: I1205 07:14:05.705589 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-p6wrp" Dec 05 07:14:05 crc kubenswrapper[4997]: I1205 07:14:05.708217 4997 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-k9rdl" Dec 05 07:14:05 crc kubenswrapper[4997]: I1205 07:14:05.710528 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-p6wrp"] Dec 05 07:14:05 crc kubenswrapper[4997]: I1205 07:14:05.732297 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bd787551-b226-4b7d-88db-785469feec70-bound-sa-token\") pod \"cert-manager-86cb77c54b-p6wrp\" (UID: \"bd787551-b226-4b7d-88db-785469feec70\") " pod="cert-manager/cert-manager-86cb77c54b-p6wrp" Dec 05 07:14:05 crc kubenswrapper[4997]: I1205 07:14:05.732384 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2td7p\" (UniqueName: \"kubernetes.io/projected/bd787551-b226-4b7d-88db-785469feec70-kube-api-access-2td7p\") pod \"cert-manager-86cb77c54b-p6wrp\" (UID: \"bd787551-b226-4b7d-88db-785469feec70\") " pod="cert-manager/cert-manager-86cb77c54b-p6wrp" Dec 05 07:14:05 crc kubenswrapper[4997]: I1205 07:14:05.833219 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bd787551-b226-4b7d-88db-785469feec70-bound-sa-token\") pod \"cert-manager-86cb77c54b-p6wrp\" (UID: \"bd787551-b226-4b7d-88db-785469feec70\") " pod="cert-manager/cert-manager-86cb77c54b-p6wrp" Dec 05 07:14:05 crc kubenswrapper[4997]: I1205 07:14:05.833296 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2td7p\" (UniqueName: \"kubernetes.io/projected/bd787551-b226-4b7d-88db-785469feec70-kube-api-access-2td7p\") pod \"cert-manager-86cb77c54b-p6wrp\" (UID: \"bd787551-b226-4b7d-88db-785469feec70\") " pod="cert-manager/cert-manager-86cb77c54b-p6wrp" Dec 05 07:14:05 crc kubenswrapper[4997]: I1205 07:14:05.855268 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2td7p\" (UniqueName: \"kubernetes.io/projected/bd787551-b226-4b7d-88db-785469feec70-kube-api-access-2td7p\") pod \"cert-manager-86cb77c54b-p6wrp\" (UID: \"bd787551-b226-4b7d-88db-785469feec70\") " pod="cert-manager/cert-manager-86cb77c54b-p6wrp" Dec 05 07:14:05 crc kubenswrapper[4997]: I1205 07:14:05.859902 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bd787551-b226-4b7d-88db-785469feec70-bound-sa-token\") pod \"cert-manager-86cb77c54b-p6wrp\" (UID: \"bd787551-b226-4b7d-88db-785469feec70\") " pod="cert-manager/cert-manager-86cb77c54b-p6wrp" Dec 05 07:14:06 crc kubenswrapper[4997]: I1205 07:14:06.031261 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-p6wrp" Dec 05 07:14:06 crc kubenswrapper[4997]: I1205 07:14:06.291504 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-qdq9n"] Dec 05 07:14:06 crc kubenswrapper[4997]: I1205 07:14:06.406434 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-p6wrp"] Dec 05 07:14:07 crc kubenswrapper[4997]: I1205 07:14:07.039558 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-8z9rx" event={"ID":"849d3f8f-a64b-4b10-a847-d9305ed4c13b","Type":"ContainerStarted","Data":"7f3de4ffa0b6fb45c61f1b3baaf13eb391b1c656876634028c9ef16ea9af592e"} Dec 05 07:14:07 crc kubenswrapper[4997]: I1205 07:14:07.040038 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-8z9rx" Dec 05 07:14:07 crc kubenswrapper[4997]: I1205 07:14:07.041876 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-qdq9n" event={"ID":"d9e89c33-9251-4912-82c5-c37e1aec78bc","Type":"ContainerStarted","Data":"4b3f7b885f73fc3c3d8eac317eb828ace7c640b29f61c4613d580bae6bb24466"} Dec 05 07:14:07 crc kubenswrapper[4997]: I1205 07:14:07.041920 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-qdq9n" event={"ID":"d9e89c33-9251-4912-82c5-c37e1aec78bc","Type":"ContainerStarted","Data":"eadb7bfb834f6bbd3288efcf88724ee1d689dc327cf7f50a8c64532a24b872bc"} Dec 05 07:14:07 crc kubenswrapper[4997]: I1205 07:14:07.045692 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-p6wrp" event={"ID":"bd787551-b226-4b7d-88db-785469feec70","Type":"ContainerStarted","Data":"f673a552eae6343c685e362e4d12c9510db5377bf96605acf6980c97bc9334c0"} Dec 05 07:14:07 crc kubenswrapper[4997]: I1205 07:14:07.045758 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-p6wrp" event={"ID":"bd787551-b226-4b7d-88db-785469feec70","Type":"ContainerStarted","Data":"1bac11237dc516d3a3eb2f433a6a31d263bc40a03f527250388b7408cb963f9b"} Dec 05 07:14:07 crc kubenswrapper[4997]: I1205 07:14:07.057683 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-8z9rx" podStartSLOduration=2.735912188 podStartE2EDuration="13.057658343s" podCreationTimestamp="2025-12-05 07:13:54 +0000 UTC" firstStartedPulling="2025-12-05 07:13:55.602306959 +0000 UTC m=+1136.131214220" lastFinishedPulling="2025-12-05 07:14:05.924053114 +0000 UTC m=+1146.452960375" observedRunningTime="2025-12-05 07:14:07.056662436 +0000 UTC m=+1147.585569707" watchObservedRunningTime="2025-12-05 07:14:07.057658343 +0000 UTC m=+1147.586565614" Dec 05 07:14:07 crc kubenswrapper[4997]: I1205 07:14:07.089192 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-p6wrp" podStartSLOduration=2.089155438 podStartE2EDuration="2.089155438s" podCreationTimestamp="2025-12-05 07:14:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:14:07.081849912 +0000 UTC m=+1147.610757193" watchObservedRunningTime="2025-12-05 07:14:07.089155438 +0000 UTC m=+1147.618062699" Dec 05 07:14:07 crc kubenswrapper[4997]: I1205 07:14:07.104852 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-qdq9n" podStartSLOduration=9.104832719 podStartE2EDuration="9.104832719s" podCreationTimestamp="2025-12-05 07:13:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:14:07.100241965 +0000 UTC m=+1147.629149286" watchObservedRunningTime="2025-12-05 07:14:07.104832719 +0000 UTC m=+1147.633739970" Dec 05 07:14:15 crc kubenswrapper[4997]: I1205 07:14:15.102103 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-8z9rx" Dec 05 07:14:18 crc kubenswrapper[4997]: I1205 07:14:18.556170 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-z7zql"] Dec 05 07:14:18 crc kubenswrapper[4997]: I1205 07:14:18.557302 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z7zql" Dec 05 07:14:18 crc kubenswrapper[4997]: I1205 07:14:18.559445 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 05 07:14:18 crc kubenswrapper[4997]: I1205 07:14:18.559552 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-rp74j" Dec 05 07:14:18 crc kubenswrapper[4997]: I1205 07:14:18.559757 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 05 07:14:18 crc kubenswrapper[4997]: I1205 07:14:18.577496 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-z7zql"] Dec 05 07:14:18 crc kubenswrapper[4997]: I1205 07:14:18.740832 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tgmt\" (UniqueName: \"kubernetes.io/projected/3c465f77-ff2b-4c8c-a16f-135d04c59337-kube-api-access-7tgmt\") pod \"openstack-operator-index-z7zql\" (UID: \"3c465f77-ff2b-4c8c-a16f-135d04c59337\") " pod="openstack-operators/openstack-operator-index-z7zql" Dec 05 07:14:18 crc kubenswrapper[4997]: I1205 07:14:18.842692 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tgmt\" (UniqueName: \"kubernetes.io/projected/3c465f77-ff2b-4c8c-a16f-135d04c59337-kube-api-access-7tgmt\") pod \"openstack-operator-index-z7zql\" (UID: \"3c465f77-ff2b-4c8c-a16f-135d04c59337\") " pod="openstack-operators/openstack-operator-index-z7zql" Dec 05 07:14:18 crc kubenswrapper[4997]: I1205 07:14:18.863360 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tgmt\" (UniqueName: \"kubernetes.io/projected/3c465f77-ff2b-4c8c-a16f-135d04c59337-kube-api-access-7tgmt\") pod \"openstack-operator-index-z7zql\" (UID: \"3c465f77-ff2b-4c8c-a16f-135d04c59337\") " pod="openstack-operators/openstack-operator-index-z7zql" Dec 05 07:14:18 crc kubenswrapper[4997]: I1205 07:14:18.879963 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z7zql" Dec 05 07:14:19 crc kubenswrapper[4997]: I1205 07:14:19.175348 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-z7zql"] Dec 05 07:14:20 crc kubenswrapper[4997]: I1205 07:14:20.153574 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z7zql" event={"ID":"3c465f77-ff2b-4c8c-a16f-135d04c59337","Type":"ContainerStarted","Data":"39dfe83232c5aac416da454528070c667ee4eb9dbaa20f343d289f638b388b99"} Dec 05 07:14:21 crc kubenswrapper[4997]: I1205 07:14:21.124954 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-z7zql"] Dec 05 07:14:21 crc kubenswrapper[4997]: I1205 07:14:21.163323 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z7zql" event={"ID":"3c465f77-ff2b-4c8c-a16f-135d04c59337","Type":"ContainerStarted","Data":"3fbffd05d163a51f98585cecec6c20ceef1f76c99612d3a24333e11e19aa51c2"} Dec 05 07:14:21 crc kubenswrapper[4997]: I1205 07:14:21.183097 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-z7zql" podStartSLOduration=1.566520987 podStartE2EDuration="3.183072041s" podCreationTimestamp="2025-12-05 07:14:18 +0000 UTC" firstStartedPulling="2025-12-05 07:14:19.184233342 +0000 UTC m=+1159.713140603" lastFinishedPulling="2025-12-05 07:14:20.800784396 +0000 UTC m=+1161.329691657" observedRunningTime="2025-12-05 07:14:21.180788819 +0000 UTC m=+1161.709696120" watchObservedRunningTime="2025-12-05 07:14:21.183072041 +0000 UTC m=+1161.711979302" Dec 05 07:14:21 crc kubenswrapper[4997]: I1205 07:14:21.735527 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-kc298"] Dec 05 07:14:21 crc kubenswrapper[4997]: I1205 07:14:21.736553 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kc298" Dec 05 07:14:21 crc kubenswrapper[4997]: I1205 07:14:21.759180 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kc298"] Dec 05 07:14:21 crc kubenswrapper[4997]: I1205 07:14:21.887787 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm8zr\" (UniqueName: \"kubernetes.io/projected/395fd5d1-b4de-4859-ba11-2f2fa89395d2-kube-api-access-bm8zr\") pod \"openstack-operator-index-kc298\" (UID: \"395fd5d1-b4de-4859-ba11-2f2fa89395d2\") " pod="openstack-operators/openstack-operator-index-kc298" Dec 05 07:14:21 crc kubenswrapper[4997]: I1205 07:14:21.989233 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm8zr\" (UniqueName: \"kubernetes.io/projected/395fd5d1-b4de-4859-ba11-2f2fa89395d2-kube-api-access-bm8zr\") pod \"openstack-operator-index-kc298\" (UID: \"395fd5d1-b4de-4859-ba11-2f2fa89395d2\") " pod="openstack-operators/openstack-operator-index-kc298" Dec 05 07:14:22 crc kubenswrapper[4997]: I1205 07:14:22.013864 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm8zr\" (UniqueName: \"kubernetes.io/projected/395fd5d1-b4de-4859-ba11-2f2fa89395d2-kube-api-access-bm8zr\") pod \"openstack-operator-index-kc298\" (UID: \"395fd5d1-b4de-4859-ba11-2f2fa89395d2\") " pod="openstack-operators/openstack-operator-index-kc298" Dec 05 07:14:22 crc kubenswrapper[4997]: I1205 07:14:22.060313 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kc298" Dec 05 07:14:22 crc kubenswrapper[4997]: I1205 07:14:22.174549 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-z7zql" podUID="3c465f77-ff2b-4c8c-a16f-135d04c59337" containerName="registry-server" containerID="cri-o://3fbffd05d163a51f98585cecec6c20ceef1f76c99612d3a24333e11e19aa51c2" gracePeriod=2 Dec 05 07:14:22 crc kubenswrapper[4997]: I1205 07:14:22.351769 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kc298"] Dec 05 07:14:22 crc kubenswrapper[4997]: I1205 07:14:22.520010 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z7zql" Dec 05 07:14:22 crc kubenswrapper[4997]: I1205 07:14:22.700128 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tgmt\" (UniqueName: \"kubernetes.io/projected/3c465f77-ff2b-4c8c-a16f-135d04c59337-kube-api-access-7tgmt\") pod \"3c465f77-ff2b-4c8c-a16f-135d04c59337\" (UID: \"3c465f77-ff2b-4c8c-a16f-135d04c59337\") " Dec 05 07:14:22 crc kubenswrapper[4997]: I1205 07:14:22.705008 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c465f77-ff2b-4c8c-a16f-135d04c59337-kube-api-access-7tgmt" (OuterVolumeSpecName: "kube-api-access-7tgmt") pod "3c465f77-ff2b-4c8c-a16f-135d04c59337" (UID: "3c465f77-ff2b-4c8c-a16f-135d04c59337"). InnerVolumeSpecName "kube-api-access-7tgmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:14:22 crc kubenswrapper[4997]: I1205 07:14:22.801960 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tgmt\" (UniqueName: \"kubernetes.io/projected/3c465f77-ff2b-4c8c-a16f-135d04c59337-kube-api-access-7tgmt\") on node \"crc\" DevicePath \"\"" Dec 05 07:14:23 crc kubenswrapper[4997]: I1205 07:14:23.180910 4997 generic.go:334] "Generic (PLEG): container finished" podID="3c465f77-ff2b-4c8c-a16f-135d04c59337" containerID="3fbffd05d163a51f98585cecec6c20ceef1f76c99612d3a24333e11e19aa51c2" exitCode=0 Dec 05 07:14:23 crc kubenswrapper[4997]: I1205 07:14:23.181016 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z7zql" event={"ID":"3c465f77-ff2b-4c8c-a16f-135d04c59337","Type":"ContainerDied","Data":"3fbffd05d163a51f98585cecec6c20ceef1f76c99612d3a24333e11e19aa51c2"} Dec 05 07:14:23 crc kubenswrapper[4997]: I1205 07:14:23.181017 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z7zql" Dec 05 07:14:23 crc kubenswrapper[4997]: I1205 07:14:23.181069 4997 scope.go:117] "RemoveContainer" containerID="3fbffd05d163a51f98585cecec6c20ceef1f76c99612d3a24333e11e19aa51c2" Dec 05 07:14:23 crc kubenswrapper[4997]: I1205 07:14:23.181055 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z7zql" event={"ID":"3c465f77-ff2b-4c8c-a16f-135d04c59337","Type":"ContainerDied","Data":"39dfe83232c5aac416da454528070c667ee4eb9dbaa20f343d289f638b388b99"} Dec 05 07:14:23 crc kubenswrapper[4997]: I1205 07:14:23.182483 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kc298" event={"ID":"395fd5d1-b4de-4859-ba11-2f2fa89395d2","Type":"ContainerStarted","Data":"5fd4d350f6d92bded48d6224f235e78aa7e485f19c87d5294441de9123290978"} Dec 05 07:14:23 crc kubenswrapper[4997]: I1205 07:14:23.182533 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kc298" event={"ID":"395fd5d1-b4de-4859-ba11-2f2fa89395d2","Type":"ContainerStarted","Data":"cdeaa00724c25c6e6773c4ba917abcd94143addb6a212b2943cb331a3b77df60"} Dec 05 07:14:23 crc kubenswrapper[4997]: I1205 07:14:23.204430 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-kc298" podStartSLOduration=1.8006076389999999 podStartE2EDuration="2.20398956s" podCreationTimestamp="2025-12-05 07:14:21 +0000 UTC" firstStartedPulling="2025-12-05 07:14:22.360980537 +0000 UTC m=+1162.889887798" lastFinishedPulling="2025-12-05 07:14:22.764362458 +0000 UTC m=+1163.293269719" observedRunningTime="2025-12-05 07:14:23.199077708 +0000 UTC m=+1163.727984979" watchObservedRunningTime="2025-12-05 07:14:23.20398956 +0000 UTC m=+1163.732896811" Dec 05 07:14:23 crc kubenswrapper[4997]: I1205 07:14:23.205021 4997 scope.go:117] "RemoveContainer" containerID="3fbffd05d163a51f98585cecec6c20ceef1f76c99612d3a24333e11e19aa51c2" Dec 05 07:14:23 crc kubenswrapper[4997]: E1205 07:14:23.205575 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fbffd05d163a51f98585cecec6c20ceef1f76c99612d3a24333e11e19aa51c2\": container with ID starting with 3fbffd05d163a51f98585cecec6c20ceef1f76c99612d3a24333e11e19aa51c2 not found: ID does not exist" containerID="3fbffd05d163a51f98585cecec6c20ceef1f76c99612d3a24333e11e19aa51c2" Dec 05 07:14:23 crc kubenswrapper[4997]: I1205 07:14:23.205681 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fbffd05d163a51f98585cecec6c20ceef1f76c99612d3a24333e11e19aa51c2"} err="failed to get container status \"3fbffd05d163a51f98585cecec6c20ceef1f76c99612d3a24333e11e19aa51c2\": rpc error: code = NotFound desc = could not find container \"3fbffd05d163a51f98585cecec6c20ceef1f76c99612d3a24333e11e19aa51c2\": container with ID starting with 3fbffd05d163a51f98585cecec6c20ceef1f76c99612d3a24333e11e19aa51c2 not found: ID does not exist" Dec 05 07:14:23 crc kubenswrapper[4997]: I1205 07:14:23.227456 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-z7zql"] Dec 05 07:14:23 crc kubenswrapper[4997]: I1205 07:14:23.231947 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-z7zql"] Dec 05 07:14:23 crc kubenswrapper[4997]: I1205 07:14:23.756840 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c465f77-ff2b-4c8c-a16f-135d04c59337" path="/var/lib/kubelet/pods/3c465f77-ff2b-4c8c-a16f-135d04c59337/volumes" Dec 05 07:14:32 crc kubenswrapper[4997]: I1205 07:14:32.061797 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-kc298" Dec 05 07:14:32 crc kubenswrapper[4997]: I1205 07:14:32.064983 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-kc298" Dec 05 07:14:32 crc kubenswrapper[4997]: I1205 07:14:32.105154 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-kc298" Dec 05 07:14:32 crc kubenswrapper[4997]: I1205 07:14:32.287689 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-kc298" Dec 05 07:14:32 crc kubenswrapper[4997]: I1205 07:14:32.979267 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9"] Dec 05 07:14:32 crc kubenswrapper[4997]: E1205 07:14:32.979653 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c465f77-ff2b-4c8c-a16f-135d04c59337" containerName="registry-server" Dec 05 07:14:32 crc kubenswrapper[4997]: I1205 07:14:32.979671 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c465f77-ff2b-4c8c-a16f-135d04c59337" containerName="registry-server" Dec 05 07:14:32 crc kubenswrapper[4997]: I1205 07:14:32.979808 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c465f77-ff2b-4c8c-a16f-135d04c59337" containerName="registry-server" Dec 05 07:14:32 crc kubenswrapper[4997]: I1205 07:14:32.980905 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" Dec 05 07:14:32 crc kubenswrapper[4997]: I1205 07:14:32.983108 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-vh6ds" Dec 05 07:14:32 crc kubenswrapper[4997]: I1205 07:14:32.997347 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9"] Dec 05 07:14:33 crc kubenswrapper[4997]: I1205 07:14:33.159408 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-bundle\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9\" (UID: \"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" Dec 05 07:14:33 crc kubenswrapper[4997]: I1205 07:14:33.159527 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-util\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9\" (UID: \"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" Dec 05 07:14:33 crc kubenswrapper[4997]: I1205 07:14:33.159565 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsrmr\" (UniqueName: \"kubernetes.io/projected/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-kube-api-access-xsrmr\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9\" (UID: \"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" Dec 05 07:14:33 crc kubenswrapper[4997]: I1205 07:14:33.261090 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-util\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9\" (UID: \"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" Dec 05 07:14:33 crc kubenswrapper[4997]: I1205 07:14:33.261639 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsrmr\" (UniqueName: \"kubernetes.io/projected/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-kube-api-access-xsrmr\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9\" (UID: \"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" Dec 05 07:14:33 crc kubenswrapper[4997]: I1205 07:14:33.261934 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-bundle\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9\" (UID: \"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" Dec 05 07:14:33 crc kubenswrapper[4997]: I1205 07:14:33.262296 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-util\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9\" (UID: \"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" Dec 05 07:14:33 crc kubenswrapper[4997]: I1205 07:14:33.262468 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-bundle\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9\" (UID: \"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" Dec 05 07:14:33 crc kubenswrapper[4997]: I1205 07:14:33.289454 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsrmr\" (UniqueName: \"kubernetes.io/projected/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-kube-api-access-xsrmr\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9\" (UID: \"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" Dec 05 07:14:33 crc kubenswrapper[4997]: I1205 07:14:33.346844 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" Dec 05 07:14:33 crc kubenswrapper[4997]: I1205 07:14:33.576090 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9"] Dec 05 07:14:34 crc kubenswrapper[4997]: I1205 07:14:34.275101 4997 generic.go:334] "Generic (PLEG): container finished" podID="aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5" containerID="2bbef53a634ba97712f62f5369d9a3d2602bf91cbe60cfe48cf0f2263efe5047" exitCode=0 Dec 05 07:14:34 crc kubenswrapper[4997]: I1205 07:14:34.275193 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" event={"ID":"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5","Type":"ContainerDied","Data":"2bbef53a634ba97712f62f5369d9a3d2602bf91cbe60cfe48cf0f2263efe5047"} Dec 05 07:14:34 crc kubenswrapper[4997]: I1205 07:14:34.275623 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" event={"ID":"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5","Type":"ContainerStarted","Data":"6856bf6f55425fbe18d484fc0e59837ef9021d3ea76776e4207eaf236a79f457"} Dec 05 07:14:36 crc kubenswrapper[4997]: I1205 07:14:36.289272 4997 generic.go:334] "Generic (PLEG): container finished" podID="aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5" containerID="65104259abef34881393ecdff69c6e1d291c320a5bbf19d7daf95869e68305e4" exitCode=0 Dec 05 07:14:36 crc kubenswrapper[4997]: I1205 07:14:36.289325 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" event={"ID":"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5","Type":"ContainerDied","Data":"65104259abef34881393ecdff69c6e1d291c320a5bbf19d7daf95869e68305e4"} Dec 05 07:14:37 crc kubenswrapper[4997]: I1205 07:14:37.299784 4997 generic.go:334] "Generic (PLEG): container finished" podID="aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5" containerID="6b54e22b12a66b89844c8c9de02fbbccee8d7508fe925df3f48c0a42326da176" exitCode=0 Dec 05 07:14:37 crc kubenswrapper[4997]: I1205 07:14:37.299883 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" event={"ID":"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5","Type":"ContainerDied","Data":"6b54e22b12a66b89844c8c9de02fbbccee8d7508fe925df3f48c0a42326da176"} Dec 05 07:14:38 crc kubenswrapper[4997]: I1205 07:14:38.576661 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" Dec 05 07:14:38 crc kubenswrapper[4997]: I1205 07:14:38.663123 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-bundle\") pod \"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5\" (UID: \"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5\") " Dec 05 07:14:38 crc kubenswrapper[4997]: I1205 07:14:38.663483 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsrmr\" (UniqueName: \"kubernetes.io/projected/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-kube-api-access-xsrmr\") pod \"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5\" (UID: \"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5\") " Dec 05 07:14:38 crc kubenswrapper[4997]: I1205 07:14:38.663586 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-util\") pod \"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5\" (UID: \"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5\") " Dec 05 07:14:38 crc kubenswrapper[4997]: I1205 07:14:38.663932 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-bundle" (OuterVolumeSpecName: "bundle") pod "aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5" (UID: "aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:14:38 crc kubenswrapper[4997]: I1205 07:14:38.664445 4997 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:14:38 crc kubenswrapper[4997]: I1205 07:14:38.675763 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-kube-api-access-xsrmr" (OuterVolumeSpecName: "kube-api-access-xsrmr") pod "aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5" (UID: "aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5"). InnerVolumeSpecName "kube-api-access-xsrmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:14:38 crc kubenswrapper[4997]: I1205 07:14:38.683137 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-util" (OuterVolumeSpecName: "util") pod "aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5" (UID: "aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:14:38 crc kubenswrapper[4997]: I1205 07:14:38.767372 4997 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-util\") on node \"crc\" DevicePath \"\"" Dec 05 07:14:38 crc kubenswrapper[4997]: I1205 07:14:38.767415 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsrmr\" (UniqueName: \"kubernetes.io/projected/aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5-kube-api-access-xsrmr\") on node \"crc\" DevicePath \"\"" Dec 05 07:14:39 crc kubenswrapper[4997]: I1205 07:14:39.314345 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" event={"ID":"aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5","Type":"ContainerDied","Data":"6856bf6f55425fbe18d484fc0e59837ef9021d3ea76776e4207eaf236a79f457"} Dec 05 07:14:39 crc kubenswrapper[4997]: I1205 07:14:39.314407 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6856bf6f55425fbe18d484fc0e59837ef9021d3ea76776e4207eaf236a79f457" Dec 05 07:14:39 crc kubenswrapper[4997]: I1205 07:14:39.314378 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9" Dec 05 07:14:45 crc kubenswrapper[4997]: I1205 07:14:45.556110 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-259kk"] Dec 05 07:14:45 crc kubenswrapper[4997]: E1205 07:14:45.557280 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5" containerName="pull" Dec 05 07:14:45 crc kubenswrapper[4997]: I1205 07:14:45.557297 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5" containerName="pull" Dec 05 07:14:45 crc kubenswrapper[4997]: E1205 07:14:45.557308 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5" containerName="util" Dec 05 07:14:45 crc kubenswrapper[4997]: I1205 07:14:45.557317 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5" containerName="util" Dec 05 07:14:45 crc kubenswrapper[4997]: E1205 07:14:45.557351 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5" containerName="extract" Dec 05 07:14:45 crc kubenswrapper[4997]: I1205 07:14:45.557358 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5" containerName="extract" Dec 05 07:14:45 crc kubenswrapper[4997]: I1205 07:14:45.557477 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5" containerName="extract" Dec 05 07:14:45 crc kubenswrapper[4997]: I1205 07:14:45.558010 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-259kk" Dec 05 07:14:45 crc kubenswrapper[4997]: I1205 07:14:45.561919 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-ggt6b" Dec 05 07:14:45 crc kubenswrapper[4997]: I1205 07:14:45.588320 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-259kk"] Dec 05 07:14:45 crc kubenswrapper[4997]: I1205 07:14:45.673551 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjj7k\" (UniqueName: \"kubernetes.io/projected/90da0714-e130-42be-acbb-845291fb3d9a-kube-api-access-gjj7k\") pod \"openstack-operator-controller-operator-55b6fb9447-259kk\" (UID: \"90da0714-e130-42be-acbb-845291fb3d9a\") " pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-259kk" Dec 05 07:14:45 crc kubenswrapper[4997]: I1205 07:14:45.775303 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjj7k\" (UniqueName: \"kubernetes.io/projected/90da0714-e130-42be-acbb-845291fb3d9a-kube-api-access-gjj7k\") pod \"openstack-operator-controller-operator-55b6fb9447-259kk\" (UID: \"90da0714-e130-42be-acbb-845291fb3d9a\") " pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-259kk" Dec 05 07:14:45 crc kubenswrapper[4997]: I1205 07:14:45.803658 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjj7k\" (UniqueName: \"kubernetes.io/projected/90da0714-e130-42be-acbb-845291fb3d9a-kube-api-access-gjj7k\") pod \"openstack-operator-controller-operator-55b6fb9447-259kk\" (UID: \"90da0714-e130-42be-acbb-845291fb3d9a\") " pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-259kk" Dec 05 07:14:45 crc kubenswrapper[4997]: I1205 07:14:45.876331 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-259kk" Dec 05 07:14:46 crc kubenswrapper[4997]: I1205 07:14:46.134347 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-259kk"] Dec 05 07:14:46 crc kubenswrapper[4997]: I1205 07:14:46.364598 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-259kk" event={"ID":"90da0714-e130-42be-acbb-845291fb3d9a","Type":"ContainerStarted","Data":"0ba6a290a2ce3c7e8f5f7ec414e95fb9a668ce5a3118eaa3177a2b0ad6a2ce14"} Dec 05 07:14:49 crc kubenswrapper[4997]: I1205 07:14:49.770743 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:14:49 crc kubenswrapper[4997]: I1205 07:14:49.771134 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:14:55 crc kubenswrapper[4997]: I1205 07:14:55.427259 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-259kk" event={"ID":"90da0714-e130-42be-acbb-845291fb3d9a","Type":"ContainerStarted","Data":"89269d9bf7a03b8d03c928238d6730868aca9e2f928749ef5275e79bc1f36a22"} Dec 05 07:14:55 crc kubenswrapper[4997]: I1205 07:14:55.427914 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-259kk" Dec 05 07:14:55 crc kubenswrapper[4997]: I1205 07:14:55.457206 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-259kk" podStartSLOduration=2.14595583 podStartE2EDuration="10.457184794s" podCreationTimestamp="2025-12-05 07:14:45 +0000 UTC" firstStartedPulling="2025-12-05 07:14:46.143192865 +0000 UTC m=+1186.672100126" lastFinishedPulling="2025-12-05 07:14:54.454421809 +0000 UTC m=+1194.983329090" observedRunningTime="2025-12-05 07:14:55.452626491 +0000 UTC m=+1195.981533772" watchObservedRunningTime="2025-12-05 07:14:55.457184794 +0000 UTC m=+1195.986092055" Dec 05 07:15:00 crc kubenswrapper[4997]: I1205 07:15:00.138314 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm"] Dec 05 07:15:00 crc kubenswrapper[4997]: I1205 07:15:00.139563 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm" Dec 05 07:15:00 crc kubenswrapper[4997]: I1205 07:15:00.142288 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 07:15:00 crc kubenswrapper[4997]: I1205 07:15:00.143254 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 07:15:00 crc kubenswrapper[4997]: I1205 07:15:00.152022 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm"] Dec 05 07:15:00 crc kubenswrapper[4997]: I1205 07:15:00.267024 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jvqg\" (UniqueName: \"kubernetes.io/projected/cb045b24-fb27-4f7e-af8c-e1206fad74cd-kube-api-access-9jvqg\") pod \"collect-profiles-29415315-m2gdm\" (UID: \"cb045b24-fb27-4f7e-af8c-e1206fad74cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm" Dec 05 07:15:00 crc kubenswrapper[4997]: I1205 07:15:00.267111 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb045b24-fb27-4f7e-af8c-e1206fad74cd-config-volume\") pod \"collect-profiles-29415315-m2gdm\" (UID: \"cb045b24-fb27-4f7e-af8c-e1206fad74cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm" Dec 05 07:15:00 crc kubenswrapper[4997]: I1205 07:15:00.268080 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb045b24-fb27-4f7e-af8c-e1206fad74cd-secret-volume\") pod \"collect-profiles-29415315-m2gdm\" (UID: \"cb045b24-fb27-4f7e-af8c-e1206fad74cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm" Dec 05 07:15:00 crc kubenswrapper[4997]: I1205 07:15:00.369042 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jvqg\" (UniqueName: \"kubernetes.io/projected/cb045b24-fb27-4f7e-af8c-e1206fad74cd-kube-api-access-9jvqg\") pod \"collect-profiles-29415315-m2gdm\" (UID: \"cb045b24-fb27-4f7e-af8c-e1206fad74cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm" Dec 05 07:15:00 crc kubenswrapper[4997]: I1205 07:15:00.369111 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb045b24-fb27-4f7e-af8c-e1206fad74cd-config-volume\") pod \"collect-profiles-29415315-m2gdm\" (UID: \"cb045b24-fb27-4f7e-af8c-e1206fad74cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm" Dec 05 07:15:00 crc kubenswrapper[4997]: I1205 07:15:00.369185 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb045b24-fb27-4f7e-af8c-e1206fad74cd-secret-volume\") pod \"collect-profiles-29415315-m2gdm\" (UID: \"cb045b24-fb27-4f7e-af8c-e1206fad74cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm" Dec 05 07:15:00 crc kubenswrapper[4997]: I1205 07:15:00.371329 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb045b24-fb27-4f7e-af8c-e1206fad74cd-config-volume\") pod \"collect-profiles-29415315-m2gdm\" (UID: \"cb045b24-fb27-4f7e-af8c-e1206fad74cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm" Dec 05 07:15:00 crc kubenswrapper[4997]: I1205 07:15:00.381718 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb045b24-fb27-4f7e-af8c-e1206fad74cd-secret-volume\") pod \"collect-profiles-29415315-m2gdm\" (UID: \"cb045b24-fb27-4f7e-af8c-e1206fad74cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm" Dec 05 07:15:00 crc kubenswrapper[4997]: I1205 07:15:00.388623 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jvqg\" (UniqueName: \"kubernetes.io/projected/cb045b24-fb27-4f7e-af8c-e1206fad74cd-kube-api-access-9jvqg\") pod \"collect-profiles-29415315-m2gdm\" (UID: \"cb045b24-fb27-4f7e-af8c-e1206fad74cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm" Dec 05 07:15:00 crc kubenswrapper[4997]: I1205 07:15:00.455517 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm" Dec 05 07:15:00 crc kubenswrapper[4997]: I1205 07:15:00.750353 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm"] Dec 05 07:15:01 crc kubenswrapper[4997]: I1205 07:15:01.472736 4997 generic.go:334] "Generic (PLEG): container finished" podID="cb045b24-fb27-4f7e-af8c-e1206fad74cd" containerID="55c6a36d35ea7edcb8f82a46e34f0555480e13f08ce0ed00ebf4d91e920a6a4e" exitCode=0 Dec 05 07:15:01 crc kubenswrapper[4997]: I1205 07:15:01.472836 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm" event={"ID":"cb045b24-fb27-4f7e-af8c-e1206fad74cd","Type":"ContainerDied","Data":"55c6a36d35ea7edcb8f82a46e34f0555480e13f08ce0ed00ebf4d91e920a6a4e"} Dec 05 07:15:01 crc kubenswrapper[4997]: I1205 07:15:01.473041 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm" event={"ID":"cb045b24-fb27-4f7e-af8c-e1206fad74cd","Type":"ContainerStarted","Data":"e59acea3e5d5cdd33089eafa5ca3f2461eacf650d44ead69e965f8a7526e8cfd"} Dec 05 07:15:02 crc kubenswrapper[4997]: I1205 07:15:02.868374 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm" Dec 05 07:15:03 crc kubenswrapper[4997]: I1205 07:15:03.004092 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb045b24-fb27-4f7e-af8c-e1206fad74cd-secret-volume\") pod \"cb045b24-fb27-4f7e-af8c-e1206fad74cd\" (UID: \"cb045b24-fb27-4f7e-af8c-e1206fad74cd\") " Dec 05 07:15:03 crc kubenswrapper[4997]: I1205 07:15:03.004173 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb045b24-fb27-4f7e-af8c-e1206fad74cd-config-volume\") pod \"cb045b24-fb27-4f7e-af8c-e1206fad74cd\" (UID: \"cb045b24-fb27-4f7e-af8c-e1206fad74cd\") " Dec 05 07:15:03 crc kubenswrapper[4997]: I1205 07:15:03.004222 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jvqg\" (UniqueName: \"kubernetes.io/projected/cb045b24-fb27-4f7e-af8c-e1206fad74cd-kube-api-access-9jvqg\") pod \"cb045b24-fb27-4f7e-af8c-e1206fad74cd\" (UID: \"cb045b24-fb27-4f7e-af8c-e1206fad74cd\") " Dec 05 07:15:03 crc kubenswrapper[4997]: I1205 07:15:03.004912 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb045b24-fb27-4f7e-af8c-e1206fad74cd-config-volume" (OuterVolumeSpecName: "config-volume") pod "cb045b24-fb27-4f7e-af8c-e1206fad74cd" (UID: "cb045b24-fb27-4f7e-af8c-e1206fad74cd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:15:03 crc kubenswrapper[4997]: I1205 07:15:03.009981 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb045b24-fb27-4f7e-af8c-e1206fad74cd-kube-api-access-9jvqg" (OuterVolumeSpecName: "kube-api-access-9jvqg") pod "cb045b24-fb27-4f7e-af8c-e1206fad74cd" (UID: "cb045b24-fb27-4f7e-af8c-e1206fad74cd"). InnerVolumeSpecName "kube-api-access-9jvqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:15:03 crc kubenswrapper[4997]: I1205 07:15:03.010228 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb045b24-fb27-4f7e-af8c-e1206fad74cd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cb045b24-fb27-4f7e-af8c-e1206fad74cd" (UID: "cb045b24-fb27-4f7e-af8c-e1206fad74cd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:15:03 crc kubenswrapper[4997]: I1205 07:15:03.119994 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jvqg\" (UniqueName: \"kubernetes.io/projected/cb045b24-fb27-4f7e-af8c-e1206fad74cd-kube-api-access-9jvqg\") on node \"crc\" DevicePath \"\"" Dec 05 07:15:03 crc kubenswrapper[4997]: I1205 07:15:03.120050 4997 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb045b24-fb27-4f7e-af8c-e1206fad74cd-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 07:15:03 crc kubenswrapper[4997]: I1205 07:15:03.120065 4997 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb045b24-fb27-4f7e-af8c-e1206fad74cd-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 07:15:03 crc kubenswrapper[4997]: I1205 07:15:03.487968 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm" event={"ID":"cb045b24-fb27-4f7e-af8c-e1206fad74cd","Type":"ContainerDied","Data":"e59acea3e5d5cdd33089eafa5ca3f2461eacf650d44ead69e965f8a7526e8cfd"} Dec 05 07:15:03 crc kubenswrapper[4997]: I1205 07:15:03.488471 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e59acea3e5d5cdd33089eafa5ca3f2461eacf650d44ead69e965f8a7526e8cfd" Dec 05 07:15:03 crc kubenswrapper[4997]: I1205 07:15:03.488075 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm" Dec 05 07:15:05 crc kubenswrapper[4997]: I1205 07:15:05.882437 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-259kk" Dec 05 07:15:19 crc kubenswrapper[4997]: I1205 07:15:19.770535 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:15:19 crc kubenswrapper[4997]: I1205 07:15:19.771120 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.630677 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-r4m8t"] Dec 05 07:15:24 crc kubenswrapper[4997]: E1205 07:15:24.631395 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb045b24-fb27-4f7e-af8c-e1206fad74cd" containerName="collect-profiles" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.631413 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb045b24-fb27-4f7e-af8c-e1206fad74cd" containerName="collect-profiles" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.631592 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb045b24-fb27-4f7e-af8c-e1206fad74cd" containerName="collect-profiles" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.632546 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r4m8t" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.635658 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-k82zl" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.644923 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-rnq5d"] Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.646112 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-rnq5d" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.649683 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-tbgsq" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.649713 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-r4m8t"] Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.668990 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-rnq5d"] Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.671543 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb6db\" (UniqueName: \"kubernetes.io/projected/fe84dde4-76f7-4df6-a6fd-d1b45d097e9b-kube-api-access-wb6db\") pod \"barbican-operator-controller-manager-7d9dfd778-r4m8t\" (UID: \"fe84dde4-76f7-4df6-a6fd-d1b45d097e9b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r4m8t" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.671671 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c2hp\" (UniqueName: \"kubernetes.io/projected/27e5617b-9979-45b2-a526-bb671edf63c2-kube-api-access-5c2hp\") pod \"cinder-operator-controller-manager-859b6ccc6-rnq5d\" (UID: \"27e5617b-9979-45b2-a526-bb671edf63c2\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-rnq5d" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.678599 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-r9mh5"] Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.679870 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-r9mh5" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.683856 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-cpf6q" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.705723 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-5zk9v"] Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.707283 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-5zk9v" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.710701 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-ksvgd" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.721552 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-r9mh5"] Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.749694 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-lpx4g"] Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.752454 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-lpx4g" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.758875 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-jmsf6" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.772586 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6k77\" (UniqueName: \"kubernetes.io/projected/351c9912-1ac6-4b8b-b009-d073b1bd28a5-kube-api-access-j6k77\") pod \"designate-operator-controller-manager-78b4bc895b-r9mh5\" (UID: \"351c9912-1ac6-4b8b-b009-d073b1bd28a5\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-r9mh5" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.772708 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb6db\" (UniqueName: \"kubernetes.io/projected/fe84dde4-76f7-4df6-a6fd-d1b45d097e9b-kube-api-access-wb6db\") pod \"barbican-operator-controller-manager-7d9dfd778-r4m8t\" (UID: \"fe84dde4-76f7-4df6-a6fd-d1b45d097e9b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r4m8t" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.772748 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c2hp\" (UniqueName: \"kubernetes.io/projected/27e5617b-9979-45b2-a526-bb671edf63c2-kube-api-access-5c2hp\") pod \"cinder-operator-controller-manager-859b6ccc6-rnq5d\" (UID: \"27e5617b-9979-45b2-a526-bb671edf63c2\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-rnq5d" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.772776 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btr69\" (UniqueName: \"kubernetes.io/projected/c488dae2-6bbc-434a-a1c8-d3bd6ba5a549-kube-api-access-btr69\") pod \"heat-operator-controller-manager-5f64f6f8bb-5zk9v\" (UID: \"c488dae2-6bbc-434a-a1c8-d3bd6ba5a549\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-5zk9v" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.772834 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gwcg\" (UniqueName: \"kubernetes.io/projected/d5054dc6-123a-4d45-8340-5315c0c57e32-kube-api-access-5gwcg\") pod \"glance-operator-controller-manager-77987cd8cd-lpx4g\" (UID: \"d5054dc6-123a-4d45-8340-5315c0c57e32\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-lpx4g" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.782597 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-jk2cb"] Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.783780 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-jk2cb" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.789924 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-j9t8q" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.803668 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-5zk9v"] Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.812698 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c2hp\" (UniqueName: \"kubernetes.io/projected/27e5617b-9979-45b2-a526-bb671edf63c2-kube-api-access-5c2hp\") pod \"cinder-operator-controller-manager-859b6ccc6-rnq5d\" (UID: \"27e5617b-9979-45b2-a526-bb671edf63c2\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-rnq5d" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.822679 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-pppgg"] Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.824207 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.835722 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb6db\" (UniqueName: \"kubernetes.io/projected/fe84dde4-76f7-4df6-a6fd-d1b45d097e9b-kube-api-access-wb6db\") pod \"barbican-operator-controller-manager-7d9dfd778-r4m8t\" (UID: \"fe84dde4-76f7-4df6-a6fd-d1b45d097e9b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r4m8t" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.836386 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-8knpj" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.836641 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.856986 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-jk2cb"] Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.875390 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert\") pod \"infra-operator-controller-manager-57548d458d-pppgg\" (UID: \"6aac09d2-991d-45b9-bba9-4cf2d70278ab\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.875458 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6k77\" (UniqueName: \"kubernetes.io/projected/351c9912-1ac6-4b8b-b009-d073b1bd28a5-kube-api-access-j6k77\") pod \"designate-operator-controller-manager-78b4bc895b-r9mh5\" (UID: \"351c9912-1ac6-4b8b-b009-d073b1bd28a5\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-r9mh5" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.875488 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfw8t\" (UniqueName: \"kubernetes.io/projected/6aac09d2-991d-45b9-bba9-4cf2d70278ab-kube-api-access-mfw8t\") pod \"infra-operator-controller-manager-57548d458d-pppgg\" (UID: \"6aac09d2-991d-45b9-bba9-4cf2d70278ab\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.875507 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjtmj\" (UniqueName: \"kubernetes.io/projected/7db21a3e-655e-4874-8366-dfdb5d2521f5-kube-api-access-wjtmj\") pod \"horizon-operator-controller-manager-68c6d99b8f-jk2cb\" (UID: \"7db21a3e-655e-4874-8366-dfdb5d2521f5\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-jk2cb" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.875546 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btr69\" (UniqueName: \"kubernetes.io/projected/c488dae2-6bbc-434a-a1c8-d3bd6ba5a549-kube-api-access-btr69\") pod \"heat-operator-controller-manager-5f64f6f8bb-5zk9v\" (UID: \"c488dae2-6bbc-434a-a1c8-d3bd6ba5a549\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-5zk9v" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.875589 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gwcg\" (UniqueName: \"kubernetes.io/projected/d5054dc6-123a-4d45-8340-5315c0c57e32-kube-api-access-5gwcg\") pod \"glance-operator-controller-manager-77987cd8cd-lpx4g\" (UID: \"d5054dc6-123a-4d45-8340-5315c0c57e32\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-lpx4g" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.907051 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-lpx4g"] Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.910821 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btr69\" (UniqueName: \"kubernetes.io/projected/c488dae2-6bbc-434a-a1c8-d3bd6ba5a549-kube-api-access-btr69\") pod \"heat-operator-controller-manager-5f64f6f8bb-5zk9v\" (UID: \"c488dae2-6bbc-434a-a1c8-d3bd6ba5a549\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-5zk9v" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.914204 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gwcg\" (UniqueName: \"kubernetes.io/projected/d5054dc6-123a-4d45-8340-5315c0c57e32-kube-api-access-5gwcg\") pod \"glance-operator-controller-manager-77987cd8cd-lpx4g\" (UID: \"d5054dc6-123a-4d45-8340-5315c0c57e32\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-lpx4g" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.916606 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6k77\" (UniqueName: \"kubernetes.io/projected/351c9912-1ac6-4b8b-b009-d073b1bd28a5-kube-api-access-j6k77\") pod \"designate-operator-controller-manager-78b4bc895b-r9mh5\" (UID: \"351c9912-1ac6-4b8b-b009-d073b1bd28a5\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-r9mh5" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.925479 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-pppgg"] Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.939216 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-q62zq"] Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.940785 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q62zq" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.942891 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-zccqz" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.957407 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r4m8t" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.968166 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-rnq5d" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.969389 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-88hm7"] Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.970596 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88hm7" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.975018 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-rttcf" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.976489 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfw8t\" (UniqueName: \"kubernetes.io/projected/6aac09d2-991d-45b9-bba9-4cf2d70278ab-kube-api-access-mfw8t\") pod \"infra-operator-controller-manager-57548d458d-pppgg\" (UID: \"6aac09d2-991d-45b9-bba9-4cf2d70278ab\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.976543 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjtmj\" (UniqueName: \"kubernetes.io/projected/7db21a3e-655e-4874-8366-dfdb5d2521f5-kube-api-access-wjtmj\") pod \"horizon-operator-controller-manager-68c6d99b8f-jk2cb\" (UID: \"7db21a3e-655e-4874-8366-dfdb5d2521f5\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-jk2cb" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.976682 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert\") pod \"infra-operator-controller-manager-57548d458d-pppgg\" (UID: \"6aac09d2-991d-45b9-bba9-4cf2d70278ab\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" Dec 05 07:15:24 crc kubenswrapper[4997]: I1205 07:15:24.976718 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2l9p\" (UniqueName: \"kubernetes.io/projected/c9d46c14-6c5e-42d4-81f3-43e7169d23a0-kube-api-access-k2l9p\") pod \"ironic-operator-controller-manager-6c548fd776-q62zq\" (UID: \"c9d46c14-6c5e-42d4-81f3-43e7169d23a0\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q62zq" Dec 05 07:15:24 crc kubenswrapper[4997]: E1205 07:15:24.977150 4997 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 07:15:24 crc kubenswrapper[4997]: E1205 07:15:24.977206 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert podName:6aac09d2-991d-45b9-bba9-4cf2d70278ab nodeName:}" failed. No retries permitted until 2025-12-05 07:15:25.477187392 +0000 UTC m=+1226.006094653 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert") pod "infra-operator-controller-manager-57548d458d-pppgg" (UID: "6aac09d2-991d-45b9-bba9-4cf2d70278ab") : secret "infra-operator-webhook-server-cert" not found Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.009920 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-r9mh5" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.030995 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-5zk9v" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.038062 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjtmj\" (UniqueName: \"kubernetes.io/projected/7db21a3e-655e-4874-8366-dfdb5d2521f5-kube-api-access-wjtmj\") pod \"horizon-operator-controller-manager-68c6d99b8f-jk2cb\" (UID: \"7db21a3e-655e-4874-8366-dfdb5d2521f5\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-jk2cb" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.057222 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-q62zq"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.079687 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfw8t\" (UniqueName: \"kubernetes.io/projected/6aac09d2-991d-45b9-bba9-4cf2d70278ab-kube-api-access-mfw8t\") pod \"infra-operator-controller-manager-57548d458d-pppgg\" (UID: \"6aac09d2-991d-45b9-bba9-4cf2d70278ab\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.090249 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-lpx4g" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.093175 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2l9p\" (UniqueName: \"kubernetes.io/projected/c9d46c14-6c5e-42d4-81f3-43e7169d23a0-kube-api-access-k2l9p\") pod \"ironic-operator-controller-manager-6c548fd776-q62zq\" (UID: \"c9d46c14-6c5e-42d4-81f3-43e7169d23a0\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q62zq" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.097906 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-sbfxr"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.101213 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-sbfxr" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.143323 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-jk2cb" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.150960 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-h4rfj" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.172278 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-88hm7"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.183356 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-sbfxr"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.195017 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fbtt\" (UniqueName: \"kubernetes.io/projected/3580ba04-55a0-4e33-aa08-932a2648bf2d-kube-api-access-5fbtt\") pod \"manila-operator-controller-manager-7c79b5df47-sbfxr\" (UID: \"3580ba04-55a0-4e33-aa08-932a2648bf2d\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-sbfxr" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.195065 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc88k\" (UniqueName: \"kubernetes.io/projected/40df044a-6922-43cf-ae37-ebc8dd637119-kube-api-access-hc88k\") pod \"keystone-operator-controller-manager-7765d96ddf-88hm7\" (UID: \"40df044a-6922-43cf-ae37-ebc8dd637119\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88hm7" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.198224 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2l9p\" (UniqueName: \"kubernetes.io/projected/c9d46c14-6c5e-42d4-81f3-43e7169d23a0-kube-api-access-k2l9p\") pod \"ironic-operator-controller-manager-6c548fd776-q62zq\" (UID: \"c9d46c14-6c5e-42d4-81f3-43e7169d23a0\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q62zq" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.202868 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tpg29"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.204281 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tpg29" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.210584 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rq99q"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.212127 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rq99q" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.219801 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rq99q"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.220194 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-5b5dk" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.232667 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tpg29"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.232741 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-qppfb"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.234086 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-qppfb" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.246345 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-sst5m" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.246773 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-cng85" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.246766 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-qppfb"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.247095 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-r9knp"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.248564 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-r9knp" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.253906 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-gk8w7" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.266595 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-r9knp"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.296481 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxf5p\" (UniqueName: \"kubernetes.io/projected/e5f7e30e-e2a8-4131-9c10-1dddfe528684-kube-api-access-gxf5p\") pod \"mariadb-operator-controller-manager-56bbcc9d85-tpg29\" (UID: \"e5f7e30e-e2a8-4131-9c10-1dddfe528684\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tpg29" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.296546 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgmf7\" (UniqueName: \"kubernetes.io/projected/7afb5248-e41e-46ec-9e46-ed9fcc7cf8d3-kube-api-access-xgmf7\") pod \"nova-operator-controller-manager-697bc559fc-qppfb\" (UID: \"7afb5248-e41e-46ec-9e46-ed9fcc7cf8d3\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-qppfb" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.296575 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qkm7\" (UniqueName: \"kubernetes.io/projected/9cc09cce-bafa-4f17-bd45-b767fa576e54-kube-api-access-9qkm7\") pod \"octavia-operator-controller-manager-998648c74-r9knp\" (UID: \"9cc09cce-bafa-4f17-bd45-b767fa576e54\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-r9knp" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.296695 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fbtt\" (UniqueName: \"kubernetes.io/projected/3580ba04-55a0-4e33-aa08-932a2648bf2d-kube-api-access-5fbtt\") pod \"manila-operator-controller-manager-7c79b5df47-sbfxr\" (UID: \"3580ba04-55a0-4e33-aa08-932a2648bf2d\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-sbfxr" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.296731 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc88k\" (UniqueName: \"kubernetes.io/projected/40df044a-6922-43cf-ae37-ebc8dd637119-kube-api-access-hc88k\") pod \"keystone-operator-controller-manager-7765d96ddf-88hm7\" (UID: \"40df044a-6922-43cf-ae37-ebc8dd637119\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88hm7" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.296760 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbgxs\" (UniqueName: \"kubernetes.io/projected/7a8b9284-3f13-4396-8f85-36f0c59e1ba3-kube-api-access-wbgxs\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-rq99q\" (UID: \"7a8b9284-3f13-4396-8f85-36f0c59e1ba3\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rq99q" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.315064 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q62zq" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.338302 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.340065 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.345832 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-lwrd4" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.346074 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.393039 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-hrs7p"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.399635 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fbtt\" (UniqueName: \"kubernetes.io/projected/3580ba04-55a0-4e33-aa08-932a2648bf2d-kube-api-access-5fbtt\") pod \"manila-operator-controller-manager-7c79b5df47-sbfxr\" (UID: \"3580ba04-55a0-4e33-aa08-932a2648bf2d\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-sbfxr" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.413065 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbgxs\" (UniqueName: \"kubernetes.io/projected/7a8b9284-3f13-4396-8f85-36f0c59e1ba3-kube-api-access-wbgxs\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-rq99q\" (UID: \"7a8b9284-3f13-4396-8f85-36f0c59e1ba3\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rq99q" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.413140 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxf5p\" (UniqueName: \"kubernetes.io/projected/e5f7e30e-e2a8-4131-9c10-1dddfe528684-kube-api-access-gxf5p\") pod \"mariadb-operator-controller-manager-56bbcc9d85-tpg29\" (UID: \"e5f7e30e-e2a8-4131-9c10-1dddfe528684\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tpg29" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.413158 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgmf7\" (UniqueName: \"kubernetes.io/projected/7afb5248-e41e-46ec-9e46-ed9fcc7cf8d3-kube-api-access-xgmf7\") pod \"nova-operator-controller-manager-697bc559fc-qppfb\" (UID: \"7afb5248-e41e-46ec-9e46-ed9fcc7cf8d3\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-qppfb" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.413177 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qkm7\" (UniqueName: \"kubernetes.io/projected/9cc09cce-bafa-4f17-bd45-b767fa576e54-kube-api-access-9qkm7\") pod \"octavia-operator-controller-manager-998648c74-r9knp\" (UID: \"9cc09cce-bafa-4f17-bd45-b767fa576e54\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-r9knp" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.423784 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hrs7p" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.432596 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-ztwsn"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.434099 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-ztwsn" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.442527 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc88k\" (UniqueName: \"kubernetes.io/projected/40df044a-6922-43cf-ae37-ebc8dd637119-kube-api-access-hc88k\") pod \"keystone-operator-controller-manager-7765d96ddf-88hm7\" (UID: \"40df044a-6922-43cf-ae37-ebc8dd637119\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88hm7" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.453997 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-xffsp" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.454370 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-9nms7" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.463732 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-hrs7p"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.494348 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxf5p\" (UniqueName: \"kubernetes.io/projected/e5f7e30e-e2a8-4131-9c10-1dddfe528684-kube-api-access-gxf5p\") pod \"mariadb-operator-controller-manager-56bbcc9d85-tpg29\" (UID: \"e5f7e30e-e2a8-4131-9c10-1dddfe528684\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tpg29" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.495045 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgmf7\" (UniqueName: \"kubernetes.io/projected/7afb5248-e41e-46ec-9e46-ed9fcc7cf8d3-kube-api-access-xgmf7\") pod \"nova-operator-controller-manager-697bc559fc-qppfb\" (UID: \"7afb5248-e41e-46ec-9e46-ed9fcc7cf8d3\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-qppfb" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.505388 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbgxs\" (UniqueName: \"kubernetes.io/projected/7a8b9284-3f13-4396-8f85-36f0c59e1ba3-kube-api-access-wbgxs\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-rq99q\" (UID: \"7a8b9284-3f13-4396-8f85-36f0c59e1ba3\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rq99q" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.507346 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qkm7\" (UniqueName: \"kubernetes.io/projected/9cc09cce-bafa-4f17-bd45-b767fa576e54-kube-api-access-9qkm7\") pod \"octavia-operator-controller-manager-998648c74-r9knp\" (UID: \"9cc09cce-bafa-4f17-bd45-b767fa576e54\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-r9knp" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.512960 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-sbfxr" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.514247 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert\") pod \"infra-operator-controller-manager-57548d458d-pppgg\" (UID: \"6aac09d2-991d-45b9-bba9-4cf2d70278ab\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.514336 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ffpx\" (UniqueName: \"kubernetes.io/projected/64d88cd2-49b5-439a-9344-bb9dc4ec5730-kube-api-access-5ffpx\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56jjnr\" (UID: \"64d88cd2-49b5-439a-9344-bb9dc4ec5730\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.514373 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56jjnr\" (UID: \"64d88cd2-49b5-439a-9344-bb9dc4ec5730\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" Dec 05 07:15:25 crc kubenswrapper[4997]: E1205 07:15:25.514969 4997 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 07:15:25 crc kubenswrapper[4997]: E1205 07:15:25.515206 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert podName:6aac09d2-991d-45b9-bba9-4cf2d70278ab nodeName:}" failed. No retries permitted until 2025-12-05 07:15:26.515189463 +0000 UTC m=+1227.044096724 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert") pod "infra-operator-controller-manager-57548d458d-pppgg" (UID: "6aac09d2-991d-45b9-bba9-4cf2d70278ab") : secret "infra-operator-webhook-server-cert" not found Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.532684 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-ztwsn"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.553684 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.581940 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-jtjgd"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.591295 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-jtjgd" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.594037 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tpg29" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.598970 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-rvh4x" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.617784 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56jjnr\" (UID: \"64d88cd2-49b5-439a-9344-bb9dc4ec5730\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" Dec 05 07:15:25 crc kubenswrapper[4997]: E1205 07:15:25.618093 4997 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 07:15:25 crc kubenswrapper[4997]: E1205 07:15:25.618191 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert podName:64d88cd2-49b5-439a-9344-bb9dc4ec5730 nodeName:}" failed. No retries permitted until 2025-12-05 07:15:26.118164381 +0000 UTC m=+1226.647071642 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f56jjnr" (UID: "64d88cd2-49b5-439a-9344-bb9dc4ec5730") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.618305 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whvgl\" (UniqueName: \"kubernetes.io/projected/3f6028fa-7b09-49ec-a8ed-4f4b25419e19-kube-api-access-whvgl\") pod \"ovn-operator-controller-manager-b6456fdb6-hrs7p\" (UID: \"3f6028fa-7b09-49ec-a8ed-4f4b25419e19\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hrs7p" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.618445 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f64br\" (UniqueName: \"kubernetes.io/projected/905ae077-9cc3-41a9-b03f-77b3eeb052b3-kube-api-access-f64br\") pod \"placement-operator-controller-manager-78f8948974-ztwsn\" (UID: \"905ae077-9cc3-41a9-b03f-77b3eeb052b3\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-ztwsn" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.618569 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k7j6\" (UniqueName: \"kubernetes.io/projected/9597d662-50ea-4066-8bbc-232cc1d90dbd-kube-api-access-9k7j6\") pod \"swift-operator-controller-manager-5f8c65bbfc-jtjgd\" (UID: \"9597d662-50ea-4066-8bbc-232cc1d90dbd\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-jtjgd" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.618760 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ffpx\" (UniqueName: \"kubernetes.io/projected/64d88cd2-49b5-439a-9344-bb9dc4ec5730-kube-api-access-5ffpx\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56jjnr\" (UID: \"64d88cd2-49b5-439a-9344-bb9dc4ec5730\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.625738 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rq99q" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.641478 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ffpx\" (UniqueName: \"kubernetes.io/projected/64d88cd2-49b5-439a-9344-bb9dc4ec5730-kube-api-access-5ffpx\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56jjnr\" (UID: \"64d88cd2-49b5-439a-9344-bb9dc4ec5730\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.678169 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-qppfb" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.688020 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88hm7" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.707335 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-r9knp" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.718259 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-cjggt"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.720244 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-cjggt" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.720785 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whvgl\" (UniqueName: \"kubernetes.io/projected/3f6028fa-7b09-49ec-a8ed-4f4b25419e19-kube-api-access-whvgl\") pod \"ovn-operator-controller-manager-b6456fdb6-hrs7p\" (UID: \"3f6028fa-7b09-49ec-a8ed-4f4b25419e19\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hrs7p" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.720847 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f64br\" (UniqueName: \"kubernetes.io/projected/905ae077-9cc3-41a9-b03f-77b3eeb052b3-kube-api-access-f64br\") pod \"placement-operator-controller-manager-78f8948974-ztwsn\" (UID: \"905ae077-9cc3-41a9-b03f-77b3eeb052b3\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-ztwsn" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.720871 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k7j6\" (UniqueName: \"kubernetes.io/projected/9597d662-50ea-4066-8bbc-232cc1d90dbd-kube-api-access-9k7j6\") pod \"swift-operator-controller-manager-5f8c65bbfc-jtjgd\" (UID: \"9597d662-50ea-4066-8bbc-232cc1d90dbd\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-jtjgd" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.727184 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-n4k8r" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.738682 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-jtjgd"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.759405 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k7j6\" (UniqueName: \"kubernetes.io/projected/9597d662-50ea-4066-8bbc-232cc1d90dbd-kube-api-access-9k7j6\") pod \"swift-operator-controller-manager-5f8c65bbfc-jtjgd\" (UID: \"9597d662-50ea-4066-8bbc-232cc1d90dbd\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-jtjgd" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.767532 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whvgl\" (UniqueName: \"kubernetes.io/projected/3f6028fa-7b09-49ec-a8ed-4f4b25419e19-kube-api-access-whvgl\") pod \"ovn-operator-controller-manager-b6456fdb6-hrs7p\" (UID: \"3f6028fa-7b09-49ec-a8ed-4f4b25419e19\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hrs7p" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.781413 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f64br\" (UniqueName: \"kubernetes.io/projected/905ae077-9cc3-41a9-b03f-77b3eeb052b3-kube-api-access-f64br\") pod \"placement-operator-controller-manager-78f8948974-ztwsn\" (UID: \"905ae077-9cc3-41a9-b03f-77b3eeb052b3\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-ztwsn" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.804043 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-cjggt"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.823892 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rm5x\" (UniqueName: \"kubernetes.io/projected/e1874b23-363f-4303-bfeb-86282b20963d-kube-api-access-5rm5x\") pod \"telemetry-operator-controller-manager-76cc84c6bb-cjggt\" (UID: \"e1874b23-363f-4303-bfeb-86282b20963d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-cjggt" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.864891 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hrs7p" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.918644 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-kbfbg"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.941558 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-kbfbg" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.944437 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rm5x\" (UniqueName: \"kubernetes.io/projected/e1874b23-363f-4303-bfeb-86282b20963d-kube-api-access-5rm5x\") pod \"telemetry-operator-controller-manager-76cc84c6bb-cjggt\" (UID: \"e1874b23-363f-4303-bfeb-86282b20963d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-cjggt" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.976773 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-kbfbg"] Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.978191 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-nd92h" Dec 05 07:15:25 crc kubenswrapper[4997]: I1205 07:15:25.999552 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-ztwsn" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.014829 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-skcxd"] Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.016899 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skcxd" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.021415 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-9twj6" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.027113 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-jtjgd" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.042843 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-skcxd"] Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.046666 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbng9\" (UniqueName: \"kubernetes.io/projected/f309b30b-9712-4175-b237-7c35ac2e5dee-kube-api-access-nbng9\") pod \"watcher-operator-controller-manager-769dc69bc-skcxd\" (UID: \"f309b30b-9712-4175-b237-7c35ac2e5dee\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skcxd" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.046749 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6bhp\" (UniqueName: \"kubernetes.io/projected/25ca1d60-a629-447e-afa7-7a9f0e6f0b7e-kube-api-access-q6bhp\") pod \"test-operator-controller-manager-5854674fcc-kbfbg\" (UID: \"25ca1d60-a629-447e-afa7-7a9f0e6f0b7e\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-kbfbg" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.048712 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rm5x\" (UniqueName: \"kubernetes.io/projected/e1874b23-363f-4303-bfeb-86282b20963d-kube-api-access-5rm5x\") pod \"telemetry-operator-controller-manager-76cc84c6bb-cjggt\" (UID: \"e1874b23-363f-4303-bfeb-86282b20963d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-cjggt" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.072228 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-cjggt" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.092129 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk"] Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.093529 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.118176 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk"] Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.122752 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-zfnt5" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.123132 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.123330 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.152604 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qsqst"] Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.153709 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qsqst" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.160288 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbng9\" (UniqueName: \"kubernetes.io/projected/f309b30b-9712-4175-b237-7c35ac2e5dee-kube-api-access-nbng9\") pod \"watcher-operator-controller-manager-769dc69bc-skcxd\" (UID: \"f309b30b-9712-4175-b237-7c35ac2e5dee\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skcxd" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.160475 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.160569 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6bhp\" (UniqueName: \"kubernetes.io/projected/25ca1d60-a629-447e-afa7-7a9f0e6f0b7e-kube-api-access-q6bhp\") pod \"test-operator-controller-manager-5854674fcc-kbfbg\" (UID: \"25ca1d60-a629-447e-afa7-7a9f0e6f0b7e\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-kbfbg" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.160656 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.160717 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgtjl\" (UniqueName: \"kubernetes.io/projected/6c477704-084c-4d28-8ab3-cd0f29ca488f-kube-api-access-dgtjl\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.160771 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56jjnr\" (UID: \"64d88cd2-49b5-439a-9344-bb9dc4ec5730\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" Dec 05 07:15:26 crc kubenswrapper[4997]: E1205 07:15:26.161058 4997 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 07:15:26 crc kubenswrapper[4997]: E1205 07:15:26.161131 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert podName:64d88cd2-49b5-439a-9344-bb9dc4ec5730 nodeName:}" failed. No retries permitted until 2025-12-05 07:15:27.161107616 +0000 UTC m=+1227.690014877 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f56jjnr" (UID: "64d88cd2-49b5-439a-9344-bb9dc4ec5730") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.170082 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-v8mnz" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.185285 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qsqst"] Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.217307 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6bhp\" (UniqueName: \"kubernetes.io/projected/25ca1d60-a629-447e-afa7-7a9f0e6f0b7e-kube-api-access-q6bhp\") pod \"test-operator-controller-manager-5854674fcc-kbfbg\" (UID: \"25ca1d60-a629-447e-afa7-7a9f0e6f0b7e\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-kbfbg" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.218528 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-kbfbg" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.228930 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbng9\" (UniqueName: \"kubernetes.io/projected/f309b30b-9712-4175-b237-7c35ac2e5dee-kube-api-access-nbng9\") pod \"watcher-operator-controller-manager-769dc69bc-skcxd\" (UID: \"f309b30b-9712-4175-b237-7c35ac2e5dee\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skcxd" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.236949 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skcxd" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.266176 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.266237 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgtjl\" (UniqueName: \"kubernetes.io/projected/6c477704-084c-4d28-8ab3-cd0f29ca488f-kube-api-access-dgtjl\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.266362 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6wtb\" (UniqueName: \"kubernetes.io/projected/8d31795c-bf5a-48fc-85cc-0c1385c2d7b0-kube-api-access-l6wtb\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qsqst\" (UID: \"8d31795c-bf5a-48fc-85cc-0c1385c2d7b0\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qsqst" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.266404 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:26 crc kubenswrapper[4997]: E1205 07:15:26.266986 4997 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 07:15:26 crc kubenswrapper[4997]: E1205 07:15:26.267098 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs podName:6c477704-084c-4d28-8ab3-cd0f29ca488f nodeName:}" failed. No retries permitted until 2025-12-05 07:15:26.767077874 +0000 UTC m=+1227.295985135 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-7wbfk" (UID: "6c477704-084c-4d28-8ab3-cd0f29ca488f") : secret "webhook-server-cert" not found Dec 05 07:15:26 crc kubenswrapper[4997]: E1205 07:15:26.269467 4997 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 07:15:26 crc kubenswrapper[4997]: E1205 07:15:26.269534 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs podName:6c477704-084c-4d28-8ab3-cd0f29ca488f nodeName:}" failed. No retries permitted until 2025-12-05 07:15:26.769514539 +0000 UTC m=+1227.298421810 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-7wbfk" (UID: "6c477704-084c-4d28-8ab3-cd0f29ca488f") : secret "metrics-server-cert" not found Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.316658 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-r9mh5"] Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.323171 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgtjl\" (UniqueName: \"kubernetes.io/projected/6c477704-084c-4d28-8ab3-cd0f29ca488f-kube-api-access-dgtjl\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.367786 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6wtb\" (UniqueName: \"kubernetes.io/projected/8d31795c-bf5a-48fc-85cc-0c1385c2d7b0-kube-api-access-l6wtb\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qsqst\" (UID: \"8d31795c-bf5a-48fc-85cc-0c1385c2d7b0\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qsqst" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.409043 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6wtb\" (UniqueName: \"kubernetes.io/projected/8d31795c-bf5a-48fc-85cc-0c1385c2d7b0-kube-api-access-l6wtb\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qsqst\" (UID: \"8d31795c-bf5a-48fc-85cc-0c1385c2d7b0\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qsqst" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.412097 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.570952 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert\") pod \"infra-operator-controller-manager-57548d458d-pppgg\" (UID: \"6aac09d2-991d-45b9-bba9-4cf2d70278ab\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" Dec 05 07:15:26 crc kubenswrapper[4997]: E1205 07:15:26.571187 4997 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 07:15:26 crc kubenswrapper[4997]: E1205 07:15:26.571699 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert podName:6aac09d2-991d-45b9-bba9-4cf2d70278ab nodeName:}" failed. No retries permitted until 2025-12-05 07:15:28.571642041 +0000 UTC m=+1229.100549302 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert") pod "infra-operator-controller-manager-57548d458d-pppgg" (UID: "6aac09d2-991d-45b9-bba9-4cf2d70278ab") : secret "infra-operator-webhook-server-cert" not found Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.899316 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.899398 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:26 crc kubenswrapper[4997]: E1205 07:15:26.899581 4997 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 07:15:26 crc kubenswrapper[4997]: E1205 07:15:26.899649 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs podName:6c477704-084c-4d28-8ab3-cd0f29ca488f nodeName:}" failed. No retries permitted until 2025-12-05 07:15:27.899633907 +0000 UTC m=+1228.428541168 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-7wbfk" (UID: "6c477704-084c-4d28-8ab3-cd0f29ca488f") : secret "webhook-server-cert" not found Dec 05 07:15:26 crc kubenswrapper[4997]: E1205 07:15:26.899701 4997 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 07:15:26 crc kubenswrapper[4997]: E1205 07:15:26.899723 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs podName:6c477704-084c-4d28-8ab3-cd0f29ca488f nodeName:}" failed. No retries permitted until 2025-12-05 07:15:27.899716809 +0000 UTC m=+1228.428624070 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-7wbfk" (UID: "6c477704-084c-4d28-8ab3-cd0f29ca488f") : secret "metrics-server-cert" not found Dec 05 07:15:26 crc kubenswrapper[4997]: I1205 07:15:26.900156 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qsqst" Dec 05 07:15:27 crc kubenswrapper[4997]: I1205 07:15:27.029763 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-r9mh5" event={"ID":"351c9912-1ac6-4b8b-b009-d073b1bd28a5","Type":"ContainerStarted","Data":"7f7b0ede8b5e2597018bacdf7e5eea953f7d60e2a351c2c3adda16d9e0bbf117"} Dec 05 07:15:27 crc kubenswrapper[4997]: I1205 07:15:27.235314 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56jjnr\" (UID: \"64d88cd2-49b5-439a-9344-bb9dc4ec5730\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" Dec 05 07:15:27 crc kubenswrapper[4997]: E1205 07:15:27.235526 4997 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 07:15:27 crc kubenswrapper[4997]: E1205 07:15:27.235882 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert podName:64d88cd2-49b5-439a-9344-bb9dc4ec5730 nodeName:}" failed. No retries permitted until 2025-12-05 07:15:29.235804143 +0000 UTC m=+1229.764711404 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f56jjnr" (UID: "64d88cd2-49b5-439a-9344-bb9dc4ec5730") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 07:15:27 crc kubenswrapper[4997]: I1205 07:15:27.793587 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-rnq5d"] Dec 05 07:15:27 crc kubenswrapper[4997]: I1205 07:15:27.834566 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-r4m8t"] Dec 05 07:15:27 crc kubenswrapper[4997]: W1205 07:15:27.843422 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe84dde4_76f7_4df6_a6fd_d1b45d097e9b.slice/crio-364b0f4558b7877749c1751962bcf5d81e7413c5dbe58714554a547c7d6a5f6e WatchSource:0}: Error finding container 364b0f4558b7877749c1751962bcf5d81e7413c5dbe58714554a547c7d6a5f6e: Status 404 returned error can't find the container with id 364b0f4558b7877749c1751962bcf5d81e7413c5dbe58714554a547c7d6a5f6e Dec 05 07:15:27 crc kubenswrapper[4997]: I1205 07:15:27.873018 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-lpx4g"] Dec 05 07:15:27 crc kubenswrapper[4997]: W1205 07:15:27.873435 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5054dc6_123a_4d45_8340_5315c0c57e32.slice/crio-9752ca3f61295a771b999d6db4c0e22a9be07fe9aba38c83b84413a3421e51a3 WatchSource:0}: Error finding container 9752ca3f61295a771b999d6db4c0e22a9be07fe9aba38c83b84413a3421e51a3: Status 404 returned error can't find the container with id 9752ca3f61295a771b999d6db4c0e22a9be07fe9aba38c83b84413a3421e51a3 Dec 05 07:15:27 crc kubenswrapper[4997]: I1205 07:15:27.906933 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:27 crc kubenswrapper[4997]: I1205 07:15:27.907081 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:27 crc kubenswrapper[4997]: E1205 07:15:27.907238 4997 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 07:15:27 crc kubenswrapper[4997]: E1205 07:15:27.907302 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs podName:6c477704-084c-4d28-8ab3-cd0f29ca488f nodeName:}" failed. No retries permitted until 2025-12-05 07:15:29.907285642 +0000 UTC m=+1230.436192893 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-7wbfk" (UID: "6c477704-084c-4d28-8ab3-cd0f29ca488f") : secret "metrics-server-cert" not found Dec 05 07:15:27 crc kubenswrapper[4997]: E1205 07:15:27.907328 4997 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 07:15:27 crc kubenswrapper[4997]: E1205 07:15:27.907455 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs podName:6c477704-084c-4d28-8ab3-cd0f29ca488f nodeName:}" failed. No retries permitted until 2025-12-05 07:15:29.907400646 +0000 UTC m=+1230.436307907 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-7wbfk" (UID: "6c477704-084c-4d28-8ab3-cd0f29ca488f") : secret "webhook-server-cert" not found Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.039072 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r4m8t" event={"ID":"fe84dde4-76f7-4df6-a6fd-d1b45d097e9b","Type":"ContainerStarted","Data":"364b0f4558b7877749c1751962bcf5d81e7413c5dbe58714554a547c7d6a5f6e"} Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.050944 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-lpx4g" event={"ID":"d5054dc6-123a-4d45-8340-5315c0c57e32","Type":"ContainerStarted","Data":"9752ca3f61295a771b999d6db4c0e22a9be07fe9aba38c83b84413a3421e51a3"} Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.057451 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-rnq5d" event={"ID":"27e5617b-9979-45b2-a526-bb671edf63c2","Type":"ContainerStarted","Data":"011ef4a4f65527b4644528d624a3ab00331cf3314047f375fa92b63aa2ee072a"} Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.255842 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-5zk9v"] Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.268851 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tpg29"] Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.304893 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-sbfxr"] Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.338315 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-jk2cb"] Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.350004 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rq99q"] Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.365667 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-q62zq"] Dec 05 07:15:28 crc kubenswrapper[4997]: W1205 07:15:28.373434 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9d46c14_6c5e_42d4_81f3_43e7169d23a0.slice/crio-f5eb8f8969bbcad408d9bbbfd60fd1c121e6107bcf80c0b101080903202d3d18 WatchSource:0}: Error finding container f5eb8f8969bbcad408d9bbbfd60fd1c121e6107bcf80c0b101080903202d3d18: Status 404 returned error can't find the container with id f5eb8f8969bbcad408d9bbbfd60fd1c121e6107bcf80c0b101080903202d3d18 Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.373506 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-r9knp"] Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.391712 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-qppfb"] Dec 05 07:15:28 crc kubenswrapper[4997]: W1205 07:15:28.417232 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod905ae077_9cc3_41a9_b03f_77b3eeb052b3.slice/crio-63d6b6676b8abc6698de5a1430e9c04873231d3e9043d751fd79973c15b65aea WatchSource:0}: Error finding container 63d6b6676b8abc6698de5a1430e9c04873231d3e9043d751fd79973c15b65aea: Status 404 returned error can't find the container with id 63d6b6676b8abc6698de5a1430e9c04873231d3e9043d751fd79973c15b65aea Dec 05 07:15:28 crc kubenswrapper[4997]: W1205 07:15:28.421705 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1874b23_363f_4303_bfeb_86282b20963d.slice/crio-363100ef139e7b029b0edd2dea2a2512c7ae51bde76f03b589f199ec88af87aa WatchSource:0}: Error finding container 363100ef139e7b029b0edd2dea2a2512c7ae51bde76f03b589f199ec88af87aa: Status 404 returned error can't find the container with id 363100ef139e7b029b0edd2dea2a2512c7ae51bde76f03b589f199ec88af87aa Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.422758 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-skcxd"] Dec 05 07:15:28 crc kubenswrapper[4997]: W1205 07:15:28.424874 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40df044a_6922_43cf_ae37_ebc8dd637119.slice/crio-640f8c8837333a8447789bf09c7f0ad57e8f793e420a92be747c5d9efbe68569 WatchSource:0}: Error finding container 640f8c8837333a8447789bf09c7f0ad57e8f793e420a92be747c5d9efbe68569: Status 404 returned error can't find the container with id 640f8c8837333a8447789bf09c7f0ad57e8f793e420a92be747c5d9efbe68569 Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.435512 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-ztwsn"] Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.438331 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5rm5x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-cjggt_openstack-operators(e1874b23-363f-4303-bfeb-86282b20963d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.440570 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5rm5x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-cjggt_openstack-operators(e1874b23-363f-4303-bfeb-86282b20963d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.441937 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-cjggt" podUID="e1874b23-363f-4303-bfeb-86282b20963d" Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.443063 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l6wtb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-qsqst_openstack-operators(8d31795c-bf5a-48fc-85cc-0c1385c2d7b0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.443497 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-whvgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-hrs7p_openstack-operators(3f6028fa-7b09-49ec-a8ed-4f4b25419e19): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.446961 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qsqst" podUID="8d31795c-bf5a-48fc-85cc-0c1385c2d7b0" Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.447812 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-whvgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-hrs7p_openstack-operators(3f6028fa-7b09-49ec-a8ed-4f4b25419e19): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.449898 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hrs7p" podUID="3f6028fa-7b09-49ec-a8ed-4f4b25419e19" Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.451143 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nbng9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-skcxd_openstack-operators(f309b30b-9712-4175-b237-7c35ac2e5dee): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.451291 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9k7j6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-jtjgd_openstack-operators(9597d662-50ea-4066-8bbc-232cc1d90dbd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.453369 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9k7j6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-jtjgd_openstack-operators(9597d662-50ea-4066-8bbc-232cc1d90dbd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.454154 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nbng9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-skcxd_openstack-operators(f309b30b-9712-4175-b237-7c35ac2e5dee): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.454488 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-jtjgd" podUID="9597d662-50ea-4066-8bbc-232cc1d90dbd" Dec 05 07:15:28 crc kubenswrapper[4997]: W1205 07:15:28.454776 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25ca1d60_a629_447e_afa7_7a9f0e6f0b7e.slice/crio-bc46cce0adb9c82695a81b501576de6da6252befcb9ec7f70058adad7c80f52e WatchSource:0}: Error finding container bc46cce0adb9c82695a81b501576de6da6252befcb9ec7f70058adad7c80f52e: Status 404 returned error can't find the container with id bc46cce0adb9c82695a81b501576de6da6252befcb9ec7f70058adad7c80f52e Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.455781 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skcxd" podUID="f309b30b-9712-4175-b237-7c35ac2e5dee" Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.458671 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-88hm7"] Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.460819 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q6bhp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-kbfbg_openstack-operators(25ca1d60-a629-447e-afa7-7a9f0e6f0b7e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.463575 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q6bhp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-kbfbg_openstack-operators(25ca1d60-a629-447e-afa7-7a9f0e6f0b7e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.464983 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-kbfbg" podUID="25ca1d60-a629-447e-afa7-7a9f0e6f0b7e" Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.473875 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-cjggt"] Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.484114 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-kbfbg"] Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.491039 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qsqst"] Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.500562 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-hrs7p"] Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.507500 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-jtjgd"] Dec 05 07:15:28 crc kubenswrapper[4997]: I1205 07:15:28.628945 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert\") pod \"infra-operator-controller-manager-57548d458d-pppgg\" (UID: \"6aac09d2-991d-45b9-bba9-4cf2d70278ab\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.629232 4997 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 07:15:28 crc kubenswrapper[4997]: E1205 07:15:28.629346 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert podName:6aac09d2-991d-45b9-bba9-4cf2d70278ab nodeName:}" failed. No retries permitted until 2025-12-05 07:15:32.62932013 +0000 UTC m=+1233.158227391 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert") pod "infra-operator-controller-manager-57548d458d-pppgg" (UID: "6aac09d2-991d-45b9-bba9-4cf2d70278ab") : secret "infra-operator-webhook-server-cert" not found Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.065544 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-kbfbg" event={"ID":"25ca1d60-a629-447e-afa7-7a9f0e6f0b7e","Type":"ContainerStarted","Data":"bc46cce0adb9c82695a81b501576de6da6252befcb9ec7f70058adad7c80f52e"} Dec 05 07:15:29 crc kubenswrapper[4997]: E1205 07:15:29.069155 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-kbfbg" podUID="25ca1d60-a629-447e-afa7-7a9f0e6f0b7e" Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.071094 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88hm7" event={"ID":"40df044a-6922-43cf-ae37-ebc8dd637119","Type":"ContainerStarted","Data":"640f8c8837333a8447789bf09c7f0ad57e8f793e420a92be747c5d9efbe68569"} Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.075813 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-qppfb" event={"ID":"7afb5248-e41e-46ec-9e46-ed9fcc7cf8d3","Type":"ContainerStarted","Data":"90d0980390798b3312901c4e230704b1adce43261c98798f904a68dfe0f04bb3"} Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.088830 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-jk2cb" event={"ID":"7db21a3e-655e-4874-8366-dfdb5d2521f5","Type":"ContainerStarted","Data":"bd7c97328f50efbf93a78bcf73df0c09578cae7d7885cea9cc89ca5846512b75"} Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.105271 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-cjggt" event={"ID":"e1874b23-363f-4303-bfeb-86282b20963d","Type":"ContainerStarted","Data":"363100ef139e7b029b0edd2dea2a2512c7ae51bde76f03b589f199ec88af87aa"} Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.108460 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-sbfxr" event={"ID":"3580ba04-55a0-4e33-aa08-932a2648bf2d","Type":"ContainerStarted","Data":"14ab8780b84506500df53d181505f700cedf79d7f0c39b8b8789d2f90f624dab"} Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.110995 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hrs7p" event={"ID":"3f6028fa-7b09-49ec-a8ed-4f4b25419e19","Type":"ContainerStarted","Data":"41b9431d03c637ac289be6b84c620330b954580c1ce2d3311cbb32259668e977"} Dec 05 07:15:29 crc kubenswrapper[4997]: E1205 07:15:29.111205 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-cjggt" podUID="e1874b23-363f-4303-bfeb-86282b20963d" Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.113572 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-jtjgd" event={"ID":"9597d662-50ea-4066-8bbc-232cc1d90dbd","Type":"ContainerStarted","Data":"8174e3e030a757f196762ab9918b85f33e2df08ad0345b06566b5dd9b5c6f0cc"} Dec 05 07:15:29 crc kubenswrapper[4997]: E1205 07:15:29.115018 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hrs7p" podUID="3f6028fa-7b09-49ec-a8ed-4f4b25419e19" Dec 05 07:15:29 crc kubenswrapper[4997]: E1205 07:15:29.115652 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-jtjgd" podUID="9597d662-50ea-4066-8bbc-232cc1d90dbd" Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.117130 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tpg29" event={"ID":"e5f7e30e-e2a8-4131-9c10-1dddfe528684","Type":"ContainerStarted","Data":"36f9f572b2a085eae83ff531e8dfce983c41ad9b24097cf9df412d14f07e940d"} Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.120057 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skcxd" event={"ID":"f309b30b-9712-4175-b237-7c35ac2e5dee","Type":"ContainerStarted","Data":"e0a4041036924ee1f521468c71da62d845a58b14401c06627c72829e2eeb8c17"} Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.129958 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-ztwsn" event={"ID":"905ae077-9cc3-41a9-b03f-77b3eeb052b3","Type":"ContainerStarted","Data":"63d6b6676b8abc6698de5a1430e9c04873231d3e9043d751fd79973c15b65aea"} Dec 05 07:15:29 crc kubenswrapper[4997]: E1205 07:15:29.146769 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skcxd" podUID="f309b30b-9712-4175-b237-7c35ac2e5dee" Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.156757 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-5zk9v" event={"ID":"c488dae2-6bbc-434a-a1c8-d3bd6ba5a549","Type":"ContainerStarted","Data":"b58e281695ae41e40caa9120c6fe966dd48392fb7d4421c502a52938c474a996"} Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.159294 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-r9knp" event={"ID":"9cc09cce-bafa-4f17-bd45-b767fa576e54","Type":"ContainerStarted","Data":"92750109543a245cfacdedc48fe5af16915c87d98bb29c32a6f0af4beb10e01f"} Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.183426 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qsqst" event={"ID":"8d31795c-bf5a-48fc-85cc-0c1385c2d7b0","Type":"ContainerStarted","Data":"08e5b73745de1ea67d940155b9f6f6b67d228656b569ecff44c24c5b39a29d20"} Dec 05 07:15:29 crc kubenswrapper[4997]: E1205 07:15:29.191957 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qsqst" podUID="8d31795c-bf5a-48fc-85cc-0c1385c2d7b0" Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.195500 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q62zq" event={"ID":"c9d46c14-6c5e-42d4-81f3-43e7169d23a0","Type":"ContainerStarted","Data":"f5eb8f8969bbcad408d9bbbfd60fd1c121e6107bcf80c0b101080903202d3d18"} Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.197287 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rq99q" event={"ID":"7a8b9284-3f13-4396-8f85-36f0c59e1ba3","Type":"ContainerStarted","Data":"6fcfd087ac98279074a5e8b201c0283660f6841734d45e3612e23b766595c325"} Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.253481 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56jjnr\" (UID: \"64d88cd2-49b5-439a-9344-bb9dc4ec5730\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" Dec 05 07:15:29 crc kubenswrapper[4997]: E1205 07:15:29.253703 4997 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 07:15:29 crc kubenswrapper[4997]: E1205 07:15:29.253769 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert podName:64d88cd2-49b5-439a-9344-bb9dc4ec5730 nodeName:}" failed. No retries permitted until 2025-12-05 07:15:33.253750645 +0000 UTC m=+1233.782657906 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f56jjnr" (UID: "64d88cd2-49b5-439a-9344-bb9dc4ec5730") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.980479 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:29 crc kubenswrapper[4997]: I1205 07:15:29.981048 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:29 crc kubenswrapper[4997]: E1205 07:15:29.981479 4997 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 07:15:29 crc kubenswrapper[4997]: E1205 07:15:29.981723 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs podName:6c477704-084c-4d28-8ab3-cd0f29ca488f nodeName:}" failed. No retries permitted until 2025-12-05 07:15:33.981700702 +0000 UTC m=+1234.510607963 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-7wbfk" (UID: "6c477704-084c-4d28-8ab3-cd0f29ca488f") : secret "webhook-server-cert" not found Dec 05 07:15:29 crc kubenswrapper[4997]: E1205 07:15:29.982215 4997 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 07:15:29 crc kubenswrapper[4997]: E1205 07:15:29.982324 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs podName:6c477704-084c-4d28-8ab3-cd0f29ca488f nodeName:}" failed. No retries permitted until 2025-12-05 07:15:33.982313469 +0000 UTC m=+1234.511220730 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-7wbfk" (UID: "6c477704-084c-4d28-8ab3-cd0f29ca488f") : secret "metrics-server-cert" not found Dec 05 07:15:30 crc kubenswrapper[4997]: E1205 07:15:30.245867 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hrs7p" podUID="3f6028fa-7b09-49ec-a8ed-4f4b25419e19" Dec 05 07:15:30 crc kubenswrapper[4997]: E1205 07:15:30.245969 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qsqst" podUID="8d31795c-bf5a-48fc-85cc-0c1385c2d7b0" Dec 05 07:15:30 crc kubenswrapper[4997]: E1205 07:15:30.246025 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-jtjgd" podUID="9597d662-50ea-4066-8bbc-232cc1d90dbd" Dec 05 07:15:30 crc kubenswrapper[4997]: E1205 07:15:30.247261 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-kbfbg" podUID="25ca1d60-a629-447e-afa7-7a9f0e6f0b7e" Dec 05 07:15:30 crc kubenswrapper[4997]: E1205 07:15:30.247351 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-cjggt" podUID="e1874b23-363f-4303-bfeb-86282b20963d" Dec 05 07:15:30 crc kubenswrapper[4997]: E1205 07:15:30.267630 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skcxd" podUID="f309b30b-9712-4175-b237-7c35ac2e5dee" Dec 05 07:15:32 crc kubenswrapper[4997]: I1205 07:15:32.670581 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert\") pod \"infra-operator-controller-manager-57548d458d-pppgg\" (UID: \"6aac09d2-991d-45b9-bba9-4cf2d70278ab\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" Dec 05 07:15:32 crc kubenswrapper[4997]: E1205 07:15:32.671085 4997 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 07:15:32 crc kubenswrapper[4997]: E1205 07:15:32.671154 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert podName:6aac09d2-991d-45b9-bba9-4cf2d70278ab nodeName:}" failed. No retries permitted until 2025-12-05 07:15:40.671135423 +0000 UTC m=+1241.200042684 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert") pod "infra-operator-controller-manager-57548d458d-pppgg" (UID: "6aac09d2-991d-45b9-bba9-4cf2d70278ab") : secret "infra-operator-webhook-server-cert" not found Dec 05 07:15:33 crc kubenswrapper[4997]: I1205 07:15:33.333743 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56jjnr\" (UID: \"64d88cd2-49b5-439a-9344-bb9dc4ec5730\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" Dec 05 07:15:33 crc kubenswrapper[4997]: E1205 07:15:33.336911 4997 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 07:15:33 crc kubenswrapper[4997]: E1205 07:15:33.337022 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert podName:64d88cd2-49b5-439a-9344-bb9dc4ec5730 nodeName:}" failed. No retries permitted until 2025-12-05 07:15:41.336999035 +0000 UTC m=+1241.865906296 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f56jjnr" (UID: "64d88cd2-49b5-439a-9344-bb9dc4ec5730") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 07:15:34 crc kubenswrapper[4997]: I1205 07:15:34.011993 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:34 crc kubenswrapper[4997]: I1205 07:15:34.012123 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:34 crc kubenswrapper[4997]: E1205 07:15:34.012224 4997 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 07:15:34 crc kubenswrapper[4997]: E1205 07:15:34.012293 4997 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 07:15:34 crc kubenswrapper[4997]: E1205 07:15:34.012344 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs podName:6c477704-084c-4d28-8ab3-cd0f29ca488f nodeName:}" failed. No retries permitted until 2025-12-05 07:15:42.012314103 +0000 UTC m=+1242.541221544 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-7wbfk" (UID: "6c477704-084c-4d28-8ab3-cd0f29ca488f") : secret "metrics-server-cert" not found Dec 05 07:15:34 crc kubenswrapper[4997]: E1205 07:15:34.012370 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs podName:6c477704-084c-4d28-8ab3-cd0f29ca488f nodeName:}" failed. No retries permitted until 2025-12-05 07:15:42.012359584 +0000 UTC m=+1242.541267055 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-7wbfk" (UID: "6c477704-084c-4d28-8ab3-cd0f29ca488f") : secret "webhook-server-cert" not found Dec 05 07:15:40 crc kubenswrapper[4997]: I1205 07:15:40.735182 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert\") pod \"infra-operator-controller-manager-57548d458d-pppgg\" (UID: \"6aac09d2-991d-45b9-bba9-4cf2d70278ab\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" Dec 05 07:15:40 crc kubenswrapper[4997]: I1205 07:15:40.755884 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aac09d2-991d-45b9-bba9-4cf2d70278ab-cert\") pod \"infra-operator-controller-manager-57548d458d-pppgg\" (UID: \"6aac09d2-991d-45b9-bba9-4cf2d70278ab\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" Dec 05 07:15:40 crc kubenswrapper[4997]: I1205 07:15:40.786982 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" Dec 05 07:15:41 crc kubenswrapper[4997]: I1205 07:15:41.348648 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56jjnr\" (UID: \"64d88cd2-49b5-439a-9344-bb9dc4ec5730\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" Dec 05 07:15:41 crc kubenswrapper[4997]: I1205 07:15:41.352684 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d88cd2-49b5-439a-9344-bb9dc4ec5730-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f56jjnr\" (UID: \"64d88cd2-49b5-439a-9344-bb9dc4ec5730\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" Dec 05 07:15:41 crc kubenswrapper[4997]: I1205 07:15:41.386994 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" Dec 05 07:15:42 crc kubenswrapper[4997]: I1205 07:15:42.063423 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:42 crc kubenswrapper[4997]: I1205 07:15:42.063536 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:42 crc kubenswrapper[4997]: I1205 07:15:42.067966 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:42 crc kubenswrapper[4997]: I1205 07:15:42.072936 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c477704-084c-4d28-8ab3-cd0f29ca488f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-7wbfk\" (UID: \"6c477704-084c-4d28-8ab3-cd0f29ca488f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:42 crc kubenswrapper[4997]: I1205 07:15:42.177405 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:43 crc kubenswrapper[4997]: E1205 07:15:43.229472 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801" Dec 05 07:15:43 crc kubenswrapper[4997]: E1205 07:15:43.229710 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5c2hp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-rnq5d_openstack-operators(27e5617b-9979-45b2-a526-bb671edf63c2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:15:44 crc kubenswrapper[4997]: E1205 07:15:44.031710 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 05 07:15:44 crc kubenswrapper[4997]: E1205 07:15:44.032484 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f64br,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-ztwsn_openstack-operators(905ae077-9cc3-41a9-b03f-77b3eeb052b3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:15:44 crc kubenswrapper[4997]: E1205 07:15:44.724762 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 05 07:15:44 crc kubenswrapper[4997]: E1205 07:15:44.724994 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wjtmj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-jk2cb_openstack-operators(7db21a3e-655e-4874-8366-dfdb5d2521f5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:15:45 crc kubenswrapper[4997]: E1205 07:15:45.402057 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530" Dec 05 07:15:45 crc kubenswrapper[4997]: E1205 07:15:45.402422 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k2l9p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-q62zq_openstack-operators(c9d46c14-6c5e-42d4-81f3-43e7169d23a0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:15:45 crc kubenswrapper[4997]: E1205 07:15:45.943087 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809" Dec 05 07:15:45 crc kubenswrapper[4997]: E1205 07:15:45.944011 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5gwcg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-lpx4g_openstack-operators(d5054dc6-123a-4d45-8340-5315c0c57e32): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:15:46 crc kubenswrapper[4997]: E1205 07:15:46.759065 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 05 07:15:46 crc kubenswrapper[4997]: E1205 07:15:46.759350 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hc88k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-88hm7_openstack-operators(40df044a-6922-43cf-ae37-ebc8dd637119): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:15:47 crc kubenswrapper[4997]: E1205 07:15:47.240354 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 05 07:15:47 crc kubenswrapper[4997]: E1205 07:15:47.240642 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xgmf7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-qppfb_openstack-operators(7afb5248-e41e-46ec-9e46-ed9fcc7cf8d3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:15:49 crc kubenswrapper[4997]: I1205 07:15:49.770260 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:15:49 crc kubenswrapper[4997]: I1205 07:15:49.770788 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:15:49 crc kubenswrapper[4997]: I1205 07:15:49.770852 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 07:15:49 crc kubenswrapper[4997]: I1205 07:15:49.771742 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6cb7be4205d2b6d3befa7784153570d8d5c45d7400acdc190e0c1608d52679b2"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 07:15:49 crc kubenswrapper[4997]: I1205 07:15:49.771824 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://6cb7be4205d2b6d3befa7784153570d8d5c45d7400acdc190e0c1608d52679b2" gracePeriod=600 Dec 05 07:15:50 crc kubenswrapper[4997]: I1205 07:15:50.469849 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="6cb7be4205d2b6d3befa7784153570d8d5c45d7400acdc190e0c1608d52679b2" exitCode=0 Dec 05 07:15:50 crc kubenswrapper[4997]: I1205 07:15:50.469913 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"6cb7be4205d2b6d3befa7784153570d8d5c45d7400acdc190e0c1608d52679b2"} Dec 05 07:15:50 crc kubenswrapper[4997]: I1205 07:15:50.469965 4997 scope.go:117] "RemoveContainer" containerID="046abc9c3b110cde88610f054e681c42bd85fac9c9e032fa744fe8940f595f0c" Dec 05 07:15:51 crc kubenswrapper[4997]: I1205 07:15:51.118509 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-pppgg"] Dec 05 07:15:55 crc kubenswrapper[4997]: I1205 07:15:55.522051 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" event={"ID":"6aac09d2-991d-45b9-bba9-4cf2d70278ab","Type":"ContainerStarted","Data":"da608772f3b710a16d030ea2d0530550d58cc49ff3ec7cb6352f71b59f8cecc2"} Dec 05 07:15:55 crc kubenswrapper[4997]: I1205 07:15:55.559716 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr"] Dec 05 07:15:55 crc kubenswrapper[4997]: I1205 07:15:55.633874 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk"] Dec 05 07:15:55 crc kubenswrapper[4997]: W1205 07:15:55.979082 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c477704_084c_4d28_8ab3_cd0f29ca488f.slice/crio-d6b04c0e0285b021cb6aa1c3fc06ba3c62075ce0be06265a81444930de03c73a WatchSource:0}: Error finding container d6b04c0e0285b021cb6aa1c3fc06ba3c62075ce0be06265a81444930de03c73a: Status 404 returned error can't find the container with id d6b04c0e0285b021cb6aa1c3fc06ba3c62075ce0be06265a81444930de03c73a Dec 05 07:15:55 crc kubenswrapper[4997]: W1205 07:15:55.981776 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64d88cd2_49b5_439a_9344_bb9dc4ec5730.slice/crio-005f6fc3327a53ce734a98fabffc8fd9478e158f8c086dd0f7e49bf432a2df59 WatchSource:0}: Error finding container 005f6fc3327a53ce734a98fabffc8fd9478e158f8c086dd0f7e49bf432a2df59: Status 404 returned error can't find the container with id 005f6fc3327a53ce734a98fabffc8fd9478e158f8c086dd0f7e49bf432a2df59 Dec 05 07:15:56 crc kubenswrapper[4997]: I1205 07:15:56.544315 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-r9knp" event={"ID":"9cc09cce-bafa-4f17-bd45-b767fa576e54","Type":"ContainerStarted","Data":"e3afc32da52c8c13f83fd20d07815033fe38b18bcf879027c4e69f4e42440e76"} Dec 05 07:15:56 crc kubenswrapper[4997]: I1205 07:15:56.556504 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tpg29" event={"ID":"e5f7e30e-e2a8-4131-9c10-1dddfe528684","Type":"ContainerStarted","Data":"695fd9755eb91158711b707b0957bad87fb33e41c60dee51aa94518d43ba46c1"} Dec 05 07:15:56 crc kubenswrapper[4997]: I1205 07:15:56.558385 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rq99q" event={"ID":"7a8b9284-3f13-4396-8f85-36f0c59e1ba3","Type":"ContainerStarted","Data":"396e5e3d24c07009ea899b7bb4f2c10caf3c3bfe82fa7c20768d8a2787e9925b"} Dec 05 07:15:56 crc kubenswrapper[4997]: I1205 07:15:56.559571 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" event={"ID":"6c477704-084c-4d28-8ab3-cd0f29ca488f","Type":"ContainerStarted","Data":"d6b04c0e0285b021cb6aa1c3fc06ba3c62075ce0be06265a81444930de03c73a"} Dec 05 07:15:56 crc kubenswrapper[4997]: I1205 07:15:56.560986 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-r9mh5" event={"ID":"351c9912-1ac6-4b8b-b009-d073b1bd28a5","Type":"ContainerStarted","Data":"9a5c97419e72e03ba07504bd5299ebf677e63f612e0b7b2310eea2690b816845"} Dec 05 07:15:56 crc kubenswrapper[4997]: I1205 07:15:56.568527 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-sbfxr" event={"ID":"3580ba04-55a0-4e33-aa08-932a2648bf2d","Type":"ContainerStarted","Data":"7fb6b52a15690d224dfd66d069ebbd50475b529f3ed31943fa9b6216c351ae66"} Dec 05 07:15:56 crc kubenswrapper[4997]: I1205 07:15:56.574231 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"1b8d25609ffa42f5e50251e3eb59f7240e05fe03de714e77c29a98e4dc7b4510"} Dec 05 07:15:56 crc kubenswrapper[4997]: I1205 07:15:56.576755 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-5zk9v" event={"ID":"c488dae2-6bbc-434a-a1c8-d3bd6ba5a549","Type":"ContainerStarted","Data":"f548593ad31a07c9c1f480b2e4cf90efd52851fb4de108c4169ce77ca30c92f1"} Dec 05 07:15:56 crc kubenswrapper[4997]: I1205 07:15:56.577894 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" event={"ID":"64d88cd2-49b5-439a-9344-bb9dc4ec5730","Type":"ContainerStarted","Data":"005f6fc3327a53ce734a98fabffc8fd9478e158f8c086dd0f7e49bf432a2df59"} Dec 05 07:15:57 crc kubenswrapper[4997]: I1205 07:15:57.588852 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r4m8t" event={"ID":"fe84dde4-76f7-4df6-a6fd-d1b45d097e9b","Type":"ContainerStarted","Data":"f022582b541641552aeba8493ef739650dae29695d61d710b089876622ef50cd"} Dec 05 07:15:57 crc kubenswrapper[4997]: I1205 07:15:57.591803 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" event={"ID":"6c477704-084c-4d28-8ab3-cd0f29ca488f","Type":"ContainerStarted","Data":"bd5c75bb1421937b54e87b5be7d5ddef480d162cf01915d062c4778d120b7b0f"} Dec 05 07:15:57 crc kubenswrapper[4997]: I1205 07:15:57.757171 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" podStartSLOduration=32.75714504 podStartE2EDuration="32.75714504s" podCreationTimestamp="2025-12-05 07:15:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:15:57.756721519 +0000 UTC m=+1258.285628780" watchObservedRunningTime="2025-12-05 07:15:57.75714504 +0000 UTC m=+1258.286052301" Dec 05 07:15:58 crc kubenswrapper[4997]: I1205 07:15:58.602678 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:15:59 crc kubenswrapper[4997]: I1205 07:15:59.628178 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skcxd" event={"ID":"f309b30b-9712-4175-b237-7c35ac2e5dee","Type":"ContainerStarted","Data":"bffea050d708e8aed8b25ed5607c7e7a44d75ca264a2b92cb57df3e4809ba82a"} Dec 05 07:16:02 crc kubenswrapper[4997]: I1205 07:16:02.184642 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-7wbfk" Dec 05 07:16:07 crc kubenswrapper[4997]: I1205 07:16:07.687894 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-jtjgd" event={"ID":"9597d662-50ea-4066-8bbc-232cc1d90dbd","Type":"ContainerStarted","Data":"a5b253ece0735181ff9b736fce700ecf6c2d8604035513bddb8ff02b001d8278"} Dec 05 07:16:08 crc kubenswrapper[4997]: I1205 07:16:08.696188 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-cjggt" event={"ID":"e1874b23-363f-4303-bfeb-86282b20963d","Type":"ContainerStarted","Data":"e3363aa130c0c6f291a5c4b58101525169f8abb98a62462b05197631974a7b54"} Dec 05 07:16:08 crc kubenswrapper[4997]: I1205 07:16:08.698034 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-kbfbg" event={"ID":"25ca1d60-a629-447e-afa7-7a9f0e6f0b7e","Type":"ContainerStarted","Data":"057cbcb286f9fb2c25fc08e52ccabdc042778541a2ae2e72e9a86c1b7a019bde"} Dec 05 07:16:11 crc kubenswrapper[4997]: E1205 07:16:11.234456 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 07:16:11 crc kubenswrapper[4997]: E1205 07:16:11.235341 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5c2hp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-rnq5d_openstack-operators(27e5617b-9979-45b2-a526-bb671edf63c2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:16:11 crc kubenswrapper[4997]: E1205 07:16:11.236534 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-rnq5d" podUID="27e5617b-9979-45b2-a526-bb671edf63c2" Dec 05 07:16:11 crc kubenswrapper[4997]: E1205 07:16:11.764201 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81" Dec 05 07:16:11 crc kubenswrapper[4997]: E1205 07:16:11.764834 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:51004bad441b97668eff122dd7b0cc5bdedfa185ba4d7533d9ff84d5ee9d51e2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:2444fe898df68969a7978bb84fd12c3c61dc371f264156ff0a877d8aab1f9f4e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api@sha256:2d87021f2f291525dda4c17e8fcd2fbef60780450d7941be423bcfd4047cabd2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator@sha256:3473a5f5c914f9ba397ffc5ea9d8eeedd85d31a3c9244df7457f3c3e74eaefc4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener@sha256:c1c8f583529e123a7105ebc2249ab19267313f30138867840d1e65b9390f1886,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier@sha256:8dcd62d8f75c4dbf0afc27fa96cd481c56d8fb174fa29abafa0d29616eded790,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24@sha256:6b929971283d69f485a7d3e449fb5a3dd65d5a4de585c73419e776821d00062c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:82006b9c64d4c5f80483cda262d960ce6be4813665158ef1a53ea7734bbe431f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener@sha256:a76d2c46403c03704dcfe7de49454496300d60d849ee81076d8637b272043c69,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker@sha256:d2fbe075d21195b746fd27a073dbd249d38b3c4f81c30d162770a338fb87e338,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:43a24796dabde68270dbfefa107205e173fdd6a0dc701502858cadbede69da31,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute@sha256:667029e1ec7e63fffa1a096f432f6160b441ba36df1bddc9066cbd1129b82009,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi@sha256:b2785dbc3ceaa930dff8068bbb8654af2e0b40a9c2632300641cb8348e9cf43d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter@sha256:7211a617ec657701ca819aa0ba28e1d5750f5bf2c1391b755cc4a48cc360b0fa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification@sha256:f17b61f2318b74648e174d73dd31deee6c0d1434605c9f32707aedf2f4378957,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:0b08861590e3646584af0fc7c7d8a743a35b4f5964d6fd355f206daa9ae999ce,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:e26fb8ad7808ca8efe268881f9229df90a755b24bd4ad5501ba3b8c5c16987a3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:cfeb4e264c00408dee5196b06003722b6dda540a3f26d3ff90abfd795729833b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api@sha256:e02c97e990781e27d0bc5319781ee19618cdb2997adea3df57376cbda9896b55,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor@sha256:1fe03701929d2f30e832a3831c87d8046806e2c35545aebe94f4a2849b1f8e67,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api@sha256:2f9748f10c87efbee801c70f46b3dc5c6532ca070af558a4fb45cb34dbbb6f04,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9@sha256:1d69ad383cb03ef808c1f737427c5ca2385e28a3af1861a4336b6e539b346c27,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central@sha256:112fed4b9de0ccf15011e8a3a26ce6efbbe8e7d8eb3d4153d1a1874b9bde6d68,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns@sha256:aa87158aeb1194f4940126197b912ea972fafe12ea5c1f89a07d6ccfafc16f77,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer@sha256:fcd3bf8112793023be72845ce3a984beabd5a3cb369c11252130076ed38b3770,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound@sha256:fee9fc72864ee217aace1cf11cb090ef41935841f9c60127d775dc2989330777,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker@sha256:239967aef48587f275c9636d8f89e476d909dbba57fea64d8196ddacf6817450,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr@sha256:7a0ade11985653bb8ad2646b0848eb6f7128d21d85b99551ac17f74293087a30,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid@sha256:8ab175d7ee42e22e0ca1ebf98d180112428758a86ef8adccaba8f3653567f6ab,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler@sha256:581b65b646301e0fcb07582150ba63438f1353a85bf9acf1eb2acb4ce71c58bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron@sha256:d7e43361d50d1e7d4c99e499eee56aa50591855836638742666303dc59096258,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd@sha256:5b59d54dc4a23373a5172f15f5497b287422c32f5702efd1e171c3f2048c9842,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent@sha256:f3227beee5b52411de42c6a37ceda7d8f68934b4671a2d661403f8c1c0eab6d6,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn@sha256:db3e3d71618c3539a2853a20f7684f016b67370157990932291b00a48fa16bd3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent@sha256:e6dfe5f67adec298afbb57aec95c9cf89b4757ccfea8d8be66ef0ffd8c58322f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent@sha256:ba46c29c79c92487b6b9f0db11a517269c6455b8b9786e9d2692f4e24e43d552,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent@sha256:4f1c6fcf33354f1cbbc914c1709310be2fa4fe0dd64e5dbf3f91d6f0634bd28f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:481073ac9deefb38bbd989aaa8dd7aedb4e0af26017f4883f85fce433380bf63,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api@sha256:ecf469bd360c2aa2e5eb57826585c19a10ebe9f683790803dc4989a46c11789e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn@sha256:d506b2ca02a16cdab757b38a86d40e0459094c7269067de89beb3edf4a50bf5e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:2f2aabcd1b45f9fb3034d28e9a49acac72d7917fd1bbfbbc498e69e8be0b7b2b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon@sha256:4f4b273dd4c6ead9bf640fb985d101a7c9adba388968fb1d71fbb08b0510eb9f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached@sha256:dad2336390cae6705133deefaa09c9e39512cf29133aa009006e3962c8022108,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis@sha256:21edb042683b37827463124ceb159fa316e8cf0ac6040dc464f5242300b9daad,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api@sha256:21334e97e6b4194d803a60d0ecfa33327bf248e7507683ea9dcb33a28a2ec858,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor@sha256:4deb460a113324762b3139301c6aacd48c57204d8d13eb1c387d7064ec19db0d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector@sha256:942f9cbe36d328caa5d68b398703b2be5d7b7dc2b034a72d2ae62416cb7be208,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent@sha256:9e2ae3ac44ed2495b0f4398d7419b1e8e1321bec32a0ab043aabf28aa8b33384,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe@sha256:7cb9e377fa81bbe84fcc006b27c45d56ea3d6ed2144fb9ebf5fb8df5b920d423,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent@sha256:9d930c44b5d90b140117dd05d976d10d29d93eed9a70118e594e00da64594562,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone@sha256:a7b6fa2f16a882674624b48939737e2bd95da7bef60db593a8e6e4d397fa516c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api@sha256:68714e821f8e4e2d905d6e5bc7fb2e713a24c02db48901fb2a11d57b80f6c584,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler@sha256:4a8b11fbc23e097869f8f347e78a409b294573732987dd8fa6493888a3ff68d2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share@sha256:57007fab45f2d8fbf929d26609a2e566fbcb006e05d78ca72b9d0b71af866305,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils@sha256:b0f8d8a4d29d8d4667205df4a94bacefcdd7a33981407c20bd7dd320f27308b7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api@sha256:255cc3471ee112b17da164148b0ec25678332061b5b488868b81a30e5afb5bb5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:d6189c79b326e4b09ccae1141528b03bc59b2533781a960e8f91f2a5dbb343d5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:a96d336d231eee461559cfe82b025874ce2b8652520297bc5143559694ebac58,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:eaf80338dc065eb9c8c1f40552793c7cc2ff052c88c789f0a5d3e34099549adb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:98a3cff4a3aae37148c4c982a0e37f21a476528cbd74734f59ae22f61fdb6fc1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:a7089bcd0a2dbc014b29391dbd14b3fbc3ba0abd0f36bd16cb3b594cfa001464,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager@sha256:36cc3ee813bccbfb639f17896bd98028521e3cc5740a5d07f91e119729a76a69,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping@sha256:61807c42b6197326d9483d65972029117cea6d373ae913fd359993d8e12fff13,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog@sha256:61dbee4a2559eda45dadf8d2b121cd85f79043d7cb2c1a62f176261042c3e39c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker@sha256:c4652e3a9c4275470c3ef1a2e4d20a420d9c7bdd5157b0bbdaafea3fa038dcab,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:c1b8da8298ec8be0ca22c7d8ba48da103e72dfe7ed5e9427b971d31eac3a8b33,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather@sha256:854a802357b4f565a366fce3bf29b20c1b768ec4ab7e822ef52dfc2fef000d2c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:dec5870172c510ae43ff98398260fe595288af59302709d71fc2a020763deb88,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:ebeb25c4a4ce978c741d166518070e05f0fd81c143bdc680ee1d8f5985ec8d6c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:1e53a53dfe9b3cb757e4d666e76c8989941eb4f0b98d629a7f697a1693aacb17,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:726a3df0e94cfdcef301fe88fa8d91972914ec2104fb6fa1d8e4c325981712a6,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:c8e13f116261ef06b59e9034c605f68d53eb6f760426c35ee6ed3785b97b1800,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:e554a5816081a60a0ae6fd1464c1f0a11cf2133707a4b220a023ecae7b302eed,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:f24234939afca841e46ea4d17bec959b63705ab0e75476465e777d44905c5f1b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account@sha256:4eb3a9c95f57df34ab88b952d8ad2057d60ac0aa4526a51070bea5d64e3aeeee,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:ae1279cd0af8af3863925d149db4c514dfda0c159a8084216b7228a35f238678,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object@sha256:fcb1f8a778d8cffa0f42efdcbde01061cb3aaaccc3453e65a4b213d553ad344c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:3c89899d53b3bca91830c259434c074f27554824a9cdcf117158c4a4329810f5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all@sha256:a1a7ba434daff518f09d8f4075e76308402e9b7a0b5b641ac2ef721fbf88752a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api@sha256:1ecb6e1be330877bf6dce091efe512045926c0dcb73b67615374ddf5c90adaee,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier@sha256:78d97df08b9931d90a2523fc4c1d670bdcd5480a6edf96a0d867565f3a6ab78f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine@sha256:6811871583a498f416300c9a5a2116907f428dbb3530c736c2243d8b6dec2bda,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5ffpx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-55c85496f56jjnr_openstack-operators(64d88cd2-49b5-439a-9344-bb9dc4ec5730): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.582205 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.582496 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mfw8t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-57548d458d-pppgg_openstack-operators(6aac09d2-991d-45b9-bba9-4cf2d70278ab): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.583311 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.583523 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hc88k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-88hm7_openstack-operators(40df044a-6922-43cf-ae37-ebc8dd637119): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.585039 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88hm7" podUID="40df044a-6922-43cf-ae37-ebc8dd637119" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.595263 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.595817 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wb6db,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-r4m8t_openstack-operators(fe84dde4-76f7-4df6-a6fd-d1b45d097e9b): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.597029 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r4m8t" podUID="fe84dde4-76f7-4df6-a6fd-d1b45d097e9b" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.610568 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.610821 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f64br,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-ztwsn_openstack-operators(905ae077-9cc3-41a9-b03f-77b3eeb052b3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.612698 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-ztwsn" podUID="905ae077-9cc3-41a9-b03f-77b3eeb052b3" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.674577 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.674854 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wbgxs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-rq99q_openstack-operators(7a8b9284-3f13-4396-8f85-36f0c59e1ba3): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.677197 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rq99q" podUID="7a8b9284-3f13-4396-8f85-36f0c59e1ba3" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.688797 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.689049 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k2l9p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-q62zq_openstack-operators(c9d46c14-6c5e-42d4-81f3-43e7169d23a0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.690300 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q62zq" podUID="c9d46c14-6c5e-42d4-81f3-43e7169d23a0" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.733847 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.734013 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-btr69,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-5zk9v_openstack-operators(c488dae2-6bbc-434a-a1c8-d3bd6ba5a549): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.736052 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-5zk9v" podUID="c488dae2-6bbc-434a-a1c8-d3bd6ba5a549" Dec 05 07:16:12 crc kubenswrapper[4997]: I1205 07:16:12.747636 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hrs7p" event={"ID":"3f6028fa-7b09-49ec-a8ed-4f4b25419e19","Type":"ContainerStarted","Data":"8fb2fec32551e30bc95f611e6a4c0c3744c67a7f7426a1cdcbec42cc7e7c589f"} Dec 05 07:16:12 crc kubenswrapper[4997]: I1205 07:16:12.749986 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qsqst" event={"ID":"8d31795c-bf5a-48fc-85cc-0c1385c2d7b0","Type":"ContainerStarted","Data":"ac1b56d5569e1d98eb20f42bb5f831961cd63d0bf558b733803d09f662d18336"} Dec 05 07:16:12 crc kubenswrapper[4997]: I1205 07:16:12.750962 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-5zk9v" Dec 05 07:16:12 crc kubenswrapper[4997]: I1205 07:16:12.754142 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-5zk9v" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.762106 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.762369 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5gwcg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-lpx4g_openstack-operators(d5054dc6-123a-4d45-8340-5315c0c57e32): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.763859 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-lpx4g" podUID="d5054dc6-123a-4d45-8340-5315c0c57e32" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.873796 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.873824 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.873938 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9qkm7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-r9knp_openstack-operators(9cc09cce-bafa-4f17-bd45-b767fa576e54): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.873946 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wjtmj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-jk2cb_openstack-operators(7db21a3e-655e-4874-8366-dfdb5d2521f5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.876038 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-jk2cb" podUID="7db21a3e-655e-4874-8366-dfdb5d2521f5" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.876087 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-r9knp" podUID="9cc09cce-bafa-4f17-bd45-b767fa576e54" Dec 05 07:16:12 crc kubenswrapper[4997]: I1205 07:16:12.901196 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qsqst" podStartSLOduration=20.1824394 podStartE2EDuration="47.901175019s" podCreationTimestamp="2025-12-05 07:15:25 +0000 UTC" firstStartedPulling="2025-12-05 07:15:28.442904149 +0000 UTC m=+1228.971811400" lastFinishedPulling="2025-12-05 07:15:56.161639758 +0000 UTC m=+1256.690547019" observedRunningTime="2025-12-05 07:16:12.895282661 +0000 UTC m=+1273.424189922" watchObservedRunningTime="2025-12-05 07:16:12.901175019 +0000 UTC m=+1273.430082280" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.935976 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:723607448b0abc536cd883abffcf6942c1c562a48117db73f6fe693d99395ee2: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:723607448b0abc536cd883abffcf6942c1c562a48117db73f6fe693d99395ee2\": context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.936246 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j6k77,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-r9mh5_openstack-operators(351c9912-1ac6-4b8b-b009-d073b1bd28a5): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:723607448b0abc536cd883abffcf6942c1c562a48117db73f6fe693d99395ee2: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:723607448b0abc536cd883abffcf6942c1c562a48117db73f6fe693d99395ee2\": context canceled" logger="UnhandledError" Dec 05 07:16:12 crc kubenswrapper[4997]: E1205 07:16:12.937497 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:723607448b0abc536cd883abffcf6942c1c562a48117db73f6fe693d99395ee2: Get \\\"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:723607448b0abc536cd883abffcf6942c1c562a48117db73f6fe693d99395ee2\\\": context canceled\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-r9mh5" podUID="351c9912-1ac6-4b8b-b009-d073b1bd28a5" Dec 05 07:16:13 crc kubenswrapper[4997]: I1205 07:16:13.844467 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-kbfbg" event={"ID":"25ca1d60-a629-447e-afa7-7a9f0e6f0b7e","Type":"ContainerStarted","Data":"40072d310442c0ab031da6487c88318259538b0ff25bce608cd1969b8f0e7a08"} Dec 05 07:16:13 crc kubenswrapper[4997]: I1205 07:16:13.845823 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-kbfbg" Dec 05 07:16:13 crc kubenswrapper[4997]: I1205 07:16:13.849736 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-kbfbg" Dec 05 07:16:13 crc kubenswrapper[4997]: I1205 07:16:13.862935 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tpg29" event={"ID":"e5f7e30e-e2a8-4131-9c10-1dddfe528684","Type":"ContainerStarted","Data":"4a5f5e150485e7a33f7698f73d9ef75809792acf35cbabd4c35a87ff0ddfdbbb"} Dec 05 07:16:13 crc kubenswrapper[4997]: I1205 07:16:13.863447 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-r9knp" Dec 05 07:16:13 crc kubenswrapper[4997]: I1205 07:16:13.864787 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-r9mh5" Dec 05 07:16:13 crc kubenswrapper[4997]: I1205 07:16:13.868591 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-r9knp" Dec 05 07:16:13 crc kubenswrapper[4997]: I1205 07:16:13.870193 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-r9mh5" Dec 05 07:16:13 crc kubenswrapper[4997]: I1205 07:16:13.873032 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-kbfbg" podStartSLOduration=4.593926589 podStartE2EDuration="48.873015984s" podCreationTimestamp="2025-12-05 07:15:25 +0000 UTC" firstStartedPulling="2025-12-05 07:15:28.460634946 +0000 UTC m=+1228.989542207" lastFinishedPulling="2025-12-05 07:16:12.739724321 +0000 UTC m=+1273.268631602" observedRunningTime="2025-12-05 07:16:13.870716622 +0000 UTC m=+1274.399623903" watchObservedRunningTime="2025-12-05 07:16:13.873015984 +0000 UTC m=+1274.401923245" Dec 05 07:16:13 crc kubenswrapper[4997]: I1205 07:16:13.934712 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tpg29" podStartSLOduration=5.567309849 podStartE2EDuration="49.934687185s" podCreationTimestamp="2025-12-05 07:15:24 +0000 UTC" firstStartedPulling="2025-12-05 07:15:28.338579396 +0000 UTC m=+1228.867486667" lastFinishedPulling="2025-12-05 07:16:12.705956702 +0000 UTC m=+1273.234864003" observedRunningTime="2025-12-05 07:16:13.933175435 +0000 UTC m=+1274.462082716" watchObservedRunningTime="2025-12-05 07:16:13.934687185 +0000 UTC m=+1274.463594446" Dec 05 07:16:14 crc kubenswrapper[4997]: E1205 07:16:14.510491 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-qppfb" podUID="7afb5248-e41e-46ec-9e46-ed9fcc7cf8d3" Dec 05 07:16:15 crc kubenswrapper[4997]: E1205 07:16:14.659601 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" podUID="6aac09d2-991d-45b9-bba9-4cf2d70278ab" Dec 05 07:16:15 crc kubenswrapper[4997]: E1205 07:16:14.660952 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" podUID="64d88cd2-49b5-439a-9344-bb9dc4ec5730" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.878016 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hrs7p" event={"ID":"3f6028fa-7b09-49ec-a8ed-4f4b25419e19","Type":"ContainerStarted","Data":"a4fc3210ab3523b2eb2fa785cdbc6a92008d18d33f6ffa5fd71e2c1780f60e09"} Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.879460 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hrs7p" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.881512 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88hm7" event={"ID":"40df044a-6922-43cf-ae37-ebc8dd637119","Type":"ContainerStarted","Data":"a53e338247c5fa8e919b94d5e036b68094d8db6d99149fb9637d4fb1633fbe5b"} Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.883026 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" event={"ID":"6aac09d2-991d-45b9-bba9-4cf2d70278ab","Type":"ContainerStarted","Data":"3197aed0c8a874ae1750311516d97e958d0eca251513460060076ddb3485a088"} Dec 05 07:16:15 crc kubenswrapper[4997]: E1205 07:16:14.884246 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" podUID="6aac09d2-991d-45b9-bba9-4cf2d70278ab" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.885352 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-rnq5d" event={"ID":"27e5617b-9979-45b2-a526-bb671edf63c2","Type":"ContainerStarted","Data":"b82f78977282c460c550feceac949ec280ce6a9f6310bfbae78c30d8da5075fe"} Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.886951 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r4m8t" event={"ID":"fe84dde4-76f7-4df6-a6fd-d1b45d097e9b","Type":"ContainerStarted","Data":"cd02b61594180653823a7ce4882a180793792f389b6b2dc009d758d8658db5f9"} Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.887770 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r4m8t" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.890775 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-5zk9v" event={"ID":"c488dae2-6bbc-434a-a1c8-d3bd6ba5a549","Type":"ContainerStarted","Data":"e8cab29ee9a695302602c99d48b2ffdb9c3b756478ae5adb6ea8d6333e79228c"} Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.891514 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r4m8t" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.892782 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" event={"ID":"64d88cd2-49b5-439a-9344-bb9dc4ec5730","Type":"ContainerStarted","Data":"317cdae7ded2ddf000263652efaafdea7f58f1a02fe3904118cbd20ea199f0ec"} Dec 05 07:16:15 crc kubenswrapper[4997]: E1205 07:16:14.893816 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" podUID="64d88cd2-49b5-439a-9344-bb9dc4ec5730" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.895422 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-jtjgd" event={"ID":"9597d662-50ea-4066-8bbc-232cc1d90dbd","Type":"ContainerStarted","Data":"7a0b5e803992ec89e37d0316c69e48cf1bdcf402fb4b4df82d5d4e63cb1bcb82"} Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.896095 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-jtjgd" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.897486 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skcxd" event={"ID":"f309b30b-9712-4175-b237-7c35ac2e5dee","Type":"ContainerStarted","Data":"e76b9a57d057260dcf45981cdaeb88d5e5ef1bfe346fc99d3b10e077fc158998"} Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.898069 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skcxd" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.899642 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-sbfxr" event={"ID":"3580ba04-55a0-4e33-aa08-932a2648bf2d","Type":"ContainerStarted","Data":"7c415255e1f4c5ee4bff249bbe81dafd08cabcc5eef90106aa095ea6407cda7c"} Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.900225 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-sbfxr" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.901925 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-sbfxr" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.902538 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rq99q" event={"ID":"7a8b9284-3f13-4396-8f85-36f0c59e1ba3","Type":"ContainerStarted","Data":"46560a155cd41cfc8b941920d2ecde362b6bf2cb405a632c494a9f7dc1ea1f85"} Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.903133 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rq99q" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.905034 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rq99q" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.905198 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-ztwsn" event={"ID":"905ae077-9cc3-41a9-b03f-77b3eeb052b3","Type":"ContainerStarted","Data":"999f16a4680125aa9990ca08dff72596bd0b1e28cdd340192f87921ff54fa6fc"} Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.906658 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-qppfb" event={"ID":"7afb5248-e41e-46ec-9e46-ed9fcc7cf8d3","Type":"ContainerStarted","Data":"5165ccc0de1fd08b54f2833f8e309bc81b8e4e6bbde14d451eb04f0f02e0e60c"} Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.910262 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-cjggt" event={"ID":"e1874b23-363f-4303-bfeb-86282b20963d","Type":"ContainerStarted","Data":"826510aca2ec88ce49609067c67550f41835d2dbc935c14d3f3535ef220e90c7"} Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.910288 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tpg29" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.910300 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-cjggt" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.912061 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-cjggt" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.916260 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tpg29" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.921967 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-jtjgd" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:14.928429 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skcxd" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:15.002426 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-r4m8t" podStartSLOduration=28.41575564 podStartE2EDuration="51.002409313s" podCreationTimestamp="2025-12-05 07:15:24 +0000 UTC" firstStartedPulling="2025-12-05 07:15:27.853071285 +0000 UTC m=+1228.381978546" lastFinishedPulling="2025-12-05 07:15:50.439724918 +0000 UTC m=+1250.968632219" observedRunningTime="2025-12-05 07:16:15.001336404 +0000 UTC m=+1275.530243675" watchObservedRunningTime="2025-12-05 07:16:15.002409313 +0000 UTC m=+1275.531316574" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:15.007138 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hrs7p" podStartSLOduration=23.427061195 podStartE2EDuration="51.007122449s" podCreationTimestamp="2025-12-05 07:15:24 +0000 UTC" firstStartedPulling="2025-12-05 07:15:28.443391173 +0000 UTC m=+1228.972298434" lastFinishedPulling="2025-12-05 07:15:56.023452427 +0000 UTC m=+1256.552359688" observedRunningTime="2025-12-05 07:16:14.987754858 +0000 UTC m=+1275.516662119" watchObservedRunningTime="2025-12-05 07:16:15.007122449 +0000 UTC m=+1275.536029710" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:15.109003 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-sbfxr" podStartSLOduration=6.176828345 podStartE2EDuration="51.108985433s" podCreationTimestamp="2025-12-05 07:15:24 +0000 UTC" firstStartedPulling="2025-12-05 07:15:28.349435238 +0000 UTC m=+1228.878342489" lastFinishedPulling="2025-12-05 07:16:13.281592316 +0000 UTC m=+1273.810499577" observedRunningTime="2025-12-05 07:16:15.104590525 +0000 UTC m=+1275.633497786" watchObservedRunningTime="2025-12-05 07:16:15.108985433 +0000 UTC m=+1275.637892694" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:15.354604 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-jtjgd" podStartSLOduration=6.099356905 podStartE2EDuration="50.354578647s" podCreationTimestamp="2025-12-05 07:15:25 +0000 UTC" firstStartedPulling="2025-12-05 07:15:28.451205832 +0000 UTC m=+1228.980113083" lastFinishedPulling="2025-12-05 07:16:12.706427564 +0000 UTC m=+1273.235334825" observedRunningTime="2025-12-05 07:16:15.328754432 +0000 UTC m=+1275.857661703" watchObservedRunningTime="2025-12-05 07:16:15.354578647 +0000 UTC m=+1275.883485908" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:15.472652 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-cjggt" podStartSLOduration=6.03672218 podStartE2EDuration="50.472607467s" podCreationTimestamp="2025-12-05 07:15:25 +0000 UTC" firstStartedPulling="2025-12-05 07:15:28.438148502 +0000 UTC m=+1228.967055763" lastFinishedPulling="2025-12-05 07:16:12.874033789 +0000 UTC m=+1273.402941050" observedRunningTime="2025-12-05 07:16:15.432856946 +0000 UTC m=+1275.961764237" watchObservedRunningTime="2025-12-05 07:16:15.472607467 +0000 UTC m=+1276.001514728" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:15.501557 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-5zk9v" podStartSLOduration=32.590691769 podStartE2EDuration="51.501537896s" podCreationTimestamp="2025-12-05 07:15:24 +0000 UTC" firstStartedPulling="2025-12-05 07:15:28.308568408 +0000 UTC m=+1228.837475669" lastFinishedPulling="2025-12-05 07:15:47.219414535 +0000 UTC m=+1247.748321796" observedRunningTime="2025-12-05 07:16:15.501171046 +0000 UTC m=+1276.030078307" watchObservedRunningTime="2025-12-05 07:16:15.501537896 +0000 UTC m=+1276.030445157" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:15.552201 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-skcxd" podStartSLOduration=6.079159742 podStartE2EDuration="50.552172589s" podCreationTimestamp="2025-12-05 07:15:25 +0000 UTC" firstStartedPulling="2025-12-05 07:15:28.451028458 +0000 UTC m=+1228.979935719" lastFinishedPulling="2025-12-05 07:16:12.924041305 +0000 UTC m=+1273.452948566" observedRunningTime="2025-12-05 07:16:15.541289667 +0000 UTC m=+1276.070196938" watchObservedRunningTime="2025-12-05 07:16:15.552172589 +0000 UTC m=+1276.081079850" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:15.603766 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rq99q" podStartSLOduration=31.714595619 podStartE2EDuration="51.603743078s" podCreationTimestamp="2025-12-05 07:15:24 +0000 UTC" firstStartedPulling="2025-12-05 07:15:28.39938821 +0000 UTC m=+1228.928295471" lastFinishedPulling="2025-12-05 07:15:48.288535669 +0000 UTC m=+1248.817442930" observedRunningTime="2025-12-05 07:16:15.59713508 +0000 UTC m=+1276.126042351" watchObservedRunningTime="2025-12-05 07:16:15.603743078 +0000 UTC m=+1276.132650339" Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:15.918784 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q62zq" event={"ID":"c9d46c14-6c5e-42d4-81f3-43e7169d23a0","Type":"ContainerStarted","Data":"24aab5622794f5a0806a3401860080a831dd7d958040bbc9d9c61f2ef7ec28db"} Dec 05 07:16:15 crc kubenswrapper[4997]: I1205 07:16:15.920648 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-rnq5d" event={"ID":"27e5617b-9979-45b2-a526-bb671edf63c2","Type":"ContainerStarted","Data":"4e410e9698e24f53a808934a1d06c50a4026cbf5e8fde43fa1620f13c966ca9e"} Dec 05 07:16:15 crc kubenswrapper[4997]: E1205 07:16:15.923381 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" podUID="64d88cd2-49b5-439a-9344-bb9dc4ec5730" Dec 05 07:16:15 crc kubenswrapper[4997]: E1205 07:16:15.924525 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" podUID="6aac09d2-991d-45b9-bba9-4cf2d70278ab" Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.223065 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-rnq5d" podStartSLOduration=7.317910587 podStartE2EDuration="52.223040259s" podCreationTimestamp="2025-12-05 07:15:24 +0000 UTC" firstStartedPulling="2025-12-05 07:15:27.81719146 +0000 UTC m=+1228.346098721" lastFinishedPulling="2025-12-05 07:16:12.722321132 +0000 UTC m=+1273.251228393" observedRunningTime="2025-12-05 07:16:16.219654347 +0000 UTC m=+1276.748561628" watchObservedRunningTime="2025-12-05 07:16:16.223040259 +0000 UTC m=+1276.751947520" Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.932173 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-lpx4g" event={"ID":"d5054dc6-123a-4d45-8340-5315c0c57e32","Type":"ContainerStarted","Data":"13662a2c514fbd833e676b1868fab11285a563635b80e653fc224fb6130776c6"} Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.932555 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-lpx4g" event={"ID":"d5054dc6-123a-4d45-8340-5315c0c57e32","Type":"ContainerStarted","Data":"2415883ae66653b5657a64f2b070263b80aa8deefb11772197d3616a22a816af"} Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.933693 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-lpx4g" Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.935942 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-ztwsn" event={"ID":"905ae077-9cc3-41a9-b03f-77b3eeb052b3","Type":"ContainerStarted","Data":"f1278c09d6a352fc92b18c88c2d524e69d172e58c37aad464cf2fea3d814a71d"} Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.936237 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-ztwsn" Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.938709 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-r9mh5" event={"ID":"351c9912-1ac6-4b8b-b009-d073b1bd28a5","Type":"ContainerStarted","Data":"bec23cc4b870e77f035bc4f0506d17140a28924fba1ffd1eef046a39f21eb757"} Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.941006 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q62zq" event={"ID":"c9d46c14-6c5e-42d4-81f3-43e7169d23a0","Type":"ContainerStarted","Data":"b3bd615a79e84cb90768a1f089d789f9f3884ba106fed98a6fef31328d11d329"} Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.941809 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q62zq" Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.944055 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88hm7" event={"ID":"40df044a-6922-43cf-ae37-ebc8dd637119","Type":"ContainerStarted","Data":"fd4fa0380d2cba0cdcaf09bcda9aa6b43a56e53e59a5574e5b57834c5e43b21e"} Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.944228 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88hm7" Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.946039 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-qppfb" event={"ID":"7afb5248-e41e-46ec-9e46-ed9fcc7cf8d3","Type":"ContainerStarted","Data":"0f11c047444ce4ad23bb9f31ccfdc528eeb6326cb6bb376e77422226aca88e6e"} Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.946734 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-qppfb" Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.948993 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-r9knp" event={"ID":"9cc09cce-bafa-4f17-bd45-b767fa576e54","Type":"ContainerStarted","Data":"48ee5727bd0553ffc0df99708486dd2050cca8778a410f3762bb74421f9f1b27"} Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.952364 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-jk2cb" event={"ID":"7db21a3e-655e-4874-8366-dfdb5d2521f5","Type":"ContainerStarted","Data":"a24e639122d3c5b791e302bef2ed3c7cc95146703105e4db72285821e4469ecd"} Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.952393 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-jk2cb" event={"ID":"7db21a3e-655e-4874-8366-dfdb5d2521f5","Type":"ContainerStarted","Data":"ee20b83873c416d51caf7869b4abef1d92fac3f8794b2df86e6f02c519a3f8e7"} Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.952762 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-jk2cb" Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.954169 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-rnq5d" Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.962961 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-lpx4g" podStartSLOduration=6.313322395 podStartE2EDuration="52.962942256s" podCreationTimestamp="2025-12-05 07:15:24 +0000 UTC" firstStartedPulling="2025-12-05 07:15:27.891130998 +0000 UTC m=+1228.420038249" lastFinishedPulling="2025-12-05 07:16:14.540750849 +0000 UTC m=+1275.069658110" observedRunningTime="2025-12-05 07:16:16.958498757 +0000 UTC m=+1277.487406038" watchObservedRunningTime="2025-12-05 07:16:16.962942256 +0000 UTC m=+1277.491849507" Dec 05 07:16:16 crc kubenswrapper[4997]: I1205 07:16:16.990632 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-jk2cb" podStartSLOduration=6.791572813 podStartE2EDuration="52.990603512s" podCreationTimestamp="2025-12-05 07:15:24 +0000 UTC" firstStartedPulling="2025-12-05 07:15:28.339110559 +0000 UTC m=+1228.868017820" lastFinishedPulling="2025-12-05 07:16:14.538141258 +0000 UTC m=+1275.067048519" observedRunningTime="2025-12-05 07:16:16.985225767 +0000 UTC m=+1277.514133038" watchObservedRunningTime="2025-12-05 07:16:16.990603512 +0000 UTC m=+1277.519510773" Dec 05 07:16:17 crc kubenswrapper[4997]: I1205 07:16:17.007514 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88hm7" podStartSLOduration=8.077470541 podStartE2EDuration="53.007488396s" podCreationTimestamp="2025-12-05 07:15:24 +0000 UTC" firstStartedPulling="2025-12-05 07:15:28.426947281 +0000 UTC m=+1228.955854542" lastFinishedPulling="2025-12-05 07:16:13.356965116 +0000 UTC m=+1273.885872397" observedRunningTime="2025-12-05 07:16:17.000918939 +0000 UTC m=+1277.529826200" watchObservedRunningTime="2025-12-05 07:16:17.007488396 +0000 UTC m=+1277.536395657" Dec 05 07:16:17 crc kubenswrapper[4997]: I1205 07:16:17.034671 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-r9mh5" podStartSLOduration=32.227006676 podStartE2EDuration="53.034646527s" podCreationTimestamp="2025-12-05 07:15:24 +0000 UTC" firstStartedPulling="2025-12-05 07:15:26.411729912 +0000 UTC m=+1226.940637173" lastFinishedPulling="2025-12-05 07:15:47.219369763 +0000 UTC m=+1247.748277024" observedRunningTime="2025-12-05 07:16:17.025380908 +0000 UTC m=+1277.554288169" watchObservedRunningTime="2025-12-05 07:16:17.034646527 +0000 UTC m=+1277.563553788" Dec 05 07:16:17 crc kubenswrapper[4997]: I1205 07:16:17.055196 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q62zq" podStartSLOduration=8.01905378 podStartE2EDuration="53.055181291s" podCreationTimestamp="2025-12-05 07:15:24 +0000 UTC" firstStartedPulling="2025-12-05 07:15:28.390172912 +0000 UTC m=+1228.919080173" lastFinishedPulling="2025-12-05 07:16:13.426300403 +0000 UTC m=+1273.955207684" observedRunningTime="2025-12-05 07:16:17.050313919 +0000 UTC m=+1277.579221180" watchObservedRunningTime="2025-12-05 07:16:17.055181291 +0000 UTC m=+1277.584088552" Dec 05 07:16:17 crc kubenswrapper[4997]: I1205 07:16:17.102028 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-qppfb" podStartSLOduration=5.874102418 podStartE2EDuration="53.102006451s" podCreationTimestamp="2025-12-05 07:15:24 +0000 UTC" firstStartedPulling="2025-12-05 07:15:28.3990203 +0000 UTC m=+1228.927927561" lastFinishedPulling="2025-12-05 07:16:15.626924343 +0000 UTC m=+1276.155831594" observedRunningTime="2025-12-05 07:16:17.10009058 +0000 UTC m=+1277.628997851" watchObservedRunningTime="2025-12-05 07:16:17.102006451 +0000 UTC m=+1277.630913712" Dec 05 07:16:17 crc kubenswrapper[4997]: I1205 07:16:17.164109 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-ztwsn" podStartSLOduration=8.235230961 podStartE2EDuration="53.164087624s" podCreationTimestamp="2025-12-05 07:15:24 +0000 UTC" firstStartedPulling="2025-12-05 07:15:28.423386425 +0000 UTC m=+1228.952293686" lastFinishedPulling="2025-12-05 07:16:13.352243088 +0000 UTC m=+1273.881150349" observedRunningTime="2025-12-05 07:16:17.162150271 +0000 UTC m=+1277.691057542" watchObservedRunningTime="2025-12-05 07:16:17.164087624 +0000 UTC m=+1277.692994895" Dec 05 07:16:24 crc kubenswrapper[4997]: I1205 07:16:24.971973 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-rnq5d" Dec 05 07:16:24 crc kubenswrapper[4997]: I1205 07:16:24.992812 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-r9knp" podStartSLOduration=42.175074564 podStartE2EDuration="1m0.992786777s" podCreationTimestamp="2025-12-05 07:15:24 +0000 UTC" firstStartedPulling="2025-12-05 07:15:28.40163483 +0000 UTC m=+1228.930542091" lastFinishedPulling="2025-12-05 07:15:47.219347043 +0000 UTC m=+1247.748254304" observedRunningTime="2025-12-05 07:16:17.195296134 +0000 UTC m=+1277.724203405" watchObservedRunningTime="2025-12-05 07:16:24.992786777 +0000 UTC m=+1285.521694038" Dec 05 07:16:25 crc kubenswrapper[4997]: I1205 07:16:25.094118 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-lpx4g" Dec 05 07:16:25 crc kubenswrapper[4997]: I1205 07:16:25.149738 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-jk2cb" Dec 05 07:16:25 crc kubenswrapper[4997]: I1205 07:16:25.318891 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q62zq" Dec 05 07:16:25 crc kubenswrapper[4997]: I1205 07:16:25.681492 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-qppfb" Dec 05 07:16:25 crc kubenswrapper[4997]: I1205 07:16:25.692383 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88hm7" Dec 05 07:16:25 crc kubenswrapper[4997]: I1205 07:16:25.869354 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-hrs7p" Dec 05 07:16:26 crc kubenswrapper[4997]: I1205 07:16:26.003504 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-ztwsn" Dec 05 07:16:29 crc kubenswrapper[4997]: I1205 07:16:29.057886 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" event={"ID":"6aac09d2-991d-45b9-bba9-4cf2d70278ab","Type":"ContainerStarted","Data":"60d39b112da6f691017e065868f2ec88eaaa94657156e8bce76379bcd3702ad0"} Dec 05 07:16:29 crc kubenswrapper[4997]: I1205 07:16:29.058752 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" Dec 05 07:16:29 crc kubenswrapper[4997]: I1205 07:16:29.124421 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" podStartSLOduration=32.225153109 podStartE2EDuration="1m5.124369064s" podCreationTimestamp="2025-12-05 07:15:24 +0000 UTC" firstStartedPulling="2025-12-05 07:15:55.27432323 +0000 UTC m=+1255.803230491" lastFinishedPulling="2025-12-05 07:16:28.173539185 +0000 UTC m=+1288.702446446" observedRunningTime="2025-12-05 07:16:29.083961956 +0000 UTC m=+1289.612869217" watchObservedRunningTime="2025-12-05 07:16:29.124369064 +0000 UTC m=+1289.653276335" Dec 05 07:16:32 crc kubenswrapper[4997]: I1205 07:16:32.079367 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" event={"ID":"64d88cd2-49b5-439a-9344-bb9dc4ec5730","Type":"ContainerStarted","Data":"f119698e88b25daf1fc020aea789c816e9b6d35c9711d0715556dcf2595649c1"} Dec 05 07:16:32 crc kubenswrapper[4997]: I1205 07:16:32.080401 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" Dec 05 07:16:32 crc kubenswrapper[4997]: I1205 07:16:32.110508 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" podStartSLOduration=32.501250345 podStartE2EDuration="1m8.110489631s" podCreationTimestamp="2025-12-05 07:15:24 +0000 UTC" firstStartedPulling="2025-12-05 07:15:55.988085433 +0000 UTC m=+1256.516992694" lastFinishedPulling="2025-12-05 07:16:31.597324719 +0000 UTC m=+1292.126231980" observedRunningTime="2025-12-05 07:16:32.110043998 +0000 UTC m=+1292.638951269" watchObservedRunningTime="2025-12-05 07:16:32.110489631 +0000 UTC m=+1292.639396892" Dec 05 07:16:40 crc kubenswrapper[4997]: I1205 07:16:40.795773 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pppgg" Dec 05 07:16:41 crc kubenswrapper[4997]: I1205 07:16:41.395207 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f56jjnr" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.602636 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-wvjqk"] Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.604774 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-wvjqk" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.610017 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.610069 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.610244 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-4fcmg" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.610376 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.657230 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-wvjqk"] Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.712341 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-567c455747-7vz4d"] Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.714161 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-7vz4d" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.720019 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.727080 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-567c455747-7vz4d"] Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.729430 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3da07353-d51d-4887-9b4b-b8e7d60520ac-config\") pod \"dnsmasq-dns-5cd484bb89-wvjqk\" (UID: \"3da07353-d51d-4887-9b4b-b8e7d60520ac\") " pod="openstack/dnsmasq-dns-5cd484bb89-wvjqk" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.729485 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fhg9\" (UniqueName: \"kubernetes.io/projected/3da07353-d51d-4887-9b4b-b8e7d60520ac-kube-api-access-2fhg9\") pod \"dnsmasq-dns-5cd484bb89-wvjqk\" (UID: \"3da07353-d51d-4887-9b4b-b8e7d60520ac\") " pod="openstack/dnsmasq-dns-5cd484bb89-wvjqk" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.831468 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3da07353-d51d-4887-9b4b-b8e7d60520ac-config\") pod \"dnsmasq-dns-5cd484bb89-wvjqk\" (UID: \"3da07353-d51d-4887-9b4b-b8e7d60520ac\") " pod="openstack/dnsmasq-dns-5cd484bb89-wvjqk" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.831531 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fhg9\" (UniqueName: \"kubernetes.io/projected/3da07353-d51d-4887-9b4b-b8e7d60520ac-kube-api-access-2fhg9\") pod \"dnsmasq-dns-5cd484bb89-wvjqk\" (UID: \"3da07353-d51d-4887-9b4b-b8e7d60520ac\") " pod="openstack/dnsmasq-dns-5cd484bb89-wvjqk" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.831561 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-dns-svc\") pod \"dnsmasq-dns-567c455747-7vz4d\" (UID: \"abd5fd09-09f8-4c55-8f80-f550cd7bdc0b\") " pod="openstack/dnsmasq-dns-567c455747-7vz4d" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.831600 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnrzh\" (UniqueName: \"kubernetes.io/projected/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-kube-api-access-fnrzh\") pod \"dnsmasq-dns-567c455747-7vz4d\" (UID: \"abd5fd09-09f8-4c55-8f80-f550cd7bdc0b\") " pod="openstack/dnsmasq-dns-567c455747-7vz4d" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.831632 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-config\") pod \"dnsmasq-dns-567c455747-7vz4d\" (UID: \"abd5fd09-09f8-4c55-8f80-f550cd7bdc0b\") " pod="openstack/dnsmasq-dns-567c455747-7vz4d" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.832652 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3da07353-d51d-4887-9b4b-b8e7d60520ac-config\") pod \"dnsmasq-dns-5cd484bb89-wvjqk\" (UID: \"3da07353-d51d-4887-9b4b-b8e7d60520ac\") " pod="openstack/dnsmasq-dns-5cd484bb89-wvjqk" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.888781 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fhg9\" (UniqueName: \"kubernetes.io/projected/3da07353-d51d-4887-9b4b-b8e7d60520ac-kube-api-access-2fhg9\") pod \"dnsmasq-dns-5cd484bb89-wvjqk\" (UID: \"3da07353-d51d-4887-9b4b-b8e7d60520ac\") " pod="openstack/dnsmasq-dns-5cd484bb89-wvjqk" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.933210 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-dns-svc\") pod \"dnsmasq-dns-567c455747-7vz4d\" (UID: \"abd5fd09-09f8-4c55-8f80-f550cd7bdc0b\") " pod="openstack/dnsmasq-dns-567c455747-7vz4d" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.933283 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnrzh\" (UniqueName: \"kubernetes.io/projected/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-kube-api-access-fnrzh\") pod \"dnsmasq-dns-567c455747-7vz4d\" (UID: \"abd5fd09-09f8-4c55-8f80-f550cd7bdc0b\") " pod="openstack/dnsmasq-dns-567c455747-7vz4d" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.933306 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-config\") pod \"dnsmasq-dns-567c455747-7vz4d\" (UID: \"abd5fd09-09f8-4c55-8f80-f550cd7bdc0b\") " pod="openstack/dnsmasq-dns-567c455747-7vz4d" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.934320 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-config\") pod \"dnsmasq-dns-567c455747-7vz4d\" (UID: \"abd5fd09-09f8-4c55-8f80-f550cd7bdc0b\") " pod="openstack/dnsmasq-dns-567c455747-7vz4d" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.934398 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-dns-svc\") pod \"dnsmasq-dns-567c455747-7vz4d\" (UID: \"abd5fd09-09f8-4c55-8f80-f550cd7bdc0b\") " pod="openstack/dnsmasq-dns-567c455747-7vz4d" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.950643 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnrzh\" (UniqueName: \"kubernetes.io/projected/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-kube-api-access-fnrzh\") pod \"dnsmasq-dns-567c455747-7vz4d\" (UID: \"abd5fd09-09f8-4c55-8f80-f550cd7bdc0b\") " pod="openstack/dnsmasq-dns-567c455747-7vz4d" Dec 05 07:16:57 crc kubenswrapper[4997]: I1205 07:16:57.955001 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-wvjqk" Dec 05 07:16:58 crc kubenswrapper[4997]: I1205 07:16:58.036757 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-7vz4d" Dec 05 07:16:58 crc kubenswrapper[4997]: I1205 07:16:58.361823 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-wvjqk"] Dec 05 07:16:58 crc kubenswrapper[4997]: I1205 07:16:58.562562 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-567c455747-7vz4d"] Dec 05 07:16:58 crc kubenswrapper[4997]: I1205 07:16:58.656431 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-567c455747-7vz4d"] Dec 05 07:16:58 crc kubenswrapper[4997]: I1205 07:16:58.684730 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-859d485f47-cdr9f"] Dec 05 07:16:58 crc kubenswrapper[4997]: I1205 07:16:58.686101 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-859d485f47-cdr9f" Dec 05 07:16:58 crc kubenswrapper[4997]: I1205 07:16:58.693827 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-859d485f47-cdr9f"] Dec 05 07:16:58 crc kubenswrapper[4997]: I1205 07:16:58.752723 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-config\") pod \"dnsmasq-dns-859d485f47-cdr9f\" (UID: \"a8cdbec8-a2eb-48cb-bfe9-b508c4e78416\") " pod="openstack/dnsmasq-dns-859d485f47-cdr9f" Dec 05 07:16:58 crc kubenswrapper[4997]: I1205 07:16:58.752833 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-dns-svc\") pod \"dnsmasq-dns-859d485f47-cdr9f\" (UID: \"a8cdbec8-a2eb-48cb-bfe9-b508c4e78416\") " pod="openstack/dnsmasq-dns-859d485f47-cdr9f" Dec 05 07:16:58 crc kubenswrapper[4997]: I1205 07:16:58.752870 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zfwr\" (UniqueName: \"kubernetes.io/projected/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-kube-api-access-5zfwr\") pod \"dnsmasq-dns-859d485f47-cdr9f\" (UID: \"a8cdbec8-a2eb-48cb-bfe9-b508c4e78416\") " pod="openstack/dnsmasq-dns-859d485f47-cdr9f" Dec 05 07:16:58 crc kubenswrapper[4997]: I1205 07:16:58.854524 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-config\") pod \"dnsmasq-dns-859d485f47-cdr9f\" (UID: \"a8cdbec8-a2eb-48cb-bfe9-b508c4e78416\") " pod="openstack/dnsmasq-dns-859d485f47-cdr9f" Dec 05 07:16:58 crc kubenswrapper[4997]: I1205 07:16:58.854603 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-dns-svc\") pod \"dnsmasq-dns-859d485f47-cdr9f\" (UID: \"a8cdbec8-a2eb-48cb-bfe9-b508c4e78416\") " pod="openstack/dnsmasq-dns-859d485f47-cdr9f" Dec 05 07:16:58 crc kubenswrapper[4997]: I1205 07:16:58.854638 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zfwr\" (UniqueName: \"kubernetes.io/projected/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-kube-api-access-5zfwr\") pod \"dnsmasq-dns-859d485f47-cdr9f\" (UID: \"a8cdbec8-a2eb-48cb-bfe9-b508c4e78416\") " pod="openstack/dnsmasq-dns-859d485f47-cdr9f" Dec 05 07:16:58 crc kubenswrapper[4997]: I1205 07:16:58.855656 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-config\") pod \"dnsmasq-dns-859d485f47-cdr9f\" (UID: \"a8cdbec8-a2eb-48cb-bfe9-b508c4e78416\") " pod="openstack/dnsmasq-dns-859d485f47-cdr9f" Dec 05 07:16:58 crc kubenswrapper[4997]: I1205 07:16:58.855656 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-dns-svc\") pod \"dnsmasq-dns-859d485f47-cdr9f\" (UID: \"a8cdbec8-a2eb-48cb-bfe9-b508c4e78416\") " pod="openstack/dnsmasq-dns-859d485f47-cdr9f" Dec 05 07:16:58 crc kubenswrapper[4997]: I1205 07:16:58.876395 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zfwr\" (UniqueName: \"kubernetes.io/projected/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-kube-api-access-5zfwr\") pod \"dnsmasq-dns-859d485f47-cdr9f\" (UID: \"a8cdbec8-a2eb-48cb-bfe9-b508c4e78416\") " pod="openstack/dnsmasq-dns-859d485f47-cdr9f" Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.008862 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-859d485f47-cdr9f" Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.320929 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd484bb89-wvjqk" event={"ID":"3da07353-d51d-4887-9b4b-b8e7d60520ac","Type":"ContainerStarted","Data":"4048e37b1997cdcaf61d32c6c267f02ee9a2b6654565bf2b8391804db6a532a6"} Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.322253 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567c455747-7vz4d" event={"ID":"abd5fd09-09f8-4c55-8f80-f550cd7bdc0b","Type":"ContainerStarted","Data":"985db3af34cf98eb4787e3aa221955afb41d045625482bfb8a71d4f220403c53"} Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.542670 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-wvjqk"] Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.554358 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-859d485f47-cdr9f"] Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.571036 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb666b895-snwmr"] Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.572552 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-snwmr" Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.592305 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-snwmr"] Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.671300 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33be1eda-85ae-44a3-85ee-e99b66f32fae-dns-svc\") pod \"dnsmasq-dns-cb666b895-snwmr\" (UID: \"33be1eda-85ae-44a3-85ee-e99b66f32fae\") " pod="openstack/dnsmasq-dns-cb666b895-snwmr" Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.671652 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33be1eda-85ae-44a3-85ee-e99b66f32fae-config\") pod \"dnsmasq-dns-cb666b895-snwmr\" (UID: \"33be1eda-85ae-44a3-85ee-e99b66f32fae\") " pod="openstack/dnsmasq-dns-cb666b895-snwmr" Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.671708 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hmdt\" (UniqueName: \"kubernetes.io/projected/33be1eda-85ae-44a3-85ee-e99b66f32fae-kube-api-access-9hmdt\") pod \"dnsmasq-dns-cb666b895-snwmr\" (UID: \"33be1eda-85ae-44a3-85ee-e99b66f32fae\") " pod="openstack/dnsmasq-dns-cb666b895-snwmr" Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.773099 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33be1eda-85ae-44a3-85ee-e99b66f32fae-dns-svc\") pod \"dnsmasq-dns-cb666b895-snwmr\" (UID: \"33be1eda-85ae-44a3-85ee-e99b66f32fae\") " pod="openstack/dnsmasq-dns-cb666b895-snwmr" Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.773166 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33be1eda-85ae-44a3-85ee-e99b66f32fae-config\") pod \"dnsmasq-dns-cb666b895-snwmr\" (UID: \"33be1eda-85ae-44a3-85ee-e99b66f32fae\") " pod="openstack/dnsmasq-dns-cb666b895-snwmr" Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.773242 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hmdt\" (UniqueName: \"kubernetes.io/projected/33be1eda-85ae-44a3-85ee-e99b66f32fae-kube-api-access-9hmdt\") pod \"dnsmasq-dns-cb666b895-snwmr\" (UID: \"33be1eda-85ae-44a3-85ee-e99b66f32fae\") " pod="openstack/dnsmasq-dns-cb666b895-snwmr" Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.774717 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33be1eda-85ae-44a3-85ee-e99b66f32fae-dns-svc\") pod \"dnsmasq-dns-cb666b895-snwmr\" (UID: \"33be1eda-85ae-44a3-85ee-e99b66f32fae\") " pod="openstack/dnsmasq-dns-cb666b895-snwmr" Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.775696 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33be1eda-85ae-44a3-85ee-e99b66f32fae-config\") pod \"dnsmasq-dns-cb666b895-snwmr\" (UID: \"33be1eda-85ae-44a3-85ee-e99b66f32fae\") " pod="openstack/dnsmasq-dns-cb666b895-snwmr" Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.800373 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hmdt\" (UniqueName: \"kubernetes.io/projected/33be1eda-85ae-44a3-85ee-e99b66f32fae-kube-api-access-9hmdt\") pod \"dnsmasq-dns-cb666b895-snwmr\" (UID: \"33be1eda-85ae-44a3-85ee-e99b66f32fae\") " pod="openstack/dnsmasq-dns-cb666b895-snwmr" Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.888174 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 07:16:59 crc kubenswrapper[4997]: I1205 07:16:59.890015 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:16:59.892259 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:16:59.892555 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-cn5f6" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:16:59.894476 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:16:59.894697 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:16:59.894842 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:16:59.894986 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:16:59.895101 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:16:59.896125 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:16:59.998050 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-snwmr" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.101348 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.101401 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.101437 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.101457 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f158874-152b-46ce-ac42-d202684853ca-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.101476 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.101761 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.101826 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.101990 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ps2f6\" (UniqueName: \"kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-kube-api-access-ps2f6\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.102044 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.102100 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f158874-152b-46ce-ac42-d202684853ca-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.102123 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.203345 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.203395 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.203435 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ps2f6\" (UniqueName: \"kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-kube-api-access-ps2f6\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.203481 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f158874-152b-46ce-ac42-d202684853ca-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.203500 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.203549 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.203571 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.203606 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.203637 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f158874-152b-46ce-ac42-d202684853ca-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.203653 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.205235 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.205261 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.205356 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.205717 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.206574 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.234697 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.238207 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.241281 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f158874-152b-46ce-ac42-d202684853ca-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.244188 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f158874-152b-46ce-ac42-d202684853ca-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.247518 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ps2f6\" (UniqueName: \"kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-kube-api-access-ps2f6\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.332230 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.334063 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.336139 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-859d485f47-cdr9f" event={"ID":"a8cdbec8-a2eb-48cb-bfe9-b508c4e78416","Type":"ContainerStarted","Data":"b47bc56d0f868d6b9f7e3dd07bdf7a40ee7d56e135ebd3908f07881e9414f89b"} Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.361319 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.806050 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.941024 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.943736 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.947298 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.948450 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-9wv9z" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.952515 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.953557 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.953739 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.953567 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.953907 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 05 07:17:00 crc kubenswrapper[4997]: I1205 07:17:00.962741 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.198035 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ngbr\" (UniqueName: \"kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-kube-api-access-5ngbr\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.198100 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.198131 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.198159 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.198191 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-config-data\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.198216 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.198278 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.198316 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.198335 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.198360 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.198384 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.350780 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ngbr\" (UniqueName: \"kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-kube-api-access-5ngbr\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.351427 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.351464 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.352328 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.352485 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-config-data\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.352514 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.352766 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.353013 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.356134 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.356199 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.356242 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.358275 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.384322 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.421436 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-config-data\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.426903 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.428147 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.428879 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.661842 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.762122 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.773330 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.773384 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.773766 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.774185 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ngbr\" (UniqueName: \"kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-kube-api-access-5ngbr\") pod \"rabbitmq-server-0\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " pod="openstack/rabbitmq-server-0" Dec 05 07:17:01 crc kubenswrapper[4997]: I1205 07:17:01.886695 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.312726 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-snwmr"] Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.446893 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.480051 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-snwmr" event={"ID":"33be1eda-85ae-44a3-85ee-e99b66f32fae","Type":"ContainerStarted","Data":"471c1559376966f1431d31f71ff39e3d87053a04b283a3b0f8260af1c6e3ce19"} Dec 05 07:17:02 crc kubenswrapper[4997]: W1205 07:17:02.505201 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f158874_152b_46ce_ac42_d202684853ca.slice/crio-9795e2ed191f04d77c06fc39017db89322459c1559b11a047ed1fee44b59593e WatchSource:0}: Error finding container 9795e2ed191f04d77c06fc39017db89322459c1559b11a047ed1fee44b59593e: Status 404 returned error can't find the container with id 9795e2ed191f04d77c06fc39017db89322459c1559b11a047ed1fee44b59593e Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.731331 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.740197 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.745537 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.755406 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-ct8h9" Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.755744 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.756072 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.759233 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.759959 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.800693 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.919722 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0601e553-5305-4f54-9a2d-43b43a1a5dd4-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.919796 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0601e553-5305-4f54-9a2d-43b43a1a5dd4-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.919830 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.919857 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0601e553-5305-4f54-9a2d-43b43a1a5dd4-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.919895 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-config-data-default\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.919944 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jwls\" (UniqueName: \"kubernetes.io/projected/0601e553-5305-4f54-9a2d-43b43a1a5dd4-kube-api-access-6jwls\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.919991 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:02 crc kubenswrapper[4997]: I1205 07:17:02.920012 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-kolla-config\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.022068 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.022136 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-kolla-config\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.022190 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0601e553-5305-4f54-9a2d-43b43a1a5dd4-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.022220 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0601e553-5305-4f54-9a2d-43b43a1a5dd4-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.022348 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.022416 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0601e553-5305-4f54-9a2d-43b43a1a5dd4-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.022451 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-config-data-default\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.022511 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jwls\" (UniqueName: \"kubernetes.io/projected/0601e553-5305-4f54-9a2d-43b43a1a5dd4-kube-api-access-6jwls\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.023535 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.023765 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0601e553-5305-4f54-9a2d-43b43a1a5dd4-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.024002 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-kolla-config\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.024890 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-config-data-default\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.025100 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.052651 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0601e553-5305-4f54-9a2d-43b43a1a5dd4-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.080671 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0601e553-5305-4f54-9a2d-43b43a1a5dd4-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.091053 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jwls\" (UniqueName: \"kubernetes.io/projected/0601e553-5305-4f54-9a2d-43b43a1a5dd4-kube-api-access-6jwls\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.116959 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.392975 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.503583 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79","Type":"ContainerStarted","Data":"27a0d3d6b8ab687d87ab58840a512f00fcdb53823f791cfa7ba9a367a6e552ad"} Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.517660 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0f158874-152b-46ce-ac42-d202684853ca","Type":"ContainerStarted","Data":"9795e2ed191f04d77c06fc39017db89322459c1559b11a047ed1fee44b59593e"} Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.813578 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.815362 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.818582 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.818935 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-mhqms" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.819149 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.845375 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.918045 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.925519 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.930848 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.931430 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.931580 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-v6btw" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.931786 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.950794 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " pod="openstack/memcached-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.950907 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-kolla-config\") pod \"memcached-0\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " pod="openstack/memcached-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.951024 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hrfb\" (UniqueName: \"kubernetes.io/projected/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-kube-api-access-9hrfb\") pod \"memcached-0\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " pod="openstack/memcached-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.951081 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-config-data\") pod \"memcached-0\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " pod="openstack/memcached-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.951135 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " pod="openstack/memcached-0" Dec 05 07:17:03 crc kubenswrapper[4997]: I1205 07:17:03.968602 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.054291 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.054365 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.054409 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffwz9\" (UniqueName: \"kubernetes.io/projected/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-kube-api-access-ffwz9\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.054438 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-config-data\") pod \"memcached-0\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " pod="openstack/memcached-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.054464 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.054497 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.054537 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " pod="openstack/memcached-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.054571 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.054591 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.054811 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " pod="openstack/memcached-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.054839 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.054875 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-kolla-config\") pod \"memcached-0\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " pod="openstack/memcached-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.054894 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hrfb\" (UniqueName: \"kubernetes.io/projected/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-kube-api-access-9hrfb\") pod \"memcached-0\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " pod="openstack/memcached-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.065258 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-config-data\") pod \"memcached-0\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " pod="openstack/memcached-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.067458 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-kolla-config\") pod \"memcached-0\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " pod="openstack/memcached-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.087990 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " pod="openstack/memcached-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.088050 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " pod="openstack/memcached-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.118312 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hrfb\" (UniqueName: \"kubernetes.io/projected/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-kube-api-access-9hrfb\") pod \"memcached-0\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " pod="openstack/memcached-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.156395 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.156461 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.156529 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.156556 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.156596 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.156650 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.156718 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.156826 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffwz9\" (UniqueName: \"kubernetes.io/projected/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-kube-api-access-ffwz9\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.156816 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.165595 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.167010 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.173329 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.179707 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.188111 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.190080 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffwz9\" (UniqueName: \"kubernetes.io/projected/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-kube-api-access-ffwz9\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.198087 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.203277 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.213508 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.248720 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.261268 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.569455 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0601e553-5305-4f54-9a2d-43b43a1a5dd4","Type":"ContainerStarted","Data":"bdc5c4f8e0f8135d4200ac4b9170d56f703176ebd353118f32f1d072b5b48501"} Dec 05 07:17:04 crc kubenswrapper[4997]: I1205 07:17:04.972218 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 05 07:17:05 crc kubenswrapper[4997]: I1205 07:17:05.247958 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 07:17:05 crc kubenswrapper[4997]: I1205 07:17:05.592424 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b","Type":"ContainerStarted","Data":"7a399ac2c7f99edcffa9919bc82a175c87e153671bf6886319821c1a6e648b0b"} Dec 05 07:17:05 crc kubenswrapper[4997]: I1205 07:17:05.594997 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe","Type":"ContainerStarted","Data":"7b458905cf9b2e7f5e8005d6d0b844702a1acbf3e7f6db00f7a5af4e951d7c76"} Dec 05 07:17:05 crc kubenswrapper[4997]: I1205 07:17:05.875716 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 07:17:05 crc kubenswrapper[4997]: I1205 07:17:05.882908 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 07:17:05 crc kubenswrapper[4997]: I1205 07:17:05.890500 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-99rbm" Dec 05 07:17:05 crc kubenswrapper[4997]: I1205 07:17:05.900121 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 07:17:06 crc kubenswrapper[4997]: I1205 07:17:06.025463 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv24p\" (UniqueName: \"kubernetes.io/projected/833f0bba-91d9-41ab-a412-9e8ae8c38ba2-kube-api-access-nv24p\") pod \"kube-state-metrics-0\" (UID: \"833f0bba-91d9-41ab-a412-9e8ae8c38ba2\") " pod="openstack/kube-state-metrics-0" Dec 05 07:17:06 crc kubenswrapper[4997]: I1205 07:17:06.135727 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv24p\" (UniqueName: \"kubernetes.io/projected/833f0bba-91d9-41ab-a412-9e8ae8c38ba2-kube-api-access-nv24p\") pod \"kube-state-metrics-0\" (UID: \"833f0bba-91d9-41ab-a412-9e8ae8c38ba2\") " pod="openstack/kube-state-metrics-0" Dec 05 07:17:06 crc kubenswrapper[4997]: I1205 07:17:06.212443 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv24p\" (UniqueName: \"kubernetes.io/projected/833f0bba-91d9-41ab-a412-9e8ae8c38ba2-kube-api-access-nv24p\") pod \"kube-state-metrics-0\" (UID: \"833f0bba-91d9-41ab-a412-9e8ae8c38ba2\") " pod="openstack/kube-state-metrics-0" Dec 05 07:17:06 crc kubenswrapper[4997]: I1205 07:17:06.230127 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 07:17:07 crc kubenswrapper[4997]: I1205 07:17:07.063729 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 07:17:07 crc kubenswrapper[4997]: I1205 07:17:07.765576 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"833f0bba-91d9-41ab-a412-9e8ae8c38ba2","Type":"ContainerStarted","Data":"60369b249264cabd66d0f28c6475b7de96dd2b3529ef70939c79e935e3edce00"} Dec 05 07:17:07 crc kubenswrapper[4997]: I1205 07:17:07.862806 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vnncj"] Dec 05 07:17:07 crc kubenswrapper[4997]: I1205 07:17:07.908636 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnncj" Dec 05 07:17:07 crc kubenswrapper[4997]: I1205 07:17:07.913199 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-xghdc" Dec 05 07:17:07 crc kubenswrapper[4997]: I1205 07:17:07.913716 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 05 07:17:07 crc kubenswrapper[4997]: I1205 07:17:07.914090 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 05 07:17:07 crc kubenswrapper[4997]: I1205 07:17:07.916777 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-sctbp"] Dec 05 07:17:07 crc kubenswrapper[4997]: I1205 07:17:07.924800 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:07 crc kubenswrapper[4997]: I1205 07:17:07.943678 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-sctbp"] Dec 05 07:17:07 crc kubenswrapper[4997]: I1205 07:17:07.960731 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vnncj"] Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.045723 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-run\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.045786 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5388a6f5-ecde-42db-9c94-a52c17e8a302-scripts\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.045808 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-log\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.045840 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-combined-ca-bundle\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.045873 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbvgw\" (UniqueName: \"kubernetes.io/projected/5388a6f5-ecde-42db-9c94-a52c17e8a302-kube-api-access-vbvgw\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.045892 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-lib\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.045921 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-scripts\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.045958 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-ovn-controller-tls-certs\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.045983 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-etc-ovs\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.046007 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-log-ovn\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.046024 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dk79\" (UniqueName: \"kubernetes.io/projected/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-kube-api-access-7dk79\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.046044 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-run-ovn\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.046076 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-run\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.147677 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-ovn-controller-tls-certs\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.147742 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-etc-ovs\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.147789 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-log-ovn\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.147813 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dk79\" (UniqueName: \"kubernetes.io/projected/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-kube-api-access-7dk79\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.147835 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-run-ovn\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.147869 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-run\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.147896 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-run\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.147913 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5388a6f5-ecde-42db-9c94-a52c17e8a302-scripts\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.147936 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-log\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.147967 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-combined-ca-bundle\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.147996 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-lib\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.148012 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbvgw\" (UniqueName: \"kubernetes.io/projected/5388a6f5-ecde-42db-9c94-a52c17e8a302-kube-api-access-vbvgw\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.148039 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-scripts\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.150227 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-scripts\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.150773 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-run\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.151051 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-etc-ovs\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.151075 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-log-ovn\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.152599 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-lib\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.152741 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-log\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.152836 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-run-ovn\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.152886 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-run\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.154230 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5388a6f5-ecde-42db-9c94-a52c17e8a302-scripts\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.159155 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-combined-ca-bundle\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.159442 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-ovn-controller-tls-certs\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.223220 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbvgw\" (UniqueName: \"kubernetes.io/projected/5388a6f5-ecde-42db-9c94-a52c17e8a302-kube-api-access-vbvgw\") pod \"ovn-controller-ovs-sctbp\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.236559 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dk79\" (UniqueName: \"kubernetes.io/projected/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-kube-api-access-7dk79\") pod \"ovn-controller-vnncj\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.275688 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnncj" Dec 05 07:17:08 crc kubenswrapper[4997]: I1205 07:17:08.292905 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.500301 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.503038 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.505746 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.506064 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-5rczl" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.506846 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.507576 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.508426 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.511699 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.673287 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.673717 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lrqm\" (UniqueName: \"kubernetes.io/projected/dd426a72-68fd-46b9-b427-e193a70aad3a-kube-api-access-5lrqm\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.673768 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.673814 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd426a72-68fd-46b9-b427-e193a70aad3a-config\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.673894 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.673930 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.673951 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd426a72-68fd-46b9-b427-e193a70aad3a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.673982 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dd426a72-68fd-46b9-b427-e193a70aad3a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.749991 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.752660 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.757403 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.759674 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-dfqvs" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.759794 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.759985 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.765323 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.777463 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd426a72-68fd-46b9-b427-e193a70aad3a-config\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.777534 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.777590 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.777639 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd426a72-68fd-46b9-b427-e193a70aad3a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.777675 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dd426a72-68fd-46b9-b427-e193a70aad3a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.777762 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.777799 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lrqm\" (UniqueName: \"kubernetes.io/projected/dd426a72-68fd-46b9-b427-e193a70aad3a-kube-api-access-5lrqm\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.777839 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.778246 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.779416 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd426a72-68fd-46b9-b427-e193a70aad3a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.779875 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd426a72-68fd-46b9-b427-e193a70aad3a-config\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.781879 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dd426a72-68fd-46b9-b427-e193a70aad3a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.789052 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.802014 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.806359 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lrqm\" (UniqueName: \"kubernetes.io/projected/dd426a72-68fd-46b9-b427-e193a70aad3a-kube-api-access-5lrqm\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.814585 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.817811 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.848779 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.879225 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c6db4c3-5869-4a82-900c-5718abfe3a3b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.879340 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c6db4c3-5869-4a82-900c-5718abfe3a3b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.879397 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.879421 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.879447 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c6db4c3-5869-4a82-900c-5718abfe3a3b-config\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.879474 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.879517 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.879550 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4bkg\" (UniqueName: \"kubernetes.io/projected/7c6db4c3-5869-4a82-900c-5718abfe3a3b-kube-api-access-q4bkg\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.993076 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c6db4c3-5869-4a82-900c-5718abfe3a3b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.993132 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c6db4c3-5869-4a82-900c-5718abfe3a3b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.993220 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.993243 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.993287 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c6db4c3-5869-4a82-900c-5718abfe3a3b-config\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.993343 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.993373 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.993430 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4bkg\" (UniqueName: \"kubernetes.io/projected/7c6db4c3-5869-4a82-900c-5718abfe3a3b-kube-api-access-q4bkg\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.994585 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c6db4c3-5869-4a82-900c-5718abfe3a3b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.994837 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.995487 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c6db4c3-5869-4a82-900c-5718abfe3a3b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:12 crc kubenswrapper[4997]: I1205 07:17:12.996840 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c6db4c3-5869-4a82-900c-5718abfe3a3b-config\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.011688 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4bkg\" (UniqueName: \"kubernetes.io/projected/7c6db4c3-5869-4a82-900c-5718abfe3a3b-kube-api-access-q4bkg\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.012474 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.024516 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.030856 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.083585 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.382286 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.701380 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-sctbp"] Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.718981 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vnncj"] Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.772097 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-89tk2"] Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.773569 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-89tk2"] Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.773761 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.778986 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.910472 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb2d44ce-8c93-493e-9393-569c208f076f-combined-ca-bundle\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.910914 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf6bt\" (UniqueName: \"kubernetes.io/projected/cb2d44ce-8c93-493e-9393-569c208f076f-kube-api-access-lf6bt\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.910998 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cb2d44ce-8c93-493e-9393-569c208f076f-ovn-rundir\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.911090 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cb2d44ce-8c93-493e-9393-569c208f076f-ovs-rundir\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.911113 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb2d44ce-8c93-493e-9393-569c208f076f-config\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:13 crc kubenswrapper[4997]: I1205 07:17:13.911242 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb2d44ce-8c93-493e-9393-569c208f076f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:14 crc kubenswrapper[4997]: I1205 07:17:14.013071 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cb2d44ce-8c93-493e-9393-569c208f076f-ovs-rundir\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:14 crc kubenswrapper[4997]: I1205 07:17:14.013128 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb2d44ce-8c93-493e-9393-569c208f076f-config\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:14 crc kubenswrapper[4997]: I1205 07:17:14.013157 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb2d44ce-8c93-493e-9393-569c208f076f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:14 crc kubenswrapper[4997]: I1205 07:17:14.013193 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb2d44ce-8c93-493e-9393-569c208f076f-combined-ca-bundle\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:14 crc kubenswrapper[4997]: I1205 07:17:14.013252 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf6bt\" (UniqueName: \"kubernetes.io/projected/cb2d44ce-8c93-493e-9393-569c208f076f-kube-api-access-lf6bt\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:14 crc kubenswrapper[4997]: I1205 07:17:14.013297 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cb2d44ce-8c93-493e-9393-569c208f076f-ovn-rundir\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:14 crc kubenswrapper[4997]: I1205 07:17:14.013865 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cb2d44ce-8c93-493e-9393-569c208f076f-ovn-rundir\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:14 crc kubenswrapper[4997]: I1205 07:17:14.013951 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cb2d44ce-8c93-493e-9393-569c208f076f-ovs-rundir\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:14 crc kubenswrapper[4997]: I1205 07:17:14.015314 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb2d44ce-8c93-493e-9393-569c208f076f-config\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:14 crc kubenswrapper[4997]: I1205 07:17:14.023469 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb2d44ce-8c93-493e-9393-569c208f076f-combined-ca-bundle\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:14 crc kubenswrapper[4997]: I1205 07:17:14.028515 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb2d44ce-8c93-493e-9393-569c208f076f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:14 crc kubenswrapper[4997]: I1205 07:17:14.038290 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf6bt\" (UniqueName: \"kubernetes.io/projected/cb2d44ce-8c93-493e-9393-569c208f076f-kube-api-access-lf6bt\") pod \"ovn-controller-metrics-89tk2\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:14 crc kubenswrapper[4997]: I1205 07:17:14.144119 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:17:25 crc kubenswrapper[4997]: I1205 07:17:25.031663 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vnncj" event={"ID":"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5","Type":"ContainerStarted","Data":"7206db5b438aaffeeecb9087bca02ee34c156f6a8cbf254755a36acf9ea0c826"} Dec 05 07:17:25 crc kubenswrapper[4997]: I1205 07:17:25.034150 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sctbp" event={"ID":"5388a6f5-ecde-42db-9c94-a52c17e8a302","Type":"ContainerStarted","Data":"24e4b91d042a02125f23e127940e3768706e261ea237a3d9728dcbc9a68346aa"} Dec 05 07:17:35 crc kubenswrapper[4997]: E1205 07:17:35.027550 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a" Dec 05 07:17:35 crc kubenswrapper[4997]: E1205 07:17:35.028383 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ffwz9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:17:35 crc kubenswrapper[4997]: E1205 07:17:35.029778 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" Dec 05 07:17:35 crc kubenswrapper[4997]: E1205 07:17:35.139920 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" Dec 05 07:17:44 crc kubenswrapper[4997]: E1205 07:17:44.569195 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d" Dec 05 07:17:44 crc kubenswrapper[4997]: E1205 07:17:44.570166 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5ngbr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:17:44 crc kubenswrapper[4997]: E1205 07:17:44.571454 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" Dec 05 07:17:45 crc kubenswrapper[4997]: E1205 07:17:45.210308 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached@sha256:dad2336390cae6705133deefaa09c9e39512cf29133aa009006e3962c8022108" Dec 05 07:17:45 crc kubenswrapper[4997]: E1205 07:17:45.210605 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached@sha256:dad2336390cae6705133deefaa09c9e39512cf29133aa009006e3962c8022108,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n67dh668hdfhd4h595h5f6hf7h57bh54dh5f9hc5h64fh5ddh585hbchch5b8h56chdh54dh57ch5ffh688h575h5fch97hcfh5b5h5c4h85h665h57dq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9hrfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(ef43a612-a3df-4b50-9bf3-e5d8098cd0fe): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:17:45 crc kubenswrapper[4997]: E1205 07:17:45.211915 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="ef43a612-a3df-4b50-9bf3-e5d8098cd0fe" Dec 05 07:17:45 crc kubenswrapper[4997]: E1205 07:17:45.227935 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a" Dec 05 07:17:45 crc kubenswrapper[4997]: E1205 07:17:45.228162 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6jwls,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(0601e553-5305-4f54-9a2d-43b43a1a5dd4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:17:45 crc kubenswrapper[4997]: E1205 07:17:45.229464 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="0601e553-5305-4f54-9a2d-43b43a1a5dd4" Dec 05 07:17:45 crc kubenswrapper[4997]: E1205 07:17:45.235197 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d" Dec 05 07:17:45 crc kubenswrapper[4997]: E1205 07:17:45.235476 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ps2f6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(0f158874-152b-46ce-ac42-d202684853ca): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:17:45 crc kubenswrapper[4997]: E1205 07:17:45.237833 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="0f158874-152b-46ce-ac42-d202684853ca" Dec 05 07:17:45 crc kubenswrapper[4997]: E1205 07:17:45.253035 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="0f158874-152b-46ce-ac42-d202684853ca" Dec 05 07:17:45 crc kubenswrapper[4997]: E1205 07:17:45.253101 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d\\\"\"" pod="openstack/rabbitmq-server-0" podUID="1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" Dec 05 07:17:45 crc kubenswrapper[4997]: E1205 07:17:45.253240 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a\\\"\"" pod="openstack/openstack-galera-0" podUID="0601e553-5305-4f54-9a2d-43b43a1a5dd4" Dec 05 07:17:45 crc kubenswrapper[4997]: E1205 07:17:45.253466 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached@sha256:dad2336390cae6705133deefaa09c9e39512cf29133aa009006e3962c8022108\\\"\"" pod="openstack/memcached-0" podUID="ef43a612-a3df-4b50-9bf3-e5d8098cd0fe" Dec 05 07:17:45 crc kubenswrapper[4997]: I1205 07:17:45.944900 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 07:17:47 crc kubenswrapper[4997]: E1205 07:17:47.045416 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:1e53a53dfe9b3cb757e4d666e76c8989941eb4f0b98d629a7f697a1693aacb17" Dec 05 07:17:47 crc kubenswrapper[4997]: E1205 07:17:47.046223 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:ovsdb-server-init,Image:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:1e53a53dfe9b3cb757e4d666e76c8989941eb4f0b98d629a7f697a1693aacb17,Command:[/usr/local/bin/container-scripts/init-ovsdb-server.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n57h556h54fhddh5d8h56dh644hf8h5dh57ch575h68bh65h599h5f9h67fh5fch5bch569h7h5f6h5d9h649h5bbhfch579h657h5c4h569h698hf6h657q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-ovs,ReadOnly:false,MountPath:/etc/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log,ReadOnly:false,MountPath:/var/log/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-lib,ReadOnly:false,MountPath:/var/lib/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vbvgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-ovs-sctbp_openstack(5388a6f5-ecde-42db-9c94-a52c17e8a302): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:17:47 crc kubenswrapper[4997]: E1205 07:17:47.047914 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdb-server-init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-ovs-sctbp" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" Dec 05 07:17:47 crc kubenswrapper[4997]: E1205 07:17:47.275776 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdb-server-init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:1e53a53dfe9b3cb757e4d666e76c8989941eb4f0b98d629a7f697a1693aacb17\\\"\"" pod="openstack/ovn-controller-ovs-sctbp" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" Dec 05 07:17:48 crc kubenswrapper[4997]: E1205 07:17:48.657517 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 05 07:17:48 crc kubenswrapper[4997]: E1205 07:17:48.657774 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fnrzh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-567c455747-7vz4d_openstack(abd5fd09-09f8-4c55-8f80-f550cd7bdc0b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:17:48 crc kubenswrapper[4997]: E1205 07:17:48.660011 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-567c455747-7vz4d" podUID="abd5fd09-09f8-4c55-8f80-f550cd7bdc0b" Dec 05 07:17:48 crc kubenswrapper[4997]: W1205 07:17:48.697525 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd426a72_68fd_46b9_b427_e193a70aad3a.slice/crio-eef2665cee84fa9c47233a9b8e042e05cba63a5f3ba4909bf0df0fd50eb9127e WatchSource:0}: Error finding container eef2665cee84fa9c47233a9b8e042e05cba63a5f3ba4909bf0df0fd50eb9127e: Status 404 returned error can't find the container with id eef2665cee84fa9c47233a9b8e042e05cba63a5f3ba4909bf0df0fd50eb9127e Dec 05 07:17:48 crc kubenswrapper[4997]: E1205 07:17:48.738438 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 05 07:17:48 crc kubenswrapper[4997]: E1205 07:17:48.739119 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5zfwr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-859d485f47-cdr9f_openstack(a8cdbec8-a2eb-48cb-bfe9-b508c4e78416): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:17:48 crc kubenswrapper[4997]: E1205 07:17:48.740367 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-859d485f47-cdr9f" podUID="a8cdbec8-a2eb-48cb-bfe9-b508c4e78416" Dec 05 07:17:48 crc kubenswrapper[4997]: E1205 07:17:48.753658 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 05 07:17:48 crc kubenswrapper[4997]: E1205 07:17:48.753859 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2fhg9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5cd484bb89-wvjqk_openstack(3da07353-d51d-4887-9b4b-b8e7d60520ac): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:17:48 crc kubenswrapper[4997]: E1205 07:17:48.755077 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5cd484bb89-wvjqk" podUID="3da07353-d51d-4887-9b4b-b8e7d60520ac" Dec 05 07:17:48 crc kubenswrapper[4997]: E1205 07:17:48.767276 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 05 07:17:48 crc kubenswrapper[4997]: E1205 07:17:48.767439 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9hmdt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-cb666b895-snwmr_openstack(33be1eda-85ae-44a3-85ee-e99b66f32fae): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:17:48 crc kubenswrapper[4997]: E1205 07:17:48.769903 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-cb666b895-snwmr" podUID="33be1eda-85ae-44a3-85ee-e99b66f32fae" Dec 05 07:17:49 crc kubenswrapper[4997]: I1205 07:17:49.307647 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 07:17:49 crc kubenswrapper[4997]: I1205 07:17:49.324906 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"dd426a72-68fd-46b9-b427-e193a70aad3a","Type":"ContainerStarted","Data":"eef2665cee84fa9c47233a9b8e042e05cba63a5f3ba4909bf0df0fd50eb9127e"} Dec 05 07:17:49 crc kubenswrapper[4997]: E1205 07:17:49.326416 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792\\\"\"" pod="openstack/dnsmasq-dns-859d485f47-cdr9f" podUID="a8cdbec8-a2eb-48cb-bfe9-b508c4e78416" Dec 05 07:17:49 crc kubenswrapper[4997]: E1205 07:17:49.326925 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792\\\"\"" pod="openstack/dnsmasq-dns-cb666b895-snwmr" podUID="33be1eda-85ae-44a3-85ee-e99b66f32fae" Dec 05 07:17:49 crc kubenswrapper[4997]: I1205 07:17:49.442034 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-89tk2"] Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.336725 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-89tk2" event={"ID":"cb2d44ce-8c93-493e-9393-569c208f076f","Type":"ContainerStarted","Data":"ffc8b8171d24cc0361e4ef7a4a54bb55664d374163a75fc1a7e0ed31f219a224"} Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.338104 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7c6db4c3-5869-4a82-900c-5718abfe3a3b","Type":"ContainerStarted","Data":"cec8912e079eb100a0bcc221020ceff6158597b2969316c396c5a9aea58e1df6"} Dec 05 07:17:50 crc kubenswrapper[4997]: E1205 07:17:50.361423 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Dec 05 07:17:50 crc kubenswrapper[4997]: E1205 07:17:50.361478 4997 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Dec 05 07:17:50 crc kubenswrapper[4997]: E1205 07:17:50.361648 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nv24p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(833f0bba-91d9-41ab-a412-9e8ae8c38ba2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 07:17:50 crc kubenswrapper[4997]: E1205 07:17:50.362926 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="833f0bba-91d9-41ab-a412-9e8ae8c38ba2" Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.442644 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-7vz4d" Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.558624 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-dns-svc\") pod \"abd5fd09-09f8-4c55-8f80-f550cd7bdc0b\" (UID: \"abd5fd09-09f8-4c55-8f80-f550cd7bdc0b\") " Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.558713 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-config\") pod \"abd5fd09-09f8-4c55-8f80-f550cd7bdc0b\" (UID: \"abd5fd09-09f8-4c55-8f80-f550cd7bdc0b\") " Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.558759 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnrzh\" (UniqueName: \"kubernetes.io/projected/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-kube-api-access-fnrzh\") pod \"abd5fd09-09f8-4c55-8f80-f550cd7bdc0b\" (UID: \"abd5fd09-09f8-4c55-8f80-f550cd7bdc0b\") " Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.559414 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "abd5fd09-09f8-4c55-8f80-f550cd7bdc0b" (UID: "abd5fd09-09f8-4c55-8f80-f550cd7bdc0b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.559515 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-config" (OuterVolumeSpecName: "config") pod "abd5fd09-09f8-4c55-8f80-f550cd7bdc0b" (UID: "abd5fd09-09f8-4c55-8f80-f550cd7bdc0b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.566686 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-kube-api-access-fnrzh" (OuterVolumeSpecName: "kube-api-access-fnrzh") pod "abd5fd09-09f8-4c55-8f80-f550cd7bdc0b" (UID: "abd5fd09-09f8-4c55-8f80-f550cd7bdc0b"). InnerVolumeSpecName "kube-api-access-fnrzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.661666 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.661701 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.661719 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnrzh\" (UniqueName: \"kubernetes.io/projected/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b-kube-api-access-fnrzh\") on node \"crc\" DevicePath \"\"" Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.745989 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-wvjqk" Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.865321 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3da07353-d51d-4887-9b4b-b8e7d60520ac-config\") pod \"3da07353-d51d-4887-9b4b-b8e7d60520ac\" (UID: \"3da07353-d51d-4887-9b4b-b8e7d60520ac\") " Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.865902 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3da07353-d51d-4887-9b4b-b8e7d60520ac-config" (OuterVolumeSpecName: "config") pod "3da07353-d51d-4887-9b4b-b8e7d60520ac" (UID: "3da07353-d51d-4887-9b4b-b8e7d60520ac"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.865941 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fhg9\" (UniqueName: \"kubernetes.io/projected/3da07353-d51d-4887-9b4b-b8e7d60520ac-kube-api-access-2fhg9\") pod \"3da07353-d51d-4887-9b4b-b8e7d60520ac\" (UID: \"3da07353-d51d-4887-9b4b-b8e7d60520ac\") " Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.866438 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3da07353-d51d-4887-9b4b-b8e7d60520ac-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.869841 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3da07353-d51d-4887-9b4b-b8e7d60520ac-kube-api-access-2fhg9" (OuterVolumeSpecName: "kube-api-access-2fhg9") pod "3da07353-d51d-4887-9b4b-b8e7d60520ac" (UID: "3da07353-d51d-4887-9b4b-b8e7d60520ac"). InnerVolumeSpecName "kube-api-access-2fhg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:17:50 crc kubenswrapper[4997]: I1205 07:17:50.972193 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fhg9\" (UniqueName: \"kubernetes.io/projected/3da07353-d51d-4887-9b4b-b8e7d60520ac-kube-api-access-2fhg9\") on node \"crc\" DevicePath \"\"" Dec 05 07:17:51 crc kubenswrapper[4997]: I1205 07:17:51.348930 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-7vz4d" Dec 05 07:17:51 crc kubenswrapper[4997]: I1205 07:17:51.348928 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567c455747-7vz4d" event={"ID":"abd5fd09-09f8-4c55-8f80-f550cd7bdc0b","Type":"ContainerDied","Data":"985db3af34cf98eb4787e3aa221955afb41d045625482bfb8a71d4f220403c53"} Dec 05 07:17:51 crc kubenswrapper[4997]: I1205 07:17:51.351496 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vnncj" event={"ID":"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5","Type":"ContainerStarted","Data":"f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38"} Dec 05 07:17:51 crc kubenswrapper[4997]: I1205 07:17:51.351647 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-vnncj" Dec 05 07:17:51 crc kubenswrapper[4997]: I1205 07:17:51.352872 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd484bb89-wvjqk" event={"ID":"3da07353-d51d-4887-9b4b-b8e7d60520ac","Type":"ContainerDied","Data":"4048e37b1997cdcaf61d32c6c267f02ee9a2b6654565bf2b8391804db6a532a6"} Dec 05 07:17:51 crc kubenswrapper[4997]: I1205 07:17:51.352913 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-wvjqk" Dec 05 07:17:51 crc kubenswrapper[4997]: I1205 07:17:51.354787 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"dd426a72-68fd-46b9-b427-e193a70aad3a","Type":"ContainerStarted","Data":"a5b9db6868281fc770b42b2466291fa6a87195e877608ca9e09e648d60355857"} Dec 05 07:17:51 crc kubenswrapper[4997]: I1205 07:17:51.357537 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b","Type":"ContainerStarted","Data":"5f2ebfbb61e660700f2975a4f58a9ce986878add969191a622893e2fcd4fd58c"} Dec 05 07:17:51 crc kubenswrapper[4997]: E1205 07:17:51.359343 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb\\\"\"" pod="openstack/kube-state-metrics-0" podUID="833f0bba-91d9-41ab-a412-9e8ae8c38ba2" Dec 05 07:17:51 crc kubenswrapper[4997]: I1205 07:17:51.388513 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-vnncj" podStartSLOduration=18.5343401 podStartE2EDuration="44.388491561s" podCreationTimestamp="2025-12-05 07:17:07 +0000 UTC" firstStartedPulling="2025-12-05 07:17:24.917370753 +0000 UTC m=+1345.446278024" lastFinishedPulling="2025-12-05 07:17:50.771522234 +0000 UTC m=+1371.300429485" observedRunningTime="2025-12-05 07:17:51.373340332 +0000 UTC m=+1371.902247603" watchObservedRunningTime="2025-12-05 07:17:51.388491561 +0000 UTC m=+1371.917398822" Dec 05 07:17:51 crc kubenswrapper[4997]: I1205 07:17:51.474907 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-567c455747-7vz4d"] Dec 05 07:17:51 crc kubenswrapper[4997]: I1205 07:17:51.490393 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-567c455747-7vz4d"] Dec 05 07:17:51 crc kubenswrapper[4997]: I1205 07:17:51.504373 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-wvjqk"] Dec 05 07:17:51 crc kubenswrapper[4997]: I1205 07:17:51.515428 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-wvjqk"] Dec 05 07:17:51 crc kubenswrapper[4997]: I1205 07:17:51.762797 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3da07353-d51d-4887-9b4b-b8e7d60520ac" path="/var/lib/kubelet/pods/3da07353-d51d-4887-9b4b-b8e7d60520ac/volumes" Dec 05 07:17:51 crc kubenswrapper[4997]: I1205 07:17:51.764083 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abd5fd09-09f8-4c55-8f80-f550cd7bdc0b" path="/var/lib/kubelet/pods/abd5fd09-09f8-4c55-8f80-f550cd7bdc0b/volumes" Dec 05 07:17:52 crc kubenswrapper[4997]: I1205 07:17:52.370703 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7c6db4c3-5869-4a82-900c-5718abfe3a3b","Type":"ContainerStarted","Data":"b92bd2595a4bc2bb590760b577f8f2341220a48a3841768940be448c3738ff62"} Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.404146 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-89tk2" event={"ID":"cb2d44ce-8c93-493e-9393-569c208f076f","Type":"ContainerStarted","Data":"c77fadb14da56ec3ffb9023a9bbf4752c6ab4d4d2977ec528e7ff29efb0be39c"} Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.409597 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7c6db4c3-5869-4a82-900c-5718abfe3a3b","Type":"ContainerStarted","Data":"5d79e06e59e6f9cd89fbe9c40174b9a6f2059177958ca28bb43f3b31b4fdf3bd"} Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.413349 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"dd426a72-68fd-46b9-b427-e193a70aad3a","Type":"ContainerStarted","Data":"8058fdc67614c4fcf41bcd46098bf3584ceead6983f82e156bf569ef8aa5918f"} Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.439674 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-89tk2" podStartSLOduration=37.178710609 podStartE2EDuration="41.439649224s" podCreationTimestamp="2025-12-05 07:17:13 +0000 UTC" firstStartedPulling="2025-12-05 07:17:49.645382468 +0000 UTC m=+1370.174289729" lastFinishedPulling="2025-12-05 07:17:53.906321073 +0000 UTC m=+1374.435228344" observedRunningTime="2025-12-05 07:17:54.424332931 +0000 UTC m=+1374.953240202" watchObservedRunningTime="2025-12-05 07:17:54.439649224 +0000 UTC m=+1374.968556495" Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.453220 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=39.169376768 podStartE2EDuration="43.45318619s" podCreationTimestamp="2025-12-05 07:17:11 +0000 UTC" firstStartedPulling="2025-12-05 07:17:49.626064647 +0000 UTC m=+1370.154971908" lastFinishedPulling="2025-12-05 07:17:53.909874069 +0000 UTC m=+1374.438781330" observedRunningTime="2025-12-05 07:17:54.451740971 +0000 UTC m=+1374.980648252" watchObservedRunningTime="2025-12-05 07:17:54.45318619 +0000 UTC m=+1374.982093451" Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.476408 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=38.287100086 podStartE2EDuration="43.476381745s" podCreationTimestamp="2025-12-05 07:17:11 +0000 UTC" firstStartedPulling="2025-12-05 07:17:48.701048063 +0000 UTC m=+1369.229955324" lastFinishedPulling="2025-12-05 07:17:53.890329722 +0000 UTC m=+1374.419236983" observedRunningTime="2025-12-05 07:17:54.475359487 +0000 UTC m=+1375.004266758" watchObservedRunningTime="2025-12-05 07:17:54.476381745 +0000 UTC m=+1375.005289006" Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.775959 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-859d485f47-cdr9f"] Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.824508 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-26h6q"] Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.826586 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.830002 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.850110 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-26h6q"] Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.850396 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.916832 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.981572 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r87rx\" (UniqueName: \"kubernetes.io/projected/4aa8aee7-517a-42f7-9035-0acb7133fa9d-kube-api-access-r87rx\") pod \"dnsmasq-dns-57db9b5bc9-26h6q\" (UID: \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\") " pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.981673 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-config\") pod \"dnsmasq-dns-57db9b5bc9-26h6q\" (UID: \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\") " pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.981720 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-ovsdbserver-nb\") pod \"dnsmasq-dns-57db9b5bc9-26h6q\" (UID: \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\") " pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.981794 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-dns-svc\") pod \"dnsmasq-dns-57db9b5bc9-26h6q\" (UID: \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\") " pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:17:54 crc kubenswrapper[4997]: I1205 07:17:54.986947 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-snwmr"] Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.037047 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-ht85n"] Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.038979 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.042522 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.084218 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-dns-svc\") pod \"dnsmasq-dns-57db9b5bc9-26h6q\" (UID: \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\") " pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.084315 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r87rx\" (UniqueName: \"kubernetes.io/projected/4aa8aee7-517a-42f7-9035-0acb7133fa9d-kube-api-access-r87rx\") pod \"dnsmasq-dns-57db9b5bc9-26h6q\" (UID: \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\") " pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.084370 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-config\") pod \"dnsmasq-dns-57db9b5bc9-26h6q\" (UID: \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\") " pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.084401 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-ovsdbserver-nb\") pod \"dnsmasq-dns-57db9b5bc9-26h6q\" (UID: \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\") " pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.085431 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-ovsdbserver-nb\") pod \"dnsmasq-dns-57db9b5bc9-26h6q\" (UID: \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\") " pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.086286 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-config\") pod \"dnsmasq-dns-57db9b5bc9-26h6q\" (UID: \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\") " pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.090101 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-dns-svc\") pod \"dnsmasq-dns-57db9b5bc9-26h6q\" (UID: \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\") " pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.102462 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-ht85n"] Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.145489 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r87rx\" (UniqueName: \"kubernetes.io/projected/4aa8aee7-517a-42f7-9035-0acb7133fa9d-kube-api-access-r87rx\") pod \"dnsmasq-dns-57db9b5bc9-26h6q\" (UID: \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\") " pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.154237 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.186671 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gtj5\" (UniqueName: \"kubernetes.io/projected/a78999af-0d52-4d34-9bf1-dda9848049bd-kube-api-access-7gtj5\") pod \"dnsmasq-dns-db7757ddc-ht85n\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.186733 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-ovsdbserver-nb\") pod \"dnsmasq-dns-db7757ddc-ht85n\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.186776 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-ovsdbserver-sb\") pod \"dnsmasq-dns-db7757ddc-ht85n\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.186827 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-dns-svc\") pod \"dnsmasq-dns-db7757ddc-ht85n\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.186854 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-config\") pod \"dnsmasq-dns-db7757ddc-ht85n\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.301728 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gtj5\" (UniqueName: \"kubernetes.io/projected/a78999af-0d52-4d34-9bf1-dda9848049bd-kube-api-access-7gtj5\") pod \"dnsmasq-dns-db7757ddc-ht85n\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.302433 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-ovsdbserver-nb\") pod \"dnsmasq-dns-db7757ddc-ht85n\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.302679 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-ovsdbserver-sb\") pod \"dnsmasq-dns-db7757ddc-ht85n\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.302853 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-dns-svc\") pod \"dnsmasq-dns-db7757ddc-ht85n\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.302901 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-config\") pod \"dnsmasq-dns-db7757ddc-ht85n\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.305111 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-ovsdbserver-nb\") pod \"dnsmasq-dns-db7757ddc-ht85n\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.304971 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-dns-svc\") pod \"dnsmasq-dns-db7757ddc-ht85n\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.304438 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-ovsdbserver-sb\") pod \"dnsmasq-dns-db7757ddc-ht85n\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.305423 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-config\") pod \"dnsmasq-dns-db7757ddc-ht85n\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.310156 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-859d485f47-cdr9f" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.331996 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gtj5\" (UniqueName: \"kubernetes.io/projected/a78999af-0d52-4d34-9bf1-dda9848049bd-kube-api-access-7gtj5\") pod \"dnsmasq-dns-db7757ddc-ht85n\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.383320 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.400836 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.407696 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zfwr\" (UniqueName: \"kubernetes.io/projected/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-kube-api-access-5zfwr\") pod \"a8cdbec8-a2eb-48cb-bfe9-b508c4e78416\" (UID: \"a8cdbec8-a2eb-48cb-bfe9-b508c4e78416\") " Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.407782 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-config\") pod \"a8cdbec8-a2eb-48cb-bfe9-b508c4e78416\" (UID: \"a8cdbec8-a2eb-48cb-bfe9-b508c4e78416\") " Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.407988 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-dns-svc\") pod \"a8cdbec8-a2eb-48cb-bfe9-b508c4e78416\" (UID: \"a8cdbec8-a2eb-48cb-bfe9-b508c4e78416\") " Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.409014 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a8cdbec8-a2eb-48cb-bfe9-b508c4e78416" (UID: "a8cdbec8-a2eb-48cb-bfe9-b508c4e78416"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.409146 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-config" (OuterVolumeSpecName: "config") pod "a8cdbec8-a2eb-48cb-bfe9-b508c4e78416" (UID: "a8cdbec8-a2eb-48cb-bfe9-b508c4e78416"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.420845 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-kube-api-access-5zfwr" (OuterVolumeSpecName: "kube-api-access-5zfwr") pod "a8cdbec8-a2eb-48cb-bfe9-b508c4e78416" (UID: "a8cdbec8-a2eb-48cb-bfe9-b508c4e78416"). InnerVolumeSpecName "kube-api-access-5zfwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.432985 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.441754 4997 generic.go:334] "Generic (PLEG): container finished" podID="7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" containerID="5f2ebfbb61e660700f2975a4f58a9ce986878add969191a622893e2fcd4fd58c" exitCode=0 Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.441829 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b","Type":"ContainerDied","Data":"5f2ebfbb61e660700f2975a4f58a9ce986878add969191a622893e2fcd4fd58c"} Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.456769 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-859d485f47-cdr9f" event={"ID":"a8cdbec8-a2eb-48cb-bfe9-b508c4e78416","Type":"ContainerDied","Data":"b47bc56d0f868d6b9f7e3dd07bdf7a40ee7d56e135ebd3908f07881e9414f89b"} Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.456957 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-859d485f47-cdr9f" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.458864 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.461533 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.486403 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-snwmr" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.490434 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-26h6q"] Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.522984 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hmdt\" (UniqueName: \"kubernetes.io/projected/33be1eda-85ae-44a3-85ee-e99b66f32fae-kube-api-access-9hmdt\") pod \"33be1eda-85ae-44a3-85ee-e99b66f32fae\" (UID: \"33be1eda-85ae-44a3-85ee-e99b66f32fae\") " Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.523266 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33be1eda-85ae-44a3-85ee-e99b66f32fae-dns-svc\") pod \"33be1eda-85ae-44a3-85ee-e99b66f32fae\" (UID: \"33be1eda-85ae-44a3-85ee-e99b66f32fae\") " Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.523309 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33be1eda-85ae-44a3-85ee-e99b66f32fae-config\") pod \"33be1eda-85ae-44a3-85ee-e99b66f32fae\" (UID: \"33be1eda-85ae-44a3-85ee-e99b66f32fae\") " Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.524122 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zfwr\" (UniqueName: \"kubernetes.io/projected/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-kube-api-access-5zfwr\") on node \"crc\" DevicePath \"\"" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.524141 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.524151 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.526880 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33be1eda-85ae-44a3-85ee-e99b66f32fae-config" (OuterVolumeSpecName: "config") pod "33be1eda-85ae-44a3-85ee-e99b66f32fae" (UID: "33be1eda-85ae-44a3-85ee-e99b66f32fae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.527074 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33be1eda-85ae-44a3-85ee-e99b66f32fae-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "33be1eda-85ae-44a3-85ee-e99b66f32fae" (UID: "33be1eda-85ae-44a3-85ee-e99b66f32fae"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.529446 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-859d485f47-cdr9f"] Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.536570 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33be1eda-85ae-44a3-85ee-e99b66f32fae-kube-api-access-9hmdt" (OuterVolumeSpecName: "kube-api-access-9hmdt") pod "33be1eda-85ae-44a3-85ee-e99b66f32fae" (UID: "33be1eda-85ae-44a3-85ee-e99b66f32fae"). InnerVolumeSpecName "kube-api-access-9hmdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.540157 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-859d485f47-cdr9f"] Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.626326 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33be1eda-85ae-44a3-85ee-e99b66f32fae-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.626872 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33be1eda-85ae-44a3-85ee-e99b66f32fae-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.626887 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hmdt\" (UniqueName: \"kubernetes.io/projected/33be1eda-85ae-44a3-85ee-e99b66f32fae-kube-api-access-9hmdt\") on node \"crc\" DevicePath \"\"" Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.740681 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-ht85n"] Dec 05 07:17:55 crc kubenswrapper[4997]: I1205 07:17:55.769038 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8cdbec8-a2eb-48cb-bfe9-b508c4e78416" path="/var/lib/kubelet/pods/a8cdbec8-a2eb-48cb-bfe9-b508c4e78416/volumes" Dec 05 07:17:56 crc kubenswrapper[4997]: I1205 07:17:56.463275 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-snwmr" event={"ID":"33be1eda-85ae-44a3-85ee-e99b66f32fae","Type":"ContainerDied","Data":"471c1559376966f1431d31f71ff39e3d87053a04b283a3b0f8260af1c6e3ce19"} Dec 05 07:17:56 crc kubenswrapper[4997]: I1205 07:17:56.463342 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-snwmr" Dec 05 07:17:56 crc kubenswrapper[4997]: I1205 07:17:56.465422 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-ht85n" event={"ID":"a78999af-0d52-4d34-9bf1-dda9848049bd","Type":"ContainerDied","Data":"a5997c2d9abe35118ff5784153570b011a42b33128ab4e9bd034de357280ffa5"} Dec 05 07:17:56 crc kubenswrapper[4997]: I1205 07:17:56.466441 4997 generic.go:334] "Generic (PLEG): container finished" podID="a78999af-0d52-4d34-9bf1-dda9848049bd" containerID="a5997c2d9abe35118ff5784153570b011a42b33128ab4e9bd034de357280ffa5" exitCode=0 Dec 05 07:17:56 crc kubenswrapper[4997]: I1205 07:17:56.466525 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-ht85n" event={"ID":"a78999af-0d52-4d34-9bf1-dda9848049bd","Type":"ContainerStarted","Data":"02ef4a19f1e03a85ad39ae42d31d2206ccbd037bb882f7e68f88a0f0391ea6d2"} Dec 05 07:17:56 crc kubenswrapper[4997]: I1205 07:17:56.469948 4997 generic.go:334] "Generic (PLEG): container finished" podID="4aa8aee7-517a-42f7-9035-0acb7133fa9d" containerID="442f4e07ce4939690c5144c004816022fd8cb2ab8a48ebad68d80f36953b5a6b" exitCode=0 Dec 05 07:17:56 crc kubenswrapper[4997]: I1205 07:17:56.470055 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" event={"ID":"4aa8aee7-517a-42f7-9035-0acb7133fa9d","Type":"ContainerDied","Data":"442f4e07ce4939690c5144c004816022fd8cb2ab8a48ebad68d80f36953b5a6b"} Dec 05 07:17:56 crc kubenswrapper[4997]: I1205 07:17:56.470109 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" event={"ID":"4aa8aee7-517a-42f7-9035-0acb7133fa9d","Type":"ContainerStarted","Data":"19c9a5c4e5e578d8363aaa2ac09dece879c9dcf4bddb06de624dd4369d5e5800"} Dec 05 07:17:56 crc kubenswrapper[4997]: I1205 07:17:56.475290 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b","Type":"ContainerStarted","Data":"5f394487f180524c3058b7242cb6bd0e8d1659b04d10a82da617c25ca61bffc0"} Dec 05 07:17:56 crc kubenswrapper[4997]: I1205 07:17:56.539567 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 05 07:17:56 crc kubenswrapper[4997]: I1205 07:17:56.551991 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 05 07:17:56 crc kubenswrapper[4997]: I1205 07:17:56.601241 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-snwmr"] Dec 05 07:17:56 crc kubenswrapper[4997]: I1205 07:17:56.615975 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-snwmr"] Dec 05 07:17:56 crc kubenswrapper[4997]: I1205 07:17:56.677638 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=9.175132279 podStartE2EDuration="54.67759071s" podCreationTimestamp="2025-12-05 07:17:02 +0000 UTC" firstStartedPulling="2025-12-05 07:17:05.270046989 +0000 UTC m=+1325.798954250" lastFinishedPulling="2025-12-05 07:17:50.77250541 +0000 UTC m=+1371.301412681" observedRunningTime="2025-12-05 07:17:56.660822878 +0000 UTC m=+1377.189730139" watchObservedRunningTime="2025-12-05 07:17:56.67759071 +0000 UTC m=+1377.206497971" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.136944 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.184843 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.189311 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.189501 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.189625 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-cqg8z" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.191058 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.192402 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.288014 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2f0e39f0-ca27-4656-879e-c7a20fcd904c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.288130 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f0e39f0-ca27-4656-879e-c7a20fcd904c-config\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.288159 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f0e39f0-ca27-4656-879e-c7a20fcd904c-scripts\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.288230 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.288314 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdbk9\" (UniqueName: \"kubernetes.io/projected/2f0e39f0-ca27-4656-879e-c7a20fcd904c-kube-api-access-qdbk9\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.288382 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.288523 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.391833 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2f0e39f0-ca27-4656-879e-c7a20fcd904c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.391918 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f0e39f0-ca27-4656-879e-c7a20fcd904c-config\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.391959 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f0e39f0-ca27-4656-879e-c7a20fcd904c-scripts\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.391994 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.392035 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdbk9\" (UniqueName: \"kubernetes.io/projected/2f0e39f0-ca27-4656-879e-c7a20fcd904c-kube-api-access-qdbk9\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.392076 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.392130 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.392475 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2f0e39f0-ca27-4656-879e-c7a20fcd904c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.393070 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f0e39f0-ca27-4656-879e-c7a20fcd904c-scripts\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.393172 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f0e39f0-ca27-4656-879e-c7a20fcd904c-config\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.398819 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.398987 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.399641 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.413239 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdbk9\" (UniqueName: \"kubernetes.io/projected/2f0e39f0-ca27-4656-879e-c7a20fcd904c-kube-api-access-qdbk9\") pod \"ovn-northd-0\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.484162 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-ht85n" event={"ID":"a78999af-0d52-4d34-9bf1-dda9848049bd","Type":"ContainerStarted","Data":"01368c32dffe061c73bd12e0225f1b9c8acd080ec609973a06f3aefb667f648e"} Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.484632 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.487532 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" event={"ID":"4aa8aee7-517a-42f7-9035-0acb7133fa9d","Type":"ContainerStarted","Data":"698dfc1fd532b847d0d00ef4837d9372b1f7ac959ded5989b9c324c1f8e401ea"} Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.487776 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.505588 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-db7757ddc-ht85n" podStartSLOduration=2.091569003 podStartE2EDuration="2.505569427s" podCreationTimestamp="2025-12-05 07:17:55 +0000 UTC" firstStartedPulling="2025-12-05 07:17:55.740799819 +0000 UTC m=+1376.269707080" lastFinishedPulling="2025-12-05 07:17:56.154800243 +0000 UTC m=+1376.683707504" observedRunningTime="2025-12-05 07:17:57.503426208 +0000 UTC m=+1378.032333459" watchObservedRunningTime="2025-12-05 07:17:57.505569427 +0000 UTC m=+1378.034476688" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.524366 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" podStartSLOduration=3.083545106 podStartE2EDuration="3.524347562s" podCreationTimestamp="2025-12-05 07:17:54 +0000 UTC" firstStartedPulling="2025-12-05 07:17:55.495030603 +0000 UTC m=+1376.023937864" lastFinishedPulling="2025-12-05 07:17:55.935833059 +0000 UTC m=+1376.464740320" observedRunningTime="2025-12-05 07:17:57.518889425 +0000 UTC m=+1378.047796686" watchObservedRunningTime="2025-12-05 07:17:57.524347562 +0000 UTC m=+1378.053254823" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.570254 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.766850 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33be1eda-85ae-44a3-85ee-e99b66f32fae" path="/var/lib/kubelet/pods/33be1eda-85ae-44a3-85ee-e99b66f32fae/volumes" Dec 05 07:17:57 crc kubenswrapper[4997]: I1205 07:17:57.838552 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 07:17:57 crc kubenswrapper[4997]: W1205 07:17:57.869544 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f0e39f0_ca27_4656_879e_c7a20fcd904c.slice/crio-ec9ff4ae268b09d7939f787bf96284c22711b8f29e8e136858254ff329c18a43 WatchSource:0}: Error finding container ec9ff4ae268b09d7939f787bf96284c22711b8f29e8e136858254ff329c18a43: Status 404 returned error can't find the container with id ec9ff4ae268b09d7939f787bf96284c22711b8f29e8e136858254ff329c18a43 Dec 05 07:17:58 crc kubenswrapper[4997]: I1205 07:17:58.498523 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0601e553-5305-4f54-9a2d-43b43a1a5dd4","Type":"ContainerStarted","Data":"f17f8b4ededafb62d528cf7bfb22c523561aaf354b59cd3fc4a9694a352a9d0b"} Dec 05 07:17:58 crc kubenswrapper[4997]: I1205 07:17:58.502874 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2f0e39f0-ca27-4656-879e-c7a20fcd904c","Type":"ContainerStarted","Data":"ec9ff4ae268b09d7939f787bf96284c22711b8f29e8e136858254ff329c18a43"} Dec 05 07:17:59 crc kubenswrapper[4997]: I1205 07:17:59.514449 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe","Type":"ContainerStarted","Data":"175b217107beb960b8632016072bdd9ad602c7c079fe319f224a856dd6e14c82"} Dec 05 07:17:59 crc kubenswrapper[4997]: I1205 07:17:59.516472 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 05 07:17:59 crc kubenswrapper[4997]: I1205 07:17:59.539015 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.773926921 podStartE2EDuration="56.538987207s" podCreationTimestamp="2025-12-05 07:17:03 +0000 UTC" firstStartedPulling="2025-12-05 07:17:04.990862496 +0000 UTC m=+1325.519769747" lastFinishedPulling="2025-12-05 07:17:58.755922772 +0000 UTC m=+1379.284830033" observedRunningTime="2025-12-05 07:17:59.535952675 +0000 UTC m=+1380.064859936" watchObservedRunningTime="2025-12-05 07:17:59.538987207 +0000 UTC m=+1380.067894468" Dec 05 07:18:00 crc kubenswrapper[4997]: I1205 07:18:00.344025 4997 scope.go:117] "RemoveContainer" containerID="48315b20317dd0c336ea1ece790aa8d0124814e195936c990f07e1c3d41809e4" Dec 05 07:18:00 crc kubenswrapper[4997]: I1205 07:18:00.395420 4997 scope.go:117] "RemoveContainer" containerID="fa2dd55190a2b5b0ae83603049210c478c36a853b896f8e7124df3b3276191ab" Dec 05 07:18:00 crc kubenswrapper[4997]: I1205 07:18:00.436381 4997 scope.go:117] "RemoveContainer" containerID="4c083a32c89ad64b4ad5b027fe8cac8a4ba0b56748e8c76949a3e9c849ba171d" Dec 05 07:18:00 crc kubenswrapper[4997]: I1205 07:18:00.528821 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79","Type":"ContainerStarted","Data":"711f304ea598c5aa2085e471f825cc1a9ccb2da36ded6e28c57872ee1771bbfd"} Dec 05 07:18:00 crc kubenswrapper[4997]: I1205 07:18:00.542872 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2f0e39f0-ca27-4656-879e-c7a20fcd904c","Type":"ContainerStarted","Data":"3caac945976009c44a6d871779a90a30a1333a9614f3564c29ec3e0c616fdbe7"} Dec 05 07:18:01 crc kubenswrapper[4997]: I1205 07:18:01.554174 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2f0e39f0-ca27-4656-879e-c7a20fcd904c","Type":"ContainerStarted","Data":"b7780da612c7177ca63143d557d61798aed63fe6008628f0f9c5e7fba16024f1"} Dec 05 07:18:01 crc kubenswrapper[4997]: I1205 07:18:01.555207 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 05 07:18:01 crc kubenswrapper[4997]: I1205 07:18:01.558180 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0f158874-152b-46ce-ac42-d202684853ca","Type":"ContainerStarted","Data":"3a4f08d0ae39c150584948dc3f19fc684275263ada8f51f39f56460972b674d2"} Dec 05 07:18:01 crc kubenswrapper[4997]: I1205 07:18:01.560469 4997 generic.go:334] "Generic (PLEG): container finished" podID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerID="13a5e6018eea0fdd10399142992fee1827749d6c2551be33953ed7d8cf75c944" exitCode=0 Dec 05 07:18:01 crc kubenswrapper[4997]: I1205 07:18:01.560522 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sctbp" event={"ID":"5388a6f5-ecde-42db-9c94-a52c17e8a302","Type":"ContainerDied","Data":"13a5e6018eea0fdd10399142992fee1827749d6c2551be33953ed7d8cf75c944"} Dec 05 07:18:01 crc kubenswrapper[4997]: I1205 07:18:01.583941 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.572412468 podStartE2EDuration="4.583907198s" podCreationTimestamp="2025-12-05 07:17:57 +0000 UTC" firstStartedPulling="2025-12-05 07:17:57.875593164 +0000 UTC m=+1378.404500425" lastFinishedPulling="2025-12-05 07:17:59.887087894 +0000 UTC m=+1380.415995155" observedRunningTime="2025-12-05 07:18:01.576605231 +0000 UTC m=+1382.105512542" watchObservedRunningTime="2025-12-05 07:18:01.583907198 +0000 UTC m=+1382.112814479" Dec 05 07:18:02 crc kubenswrapper[4997]: I1205 07:18:02.571180 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sctbp" event={"ID":"5388a6f5-ecde-42db-9c94-a52c17e8a302","Type":"ContainerStarted","Data":"68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb"} Dec 05 07:18:02 crc kubenswrapper[4997]: I1205 07:18:02.571700 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sctbp" event={"ID":"5388a6f5-ecde-42db-9c94-a52c17e8a302","Type":"ContainerStarted","Data":"85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187"} Dec 05 07:18:02 crc kubenswrapper[4997]: I1205 07:18:02.571938 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:18:02 crc kubenswrapper[4997]: I1205 07:18:02.594512 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-sctbp" podStartSLOduration=20.338476746 podStartE2EDuration="55.594488937s" podCreationTimestamp="2025-12-05 07:17:07 +0000 UTC" firstStartedPulling="2025-12-05 07:17:24.917844545 +0000 UTC m=+1345.446751806" lastFinishedPulling="2025-12-05 07:18:00.173856736 +0000 UTC m=+1380.702763997" observedRunningTime="2025-12-05 07:18:02.589464192 +0000 UTC m=+1383.118371473" watchObservedRunningTime="2025-12-05 07:18:02.594488937 +0000 UTC m=+1383.123396198" Dec 05 07:18:03 crc kubenswrapper[4997]: I1205 07:18:03.294146 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:18:03 crc kubenswrapper[4997]: I1205 07:18:03.581640 4997 generic.go:334] "Generic (PLEG): container finished" podID="0601e553-5305-4f54-9a2d-43b43a1a5dd4" containerID="f17f8b4ededafb62d528cf7bfb22c523561aaf354b59cd3fc4a9694a352a9d0b" exitCode=0 Dec 05 07:18:03 crc kubenswrapper[4997]: I1205 07:18:03.581737 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0601e553-5305-4f54-9a2d-43b43a1a5dd4","Type":"ContainerDied","Data":"f17f8b4ededafb62d528cf7bfb22c523561aaf354b59cd3fc4a9694a352a9d0b"} Dec 05 07:18:04 crc kubenswrapper[4997]: I1205 07:18:04.204004 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 05 07:18:04 crc kubenswrapper[4997]: I1205 07:18:04.262654 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 05 07:18:04 crc kubenswrapper[4997]: I1205 07:18:04.262989 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 05 07:18:04 crc kubenswrapper[4997]: I1205 07:18:04.371548 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 05 07:18:04 crc kubenswrapper[4997]: I1205 07:18:04.602717 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0601e553-5305-4f54-9a2d-43b43a1a5dd4","Type":"ContainerStarted","Data":"22c17242a1211bc3c93181daebd1e55ec7c7f3dfb0c90a0942488ba51f04fe30"} Dec 05 07:18:04 crc kubenswrapper[4997]: I1205 07:18:04.639494 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371973.215311 podStartE2EDuration="1m3.6394648s" podCreationTimestamp="2025-12-05 07:17:01 +0000 UTC" firstStartedPulling="2025-12-05 07:17:04.320477188 +0000 UTC m=+1324.849384449" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:18:04.633460509 +0000 UTC m=+1385.162367770" watchObservedRunningTime="2025-12-05 07:18:04.6394648 +0000 UTC m=+1385.168372061" Dec 05 07:18:04 crc kubenswrapper[4997]: I1205 07:18:04.711431 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 05 07:18:05 crc kubenswrapper[4997]: I1205 07:18:05.157007 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:18:05 crc kubenswrapper[4997]: I1205 07:18:05.402830 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:18:05 crc kubenswrapper[4997]: I1205 07:18:05.461599 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-26h6q"] Dec 05 07:18:05 crc kubenswrapper[4997]: I1205 07:18:05.613453 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"833f0bba-91d9-41ab-a412-9e8ae8c38ba2","Type":"ContainerStarted","Data":"a2793e5e578daef37735f5bea301361550f5c913442e8d7f862245cb9db435d5"} Dec 05 07:18:05 crc kubenswrapper[4997]: I1205 07:18:05.613812 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" podUID="4aa8aee7-517a-42f7-9035-0acb7133fa9d" containerName="dnsmasq-dns" containerID="cri-o://698dfc1fd532b847d0d00ef4837d9372b1f7ac959ded5989b9c324c1f8e401ea" gracePeriod=10 Dec 05 07:18:05 crc kubenswrapper[4997]: I1205 07:18:05.614098 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 05 07:18:05 crc kubenswrapper[4997]: I1205 07:18:05.637357 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.187552415 podStartE2EDuration="1m0.637324588s" podCreationTimestamp="2025-12-05 07:17:05 +0000 UTC" firstStartedPulling="2025-12-05 07:17:07.072131357 +0000 UTC m=+1327.601038618" lastFinishedPulling="2025-12-05 07:18:04.52190353 +0000 UTC m=+1385.050810791" observedRunningTime="2025-12-05 07:18:05.635553039 +0000 UTC m=+1386.164460300" watchObservedRunningTime="2025-12-05 07:18:05.637324588 +0000 UTC m=+1386.166231859" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.081839 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.158057 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-config\") pod \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\" (UID: \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\") " Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.158271 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-dns-svc\") pod \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\" (UID: \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\") " Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.158401 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r87rx\" (UniqueName: \"kubernetes.io/projected/4aa8aee7-517a-42f7-9035-0acb7133fa9d-kube-api-access-r87rx\") pod \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\" (UID: \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\") " Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.158484 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-ovsdbserver-nb\") pod \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\" (UID: \"4aa8aee7-517a-42f7-9035-0acb7133fa9d\") " Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.173397 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aa8aee7-517a-42f7-9035-0acb7133fa9d-kube-api-access-r87rx" (OuterVolumeSpecName: "kube-api-access-r87rx") pod "4aa8aee7-517a-42f7-9035-0acb7133fa9d" (UID: "4aa8aee7-517a-42f7-9035-0acb7133fa9d"). InnerVolumeSpecName "kube-api-access-r87rx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.240174 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4aa8aee7-517a-42f7-9035-0acb7133fa9d" (UID: "4aa8aee7-517a-42f7-9035-0acb7133fa9d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.253563 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4aa8aee7-517a-42f7-9035-0acb7133fa9d" (UID: "4aa8aee7-517a-42f7-9035-0acb7133fa9d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.260898 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.260948 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.260960 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r87rx\" (UniqueName: \"kubernetes.io/projected/4aa8aee7-517a-42f7-9035-0acb7133fa9d-kube-api-access-r87rx\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.261650 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-config" (OuterVolumeSpecName: "config") pod "4aa8aee7-517a-42f7-9035-0acb7133fa9d" (UID: "4aa8aee7-517a-42f7-9035-0acb7133fa9d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.362419 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4aa8aee7-517a-42f7-9035-0acb7133fa9d-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.420981 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-bp85j"] Dec 05 07:18:06 crc kubenswrapper[4997]: E1205 07:18:06.421766 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa8aee7-517a-42f7-9035-0acb7133fa9d" containerName="init" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.421789 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa8aee7-517a-42f7-9035-0acb7133fa9d" containerName="init" Dec 05 07:18:06 crc kubenswrapper[4997]: E1205 07:18:06.421811 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa8aee7-517a-42f7-9035-0acb7133fa9d" containerName="dnsmasq-dns" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.421819 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa8aee7-517a-42f7-9035-0acb7133fa9d" containerName="dnsmasq-dns" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.421994 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aa8aee7-517a-42f7-9035-0acb7133fa9d" containerName="dnsmasq-dns" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.422948 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.449985 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-bp85j"] Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.464706 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5fbdd8c-bp85j\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.464780 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shdnn\" (UniqueName: \"kubernetes.io/projected/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-kube-api-access-shdnn\") pod \"dnsmasq-dns-59d5fbdd8c-bp85j\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.465083 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-dns-svc\") pod \"dnsmasq-dns-59d5fbdd8c-bp85j\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.465173 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5fbdd8c-bp85j\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.465220 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-config\") pod \"dnsmasq-dns-59d5fbdd8c-bp85j\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.566766 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5fbdd8c-bp85j\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.566829 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shdnn\" (UniqueName: \"kubernetes.io/projected/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-kube-api-access-shdnn\") pod \"dnsmasq-dns-59d5fbdd8c-bp85j\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.566898 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-dns-svc\") pod \"dnsmasq-dns-59d5fbdd8c-bp85j\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.566926 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5fbdd8c-bp85j\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.566948 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-config\") pod \"dnsmasq-dns-59d5fbdd8c-bp85j\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.568017 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-config\") pod \"dnsmasq-dns-59d5fbdd8c-bp85j\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.568250 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5fbdd8c-bp85j\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.568649 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-dns-svc\") pod \"dnsmasq-dns-59d5fbdd8c-bp85j\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.568977 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5fbdd8c-bp85j\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.600901 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shdnn\" (UniqueName: \"kubernetes.io/projected/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-kube-api-access-shdnn\") pod \"dnsmasq-dns-59d5fbdd8c-bp85j\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.669258 4997 generic.go:334] "Generic (PLEG): container finished" podID="4aa8aee7-517a-42f7-9035-0acb7133fa9d" containerID="698dfc1fd532b847d0d00ef4837d9372b1f7ac959ded5989b9c324c1f8e401ea" exitCode=0 Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.669466 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.670971 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" event={"ID":"4aa8aee7-517a-42f7-9035-0acb7133fa9d","Type":"ContainerDied","Data":"698dfc1fd532b847d0d00ef4837d9372b1f7ac959ded5989b9c324c1f8e401ea"} Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.671014 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57db9b5bc9-26h6q" event={"ID":"4aa8aee7-517a-42f7-9035-0acb7133fa9d","Type":"ContainerDied","Data":"19c9a5c4e5e578d8363aaa2ac09dece879c9dcf4bddb06de624dd4369d5e5800"} Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.671040 4997 scope.go:117] "RemoveContainer" containerID="698dfc1fd532b847d0d00ef4837d9372b1f7ac959ded5989b9c324c1f8e401ea" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.735955 4997 scope.go:117] "RemoveContainer" containerID="442f4e07ce4939690c5144c004816022fd8cb2ab8a48ebad68d80f36953b5a6b" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.756580 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.778599 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-26h6q"] Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.784547 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-26h6q"] Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.970374 4997 scope.go:117] "RemoveContainer" containerID="698dfc1fd532b847d0d00ef4837d9372b1f7ac959ded5989b9c324c1f8e401ea" Dec 05 07:18:06 crc kubenswrapper[4997]: E1205 07:18:06.971639 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"698dfc1fd532b847d0d00ef4837d9372b1f7ac959ded5989b9c324c1f8e401ea\": container with ID starting with 698dfc1fd532b847d0d00ef4837d9372b1f7ac959ded5989b9c324c1f8e401ea not found: ID does not exist" containerID="698dfc1fd532b847d0d00ef4837d9372b1f7ac959ded5989b9c324c1f8e401ea" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.971701 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"698dfc1fd532b847d0d00ef4837d9372b1f7ac959ded5989b9c324c1f8e401ea"} err="failed to get container status \"698dfc1fd532b847d0d00ef4837d9372b1f7ac959ded5989b9c324c1f8e401ea\": rpc error: code = NotFound desc = could not find container \"698dfc1fd532b847d0d00ef4837d9372b1f7ac959ded5989b9c324c1f8e401ea\": container with ID starting with 698dfc1fd532b847d0d00ef4837d9372b1f7ac959ded5989b9c324c1f8e401ea not found: ID does not exist" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.971738 4997 scope.go:117] "RemoveContainer" containerID="442f4e07ce4939690c5144c004816022fd8cb2ab8a48ebad68d80f36953b5a6b" Dec 05 07:18:06 crc kubenswrapper[4997]: E1205 07:18:06.972371 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"442f4e07ce4939690c5144c004816022fd8cb2ab8a48ebad68d80f36953b5a6b\": container with ID starting with 442f4e07ce4939690c5144c004816022fd8cb2ab8a48ebad68d80f36953b5a6b not found: ID does not exist" containerID="442f4e07ce4939690c5144c004816022fd8cb2ab8a48ebad68d80f36953b5a6b" Dec 05 07:18:06 crc kubenswrapper[4997]: I1205 07:18:06.972432 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"442f4e07ce4939690c5144c004816022fd8cb2ab8a48ebad68d80f36953b5a6b"} err="failed to get container status \"442f4e07ce4939690c5144c004816022fd8cb2ab8a48ebad68d80f36953b5a6b\": rpc error: code = NotFound desc = could not find container \"442f4e07ce4939690c5144c004816022fd8cb2ab8a48ebad68d80f36953b5a6b\": container with ID starting with 442f4e07ce4939690c5144c004816022fd8cb2ab8a48ebad68d80f36953b5a6b not found: ID does not exist" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.418830 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-bp85j"] Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.693878 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.732360 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" event={"ID":"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9","Type":"ContainerStarted","Data":"edf8782ebc37c005e6ed6cd2bff03e1d279e7ad9bd4cdcab8c227f055645846f"} Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.732411 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.732569 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.742091 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.742202 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-95zxr" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.742391 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.742476 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.783587 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aa8aee7-517a-42f7-9035-0acb7133fa9d" path="/var/lib/kubelet/pods/4aa8aee7-517a-42f7-9035-0acb7133fa9d/volumes" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.859838 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5cc70c62-ad3e-4178-932a-d14571a06daa-cache\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.860280 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.860345 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.860427 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxmv5\" (UniqueName: \"kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-kube-api-access-hxmv5\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.860453 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5cc70c62-ad3e-4178-932a-d14571a06daa-lock\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.962579 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxmv5\" (UniqueName: \"kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-kube-api-access-hxmv5\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.962659 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5cc70c62-ad3e-4178-932a-d14571a06daa-lock\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.962708 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5cc70c62-ad3e-4178-932a-d14571a06daa-cache\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.962744 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.962791 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:07 crc kubenswrapper[4997]: E1205 07:18:07.962956 4997 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 07:18:07 crc kubenswrapper[4997]: E1205 07:18:07.962973 4997 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 07:18:07 crc kubenswrapper[4997]: E1205 07:18:07.963028 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift podName:5cc70c62-ad3e-4178-932a-d14571a06daa nodeName:}" failed. No retries permitted until 2025-12-05 07:18:08.46300894 +0000 UTC m=+1388.991916201 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift") pod "swift-storage-0" (UID: "5cc70c62-ad3e-4178-932a-d14571a06daa") : configmap "swift-ring-files" not found Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.963411 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5cc70c62-ad3e-4178-932a-d14571a06daa-lock\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.963546 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5cc70c62-ad3e-4178-932a-d14571a06daa-cache\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.963556 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/swift-storage-0" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.989003 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:07 crc kubenswrapper[4997]: I1205 07:18:07.990821 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxmv5\" (UniqueName: \"kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-kube-api-access-hxmv5\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.012762 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-c84rv"] Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.019672 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.022429 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.022513 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.022657 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.054277 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-c84rv"] Dec 05 07:18:08 crc kubenswrapper[4997]: E1205 07:18:08.055338 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-rxxrb ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-rxxrb ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-c84rv" podUID="2be2c1d5-9178-4300-a3d0-b6393b6814aa" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.064091 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-jqnrc"] Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.065371 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.079293 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-jqnrc"] Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.103728 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-c84rv"] Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.166418 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2be2c1d5-9178-4300-a3d0-b6393b6814aa-ring-data-devices\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.166501 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-dispersionconf\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.166564 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-combined-ca-bundle\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.166598 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2be2c1d5-9178-4300-a3d0-b6393b6814aa-etc-swift\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.166658 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/508d4dec-7bf0-4053-a665-b1bf87897981-ring-data-devices\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.166745 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-combined-ca-bundle\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.166787 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2be2c1d5-9178-4300-a3d0-b6393b6814aa-scripts\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.166812 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/508d4dec-7bf0-4053-a665-b1bf87897981-scripts\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.166857 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nvw8\" (UniqueName: \"kubernetes.io/projected/508d4dec-7bf0-4053-a665-b1bf87897981-kube-api-access-4nvw8\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.166897 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-swiftconf\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.166924 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxxrb\" (UniqueName: \"kubernetes.io/projected/2be2c1d5-9178-4300-a3d0-b6393b6814aa-kube-api-access-rxxrb\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.166950 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-swiftconf\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.166979 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/508d4dec-7bf0-4053-a665-b1bf87897981-etc-swift\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.167005 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-dispersionconf\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.268666 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nvw8\" (UniqueName: \"kubernetes.io/projected/508d4dec-7bf0-4053-a665-b1bf87897981-kube-api-access-4nvw8\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.268761 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-swiftconf\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.268790 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxxrb\" (UniqueName: \"kubernetes.io/projected/2be2c1d5-9178-4300-a3d0-b6393b6814aa-kube-api-access-rxxrb\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.268825 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-swiftconf\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.268858 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/508d4dec-7bf0-4053-a665-b1bf87897981-etc-swift\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.268909 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-dispersionconf\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.268956 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2be2c1d5-9178-4300-a3d0-b6393b6814aa-ring-data-devices\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.268987 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-dispersionconf\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.269061 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-combined-ca-bundle\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.269091 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2be2c1d5-9178-4300-a3d0-b6393b6814aa-etc-swift\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.269129 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/508d4dec-7bf0-4053-a665-b1bf87897981-ring-data-devices\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.269198 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-combined-ca-bundle\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.269232 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2be2c1d5-9178-4300-a3d0-b6393b6814aa-scripts\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.269255 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/508d4dec-7bf0-4053-a665-b1bf87897981-scripts\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.270352 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/508d4dec-7bf0-4053-a665-b1bf87897981-scripts\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.271029 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/508d4dec-7bf0-4053-a665-b1bf87897981-etc-swift\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.271718 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/508d4dec-7bf0-4053-a665-b1bf87897981-ring-data-devices\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.273839 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2be2c1d5-9178-4300-a3d0-b6393b6814aa-ring-data-devices\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.274965 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2be2c1d5-9178-4300-a3d0-b6393b6814aa-etc-swift\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.275672 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2be2c1d5-9178-4300-a3d0-b6393b6814aa-scripts\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.276405 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-dispersionconf\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.277305 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-swiftconf\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.277920 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-combined-ca-bundle\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.278526 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-dispersionconf\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.284461 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-combined-ca-bundle\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.301352 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxxrb\" (UniqueName: \"kubernetes.io/projected/2be2c1d5-9178-4300-a3d0-b6393b6814aa-kube-api-access-rxxrb\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.303367 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-swiftconf\") pod \"swift-ring-rebalance-c84rv\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.308458 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nvw8\" (UniqueName: \"kubernetes.io/projected/508d4dec-7bf0-4053-a665-b1bf87897981-kube-api-access-4nvw8\") pod \"swift-ring-rebalance-jqnrc\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.390955 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.472425 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:08 crc kubenswrapper[4997]: E1205 07:18:08.472702 4997 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 07:18:08 crc kubenswrapper[4997]: E1205 07:18:08.472740 4997 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 07:18:08 crc kubenswrapper[4997]: E1205 07:18:08.472845 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift podName:5cc70c62-ad3e-4178-932a-d14571a06daa nodeName:}" failed. No retries permitted until 2025-12-05 07:18:09.472799616 +0000 UTC m=+1390.001706877 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift") pod "swift-storage-0" (UID: "5cc70c62-ad3e-4178-932a-d14571a06daa") : configmap "swift-ring-files" not found Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.717580 4997 generic.go:334] "Generic (PLEG): container finished" podID="b0f3255a-0cbc-4d2a-bd39-526cbf142fa9" containerID="34d43de0b1e528710751b761feea30add9f37a9a1041cf3c2f563ea7891de52a" exitCode=0 Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.718031 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.719772 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" event={"ID":"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9","Type":"ContainerDied","Data":"34d43de0b1e528710751b761feea30add9f37a9a1041cf3c2f563ea7891de52a"} Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.871770 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.982597 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxxrb\" (UniqueName: \"kubernetes.io/projected/2be2c1d5-9178-4300-a3d0-b6393b6814aa-kube-api-access-rxxrb\") pod \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.982959 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-dispersionconf\") pod \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.982984 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-combined-ca-bundle\") pod \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.983030 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2be2c1d5-9178-4300-a3d0-b6393b6814aa-etc-swift\") pod \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.983066 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2be2c1d5-9178-4300-a3d0-b6393b6814aa-scripts\") pod \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.983097 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-swiftconf\") pod \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.983121 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2be2c1d5-9178-4300-a3d0-b6393b6814aa-ring-data-devices\") pod \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\" (UID: \"2be2c1d5-9178-4300-a3d0-b6393b6814aa\") " Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.983636 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2be2c1d5-9178-4300-a3d0-b6393b6814aa-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2be2c1d5-9178-4300-a3d0-b6393b6814aa" (UID: "2be2c1d5-9178-4300-a3d0-b6393b6814aa"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.983898 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2be2c1d5-9178-4300-a3d0-b6393b6814aa-scripts" (OuterVolumeSpecName: "scripts") pod "2be2c1d5-9178-4300-a3d0-b6393b6814aa" (UID: "2be2c1d5-9178-4300-a3d0-b6393b6814aa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.984165 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2be2c1d5-9178-4300-a3d0-b6393b6814aa-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2be2c1d5-9178-4300-a3d0-b6393b6814aa" (UID: "2be2c1d5-9178-4300-a3d0-b6393b6814aa"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.989500 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2be2c1d5-9178-4300-a3d0-b6393b6814aa" (UID: "2be2c1d5-9178-4300-a3d0-b6393b6814aa"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:18:08 crc kubenswrapper[4997]: I1205 07:18:08.990722 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2be2c1d5-9178-4300-a3d0-b6393b6814aa-kube-api-access-rxxrb" (OuterVolumeSpecName: "kube-api-access-rxxrb") pod "2be2c1d5-9178-4300-a3d0-b6393b6814aa" (UID: "2be2c1d5-9178-4300-a3d0-b6393b6814aa"). InnerVolumeSpecName "kube-api-access-rxxrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:18:09 crc kubenswrapper[4997]: I1205 07:18:09.001474 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2be2c1d5-9178-4300-a3d0-b6393b6814aa" (UID: "2be2c1d5-9178-4300-a3d0-b6393b6814aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:18:09 crc kubenswrapper[4997]: I1205 07:18:09.001902 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2be2c1d5-9178-4300-a3d0-b6393b6814aa" (UID: "2be2c1d5-9178-4300-a3d0-b6393b6814aa"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:18:09 crc kubenswrapper[4997]: I1205 07:18:09.051277 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-jqnrc"] Dec 05 07:18:09 crc kubenswrapper[4997]: I1205 07:18:09.085325 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxxrb\" (UniqueName: \"kubernetes.io/projected/2be2c1d5-9178-4300-a3d0-b6393b6814aa-kube-api-access-rxxrb\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:09 crc kubenswrapper[4997]: I1205 07:18:09.085375 4997 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:09 crc kubenswrapper[4997]: I1205 07:18:09.085391 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:09 crc kubenswrapper[4997]: I1205 07:18:09.085403 4997 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2be2c1d5-9178-4300-a3d0-b6393b6814aa-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:09 crc kubenswrapper[4997]: I1205 07:18:09.085413 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2be2c1d5-9178-4300-a3d0-b6393b6814aa-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:09 crc kubenswrapper[4997]: I1205 07:18:09.085424 4997 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2be2c1d5-9178-4300-a3d0-b6393b6814aa-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:09 crc kubenswrapper[4997]: I1205 07:18:09.085437 4997 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2be2c1d5-9178-4300-a3d0-b6393b6814aa-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:09 crc kubenswrapper[4997]: I1205 07:18:09.477201 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:09 crc kubenswrapper[4997]: E1205 07:18:09.477385 4997 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 07:18:09 crc kubenswrapper[4997]: E1205 07:18:09.477414 4997 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 07:18:09 crc kubenswrapper[4997]: E1205 07:18:09.477479 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift podName:5cc70c62-ad3e-4178-932a-d14571a06daa nodeName:}" failed. No retries permitted until 2025-12-05 07:18:11.477459896 +0000 UTC m=+1392.006367157 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift") pod "swift-storage-0" (UID: "5cc70c62-ad3e-4178-932a-d14571a06daa") : configmap "swift-ring-files" not found Dec 05 07:18:09 crc kubenswrapper[4997]: I1205 07:18:09.727382 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-c84rv" Dec 05 07:18:09 crc kubenswrapper[4997]: I1205 07:18:09.734059 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jqnrc" event={"ID":"508d4dec-7bf0-4053-a665-b1bf87897981","Type":"ContainerStarted","Data":"b24f60982d9e47f22e03b07911f02809961869fcf3d88cb2d1f637ccd4d37403"} Dec 05 07:18:09 crc kubenswrapper[4997]: I1205 07:18:09.890709 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-c84rv"] Dec 05 07:18:09 crc kubenswrapper[4997]: I1205 07:18:09.899726 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-c84rv"] Dec 05 07:18:11 crc kubenswrapper[4997]: I1205 07:18:11.494756 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:11 crc kubenswrapper[4997]: E1205 07:18:11.495021 4997 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 07:18:11 crc kubenswrapper[4997]: E1205 07:18:11.495311 4997 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 07:18:11 crc kubenswrapper[4997]: E1205 07:18:11.495398 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift podName:5cc70c62-ad3e-4178-932a-d14571a06daa nodeName:}" failed. No retries permitted until 2025-12-05 07:18:15.495371289 +0000 UTC m=+1396.024278560 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift") pod "swift-storage-0" (UID: "5cc70c62-ad3e-4178-932a-d14571a06daa") : configmap "swift-ring-files" not found Dec 05 07:18:11 crc kubenswrapper[4997]: I1205 07:18:11.765390 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2be2c1d5-9178-4300-a3d0-b6393b6814aa" path="/var/lib/kubelet/pods/2be2c1d5-9178-4300-a3d0-b6393b6814aa/volumes" Dec 05 07:18:11 crc kubenswrapper[4997]: I1205 07:18:11.765922 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:11 crc kubenswrapper[4997]: I1205 07:18:11.765961 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" event={"ID":"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9","Type":"ContainerStarted","Data":"792bc50db82e23371bcb60395d2f639459b0e1762fba29595170fd9dc47ea22e"} Dec 05 07:18:11 crc kubenswrapper[4997]: I1205 07:18:11.788069 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" podStartSLOduration=5.788044461 podStartE2EDuration="5.788044461s" podCreationTimestamp="2025-12-05 07:18:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:18:11.781761532 +0000 UTC m=+1392.310668813" watchObservedRunningTime="2025-12-05 07:18:11.788044461 +0000 UTC m=+1392.316951722" Dec 05 07:18:12 crc kubenswrapper[4997]: I1205 07:18:12.642605 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 05 07:18:13 crc kubenswrapper[4997]: I1205 07:18:13.393303 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 05 07:18:13 crc kubenswrapper[4997]: I1205 07:18:13.393784 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 05 07:18:13 crc kubenswrapper[4997]: I1205 07:18:13.500399 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 05 07:18:13 crc kubenswrapper[4997]: I1205 07:18:13.862448 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.191355 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-94df-account-create-update-gsv4j"] Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.192523 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-94df-account-create-update-gsv4j" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.196007 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.212337 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-94df-account-create-update-gsv4j"] Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.266593 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-6t8lv"] Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.267954 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-6t8lv" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.280494 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-6t8lv"] Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.321206 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fp6m\" (UniqueName: \"kubernetes.io/projected/16042c4a-2200-4f8e-b899-6b75e2b1709b-kube-api-access-9fp6m\") pod \"placement-94df-account-create-update-gsv4j\" (UID: \"16042c4a-2200-4f8e-b899-6b75e2b1709b\") " pod="openstack/placement-94df-account-create-update-gsv4j" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.321329 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16042c4a-2200-4f8e-b899-6b75e2b1709b-operator-scripts\") pod \"placement-94df-account-create-update-gsv4j\" (UID: \"16042c4a-2200-4f8e-b899-6b75e2b1709b\") " pod="openstack/placement-94df-account-create-update-gsv4j" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.406788 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-zf87t"] Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.407914 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zf87t" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.423234 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fp6m\" (UniqueName: \"kubernetes.io/projected/16042c4a-2200-4f8e-b899-6b75e2b1709b-kube-api-access-9fp6m\") pod \"placement-94df-account-create-update-gsv4j\" (UID: \"16042c4a-2200-4f8e-b899-6b75e2b1709b\") " pod="openstack/placement-94df-account-create-update-gsv4j" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.423280 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16042c4a-2200-4f8e-b899-6b75e2b1709b-operator-scripts\") pod \"placement-94df-account-create-update-gsv4j\" (UID: \"16042c4a-2200-4f8e-b899-6b75e2b1709b\") " pod="openstack/placement-94df-account-create-update-gsv4j" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.423369 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxm68\" (UniqueName: \"kubernetes.io/projected/eec54251-6dde-4c1b-a588-3747c8e81fd9-kube-api-access-pxm68\") pod \"placement-db-create-6t8lv\" (UID: \"eec54251-6dde-4c1b-a588-3747c8e81fd9\") " pod="openstack/placement-db-create-6t8lv" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.423401 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eec54251-6dde-4c1b-a588-3747c8e81fd9-operator-scripts\") pod \"placement-db-create-6t8lv\" (UID: \"eec54251-6dde-4c1b-a588-3747c8e81fd9\") " pod="openstack/placement-db-create-6t8lv" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.424396 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16042c4a-2200-4f8e-b899-6b75e2b1709b-operator-scripts\") pod \"placement-94df-account-create-update-gsv4j\" (UID: \"16042c4a-2200-4f8e-b899-6b75e2b1709b\") " pod="openstack/placement-94df-account-create-update-gsv4j" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.439155 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-zf87t"] Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.466915 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fp6m\" (UniqueName: \"kubernetes.io/projected/16042c4a-2200-4f8e-b899-6b75e2b1709b-kube-api-access-9fp6m\") pod \"placement-94df-account-create-update-gsv4j\" (UID: \"16042c4a-2200-4f8e-b899-6b75e2b1709b\") " pod="openstack/placement-94df-account-create-update-gsv4j" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.508555 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-7904-account-create-update-j4c7h"] Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.510194 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7904-account-create-update-j4c7h" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.512311 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.516169 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-7904-account-create-update-j4c7h"] Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.525100 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxm68\" (UniqueName: \"kubernetes.io/projected/eec54251-6dde-4c1b-a588-3747c8e81fd9-kube-api-access-pxm68\") pod \"placement-db-create-6t8lv\" (UID: \"eec54251-6dde-4c1b-a588-3747c8e81fd9\") " pod="openstack/placement-db-create-6t8lv" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.525148 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltcrn\" (UniqueName: \"kubernetes.io/projected/c66780bb-35b5-460f-bf9f-4c30e1a640c5-kube-api-access-ltcrn\") pod \"glance-db-create-zf87t\" (UID: \"c66780bb-35b5-460f-bf9f-4c30e1a640c5\") " pod="openstack/glance-db-create-zf87t" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.525182 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eec54251-6dde-4c1b-a588-3747c8e81fd9-operator-scripts\") pod \"placement-db-create-6t8lv\" (UID: \"eec54251-6dde-4c1b-a588-3747c8e81fd9\") " pod="openstack/placement-db-create-6t8lv" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.525216 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c66780bb-35b5-460f-bf9f-4c30e1a640c5-operator-scripts\") pod \"glance-db-create-zf87t\" (UID: \"c66780bb-35b5-460f-bf9f-4c30e1a640c5\") " pod="openstack/glance-db-create-zf87t" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.526847 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eec54251-6dde-4c1b-a588-3747c8e81fd9-operator-scripts\") pod \"placement-db-create-6t8lv\" (UID: \"eec54251-6dde-4c1b-a588-3747c8e81fd9\") " pod="openstack/placement-db-create-6t8lv" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.531179 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-94df-account-create-update-gsv4j" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.545669 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxm68\" (UniqueName: \"kubernetes.io/projected/eec54251-6dde-4c1b-a588-3747c8e81fd9-kube-api-access-pxm68\") pod \"placement-db-create-6t8lv\" (UID: \"eec54251-6dde-4c1b-a588-3747c8e81fd9\") " pod="openstack/placement-db-create-6t8lv" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.595013 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-6t8lv" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.627663 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgdtr\" (UniqueName: \"kubernetes.io/projected/6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56-kube-api-access-vgdtr\") pod \"glance-7904-account-create-update-j4c7h\" (UID: \"6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56\") " pod="openstack/glance-7904-account-create-update-j4c7h" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.628046 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56-operator-scripts\") pod \"glance-7904-account-create-update-j4c7h\" (UID: \"6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56\") " pod="openstack/glance-7904-account-create-update-j4c7h" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.628181 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltcrn\" (UniqueName: \"kubernetes.io/projected/c66780bb-35b5-460f-bf9f-4c30e1a640c5-kube-api-access-ltcrn\") pod \"glance-db-create-zf87t\" (UID: \"c66780bb-35b5-460f-bf9f-4c30e1a640c5\") " pod="openstack/glance-db-create-zf87t" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.628334 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c66780bb-35b5-460f-bf9f-4c30e1a640c5-operator-scripts\") pod \"glance-db-create-zf87t\" (UID: \"c66780bb-35b5-460f-bf9f-4c30e1a640c5\") " pod="openstack/glance-db-create-zf87t" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.629300 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c66780bb-35b5-460f-bf9f-4c30e1a640c5-operator-scripts\") pod \"glance-db-create-zf87t\" (UID: \"c66780bb-35b5-460f-bf9f-4c30e1a640c5\") " pod="openstack/glance-db-create-zf87t" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.645849 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltcrn\" (UniqueName: \"kubernetes.io/projected/c66780bb-35b5-460f-bf9f-4c30e1a640c5-kube-api-access-ltcrn\") pod \"glance-db-create-zf87t\" (UID: \"c66780bb-35b5-460f-bf9f-4c30e1a640c5\") " pod="openstack/glance-db-create-zf87t" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.729753 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgdtr\" (UniqueName: \"kubernetes.io/projected/6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56-kube-api-access-vgdtr\") pod \"glance-7904-account-create-update-j4c7h\" (UID: \"6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56\") " pod="openstack/glance-7904-account-create-update-j4c7h" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.729820 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56-operator-scripts\") pod \"glance-7904-account-create-update-j4c7h\" (UID: \"6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56\") " pod="openstack/glance-7904-account-create-update-j4c7h" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.730631 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56-operator-scripts\") pod \"glance-7904-account-create-update-j4c7h\" (UID: \"6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56\") " pod="openstack/glance-7904-account-create-update-j4c7h" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.732163 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zf87t" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.748955 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgdtr\" (UniqueName: \"kubernetes.io/projected/6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56-kube-api-access-vgdtr\") pod \"glance-7904-account-create-update-j4c7h\" (UID: \"6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56\") " pod="openstack/glance-7904-account-create-update-j4c7h" Dec 05 07:18:14 crc kubenswrapper[4997]: I1205 07:18:14.839676 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7904-account-create-update-j4c7h" Dec 05 07:18:15 crc kubenswrapper[4997]: I1205 07:18:15.547695 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:15 crc kubenswrapper[4997]: E1205 07:18:15.547922 4997 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 07:18:15 crc kubenswrapper[4997]: E1205 07:18:15.547947 4997 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 07:18:15 crc kubenswrapper[4997]: E1205 07:18:15.548012 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift podName:5cc70c62-ad3e-4178-932a-d14571a06daa nodeName:}" failed. No retries permitted until 2025-12-05 07:18:23.547992298 +0000 UTC m=+1404.076899559 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift") pod "swift-storage-0" (UID: "5cc70c62-ad3e-4178-932a-d14571a06daa") : configmap "swift-ring-files" not found Dec 05 07:18:16 crc kubenswrapper[4997]: I1205 07:18:16.235834 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 05 07:18:16 crc kubenswrapper[4997]: I1205 07:18:16.763287 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:18:16 crc kubenswrapper[4997]: I1205 07:18:16.836754 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-ht85n"] Dec 05 07:18:16 crc kubenswrapper[4997]: I1205 07:18:16.837020 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-db7757ddc-ht85n" podUID="a78999af-0d52-4d34-9bf1-dda9848049bd" containerName="dnsmasq-dns" containerID="cri-o://01368c32dffe061c73bd12e0225f1b9c8acd080ec609973a06f3aefb667f648e" gracePeriod=10 Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.018407 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-zf87t"] Dec 05 07:18:17 crc kubenswrapper[4997]: W1205 07:18:17.031209 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc66780bb_35b5_460f_bf9f_4c30e1a640c5.slice/crio-ba18ebfe2fb057fc995ab8d29aab4499a49d2d8bbb3ff2a73e466dc4d944049a WatchSource:0}: Error finding container ba18ebfe2fb057fc995ab8d29aab4499a49d2d8bbb3ff2a73e466dc4d944049a: Status 404 returned error can't find the container with id ba18ebfe2fb057fc995ab8d29aab4499a49d2d8bbb3ff2a73e466dc4d944049a Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.308769 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-7904-account-create-update-j4c7h"] Dec 05 07:18:17 crc kubenswrapper[4997]: W1205 07:18:17.332757 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f5dfa02_1cb3_47ca_a8a3_53ab7980ac56.slice/crio-38d506fd0662eedb1184cce5f5801d342297e05b26bccf38fcbf36cf8a411dda WatchSource:0}: Error finding container 38d506fd0662eedb1184cce5f5801d342297e05b26bccf38fcbf36cf8a411dda: Status 404 returned error can't find the container with id 38d506fd0662eedb1184cce5f5801d342297e05b26bccf38fcbf36cf8a411dda Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.382855 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-6t8lv"] Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.408365 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-94df-account-create-update-gsv4j"] Dec 05 07:18:17 crc kubenswrapper[4997]: W1205 07:18:17.428347 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16042c4a_2200_4f8e_b899_6b75e2b1709b.slice/crio-6313b7b984f1d75bf0926e33a9d3563872ab91d9f3f87160b723a75d3c079db4 WatchSource:0}: Error finding container 6313b7b984f1d75bf0926e33a9d3563872ab91d9f3f87160b723a75d3c079db4: Status 404 returned error can't find the container with id 6313b7b984f1d75bf0926e33a9d3563872ab91d9f3f87160b723a75d3c079db4 Dec 05 07:18:17 crc kubenswrapper[4997]: W1205 07:18:17.430914 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeec54251_6dde_4c1b_a588_3747c8e81fd9.slice/crio-dc695b47e5abf676e145f8a356e61db6497786ec0b4319b8c936ad6f13d40743 WatchSource:0}: Error finding container dc695b47e5abf676e145f8a356e61db6497786ec0b4319b8c936ad6f13d40743: Status 404 returned error can't find the container with id dc695b47e5abf676e145f8a356e61db6497786ec0b4319b8c936ad6f13d40743 Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.597251 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.612369 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-dns-svc\") pod \"a78999af-0d52-4d34-9bf1-dda9848049bd\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.612532 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-ovsdbserver-sb\") pod \"a78999af-0d52-4d34-9bf1-dda9848049bd\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.612586 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gtj5\" (UniqueName: \"kubernetes.io/projected/a78999af-0d52-4d34-9bf1-dda9848049bd-kube-api-access-7gtj5\") pod \"a78999af-0d52-4d34-9bf1-dda9848049bd\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.612704 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-config\") pod \"a78999af-0d52-4d34-9bf1-dda9848049bd\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.612847 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-ovsdbserver-nb\") pod \"a78999af-0d52-4d34-9bf1-dda9848049bd\" (UID: \"a78999af-0d52-4d34-9bf1-dda9848049bd\") " Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.649861 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a78999af-0d52-4d34-9bf1-dda9848049bd-kube-api-access-7gtj5" (OuterVolumeSpecName: "kube-api-access-7gtj5") pod "a78999af-0d52-4d34-9bf1-dda9848049bd" (UID: "a78999af-0d52-4d34-9bf1-dda9848049bd"). InnerVolumeSpecName "kube-api-access-7gtj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.706762 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a78999af-0d52-4d34-9bf1-dda9848049bd" (UID: "a78999af-0d52-4d34-9bf1-dda9848049bd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.723762 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gtj5\" (UniqueName: \"kubernetes.io/projected/a78999af-0d52-4d34-9bf1-dda9848049bd-kube-api-access-7gtj5\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.723806 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.737983 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-config" (OuterVolumeSpecName: "config") pod "a78999af-0d52-4d34-9bf1-dda9848049bd" (UID: "a78999af-0d52-4d34-9bf1-dda9848049bd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.770384 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a78999af-0d52-4d34-9bf1-dda9848049bd" (UID: "a78999af-0d52-4d34-9bf1-dda9848049bd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.770633 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a78999af-0d52-4d34-9bf1-dda9848049bd" (UID: "a78999af-0d52-4d34-9bf1-dda9848049bd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.829273 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.829307 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.829319 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a78999af-0d52-4d34-9bf1-dda9848049bd-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.844886 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-6t8lv" event={"ID":"eec54251-6dde-4c1b-a588-3747c8e81fd9","Type":"ContainerStarted","Data":"abdd9ffacada164260fbd3ad6870541d93f6c0272c1c8ff72270c8ed33bcbe2d"} Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.845002 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-6t8lv" event={"ID":"eec54251-6dde-4c1b-a588-3747c8e81fd9","Type":"ContainerStarted","Data":"dc695b47e5abf676e145f8a356e61db6497786ec0b4319b8c936ad6f13d40743"} Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.857412 4997 generic.go:334] "Generic (PLEG): container finished" podID="a78999af-0d52-4d34-9bf1-dda9848049bd" containerID="01368c32dffe061c73bd12e0225f1b9c8acd080ec609973a06f3aefb667f648e" exitCode=0 Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.857499 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-ht85n" event={"ID":"a78999af-0d52-4d34-9bf1-dda9848049bd","Type":"ContainerDied","Data":"01368c32dffe061c73bd12e0225f1b9c8acd080ec609973a06f3aefb667f648e"} Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.857541 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-ht85n" event={"ID":"a78999af-0d52-4d34-9bf1-dda9848049bd","Type":"ContainerDied","Data":"02ef4a19f1e03a85ad39ae42d31d2206ccbd037bb882f7e68f88a0f0391ea6d2"} Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.857588 4997 scope.go:117] "RemoveContainer" containerID="01368c32dffe061c73bd12e0225f1b9c8acd080ec609973a06f3aefb667f648e" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.857657 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-ht85n" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.868578 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-7904-account-create-update-j4c7h" event={"ID":"6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56","Type":"ContainerStarted","Data":"4d6b87afcee18a48538358bac1dcd98772c8ae50d88aed51036e1d032a49b121"} Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.868658 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-7904-account-create-update-j4c7h" event={"ID":"6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56","Type":"ContainerStarted","Data":"38d506fd0662eedb1184cce5f5801d342297e05b26bccf38fcbf36cf8a411dda"} Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.871563 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jqnrc" event={"ID":"508d4dec-7bf0-4053-a665-b1bf87897981","Type":"ContainerStarted","Data":"5f739def7163631384233edc764a3f8a516103dbcdc15f0624570994d94c96c8"} Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.876762 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-94df-account-create-update-gsv4j" event={"ID":"16042c4a-2200-4f8e-b899-6b75e2b1709b","Type":"ContainerStarted","Data":"6313b7b984f1d75bf0926e33a9d3563872ab91d9f3f87160b723a75d3c079db4"} Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.876961 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-6t8lv" podStartSLOduration=3.8769363869999998 podStartE2EDuration="3.876936387s" podCreationTimestamp="2025-12-05 07:18:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:18:17.866176426 +0000 UTC m=+1398.395083727" watchObservedRunningTime="2025-12-05 07:18:17.876936387 +0000 UTC m=+1398.405843638" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.882167 4997 generic.go:334] "Generic (PLEG): container finished" podID="c66780bb-35b5-460f-bf9f-4c30e1a640c5" containerID="d02644e37a2cd2654ffa8c46e55402e1791c721b12088c2864e45e521fffdba0" exitCode=0 Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.882208 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zf87t" event={"ID":"c66780bb-35b5-460f-bf9f-4c30e1a640c5","Type":"ContainerDied","Data":"d02644e37a2cd2654ffa8c46e55402e1791c721b12088c2864e45e521fffdba0"} Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.882276 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zf87t" event={"ID":"c66780bb-35b5-460f-bf9f-4c30e1a640c5","Type":"ContainerStarted","Data":"ba18ebfe2fb057fc995ab8d29aab4499a49d2d8bbb3ff2a73e466dc4d944049a"} Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.902282 4997 scope.go:117] "RemoveContainer" containerID="a5997c2d9abe35118ff5784153570b011a42b33128ab4e9bd034de357280ffa5" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.905562 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-7904-account-create-update-j4c7h" podStartSLOduration=3.905511878 podStartE2EDuration="3.905511878s" podCreationTimestamp="2025-12-05 07:18:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:18:17.902812065 +0000 UTC m=+1398.431719326" watchObservedRunningTime="2025-12-05 07:18:17.905511878 +0000 UTC m=+1398.434419139" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.951709 4997 scope.go:117] "RemoveContainer" containerID="01368c32dffe061c73bd12e0225f1b9c8acd080ec609973a06f3aefb667f648e" Dec 05 07:18:17 crc kubenswrapper[4997]: E1205 07:18:17.952422 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01368c32dffe061c73bd12e0225f1b9c8acd080ec609973a06f3aefb667f648e\": container with ID starting with 01368c32dffe061c73bd12e0225f1b9c8acd080ec609973a06f3aefb667f648e not found: ID does not exist" containerID="01368c32dffe061c73bd12e0225f1b9c8acd080ec609973a06f3aefb667f648e" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.952480 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01368c32dffe061c73bd12e0225f1b9c8acd080ec609973a06f3aefb667f648e"} err="failed to get container status \"01368c32dffe061c73bd12e0225f1b9c8acd080ec609973a06f3aefb667f648e\": rpc error: code = NotFound desc = could not find container \"01368c32dffe061c73bd12e0225f1b9c8acd080ec609973a06f3aefb667f648e\": container with ID starting with 01368c32dffe061c73bd12e0225f1b9c8acd080ec609973a06f3aefb667f648e not found: ID does not exist" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.952508 4997 scope.go:117] "RemoveContainer" containerID="a5997c2d9abe35118ff5784153570b011a42b33128ab4e9bd034de357280ffa5" Dec 05 07:18:17 crc kubenswrapper[4997]: E1205 07:18:17.953046 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5997c2d9abe35118ff5784153570b011a42b33128ab4e9bd034de357280ffa5\": container with ID starting with a5997c2d9abe35118ff5784153570b011a42b33128ab4e9bd034de357280ffa5 not found: ID does not exist" containerID="a5997c2d9abe35118ff5784153570b011a42b33128ab4e9bd034de357280ffa5" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.953078 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5997c2d9abe35118ff5784153570b011a42b33128ab4e9bd034de357280ffa5"} err="failed to get container status \"a5997c2d9abe35118ff5784153570b011a42b33128ab4e9bd034de357280ffa5\": rpc error: code = NotFound desc = could not find container \"a5997c2d9abe35118ff5784153570b011a42b33128ab4e9bd034de357280ffa5\": container with ID starting with a5997c2d9abe35118ff5784153570b011a42b33128ab4e9bd034de357280ffa5 not found: ID does not exist" Dec 05 07:18:17 crc kubenswrapper[4997]: I1205 07:18:17.982957 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-jqnrc" podStartSLOduration=2.554857408 podStartE2EDuration="9.982929705s" podCreationTimestamp="2025-12-05 07:18:08 +0000 UTC" firstStartedPulling="2025-12-05 07:18:09.055904189 +0000 UTC m=+1389.584811450" lastFinishedPulling="2025-12-05 07:18:16.483976486 +0000 UTC m=+1397.012883747" observedRunningTime="2025-12-05 07:18:17.929924446 +0000 UTC m=+1398.458831717" watchObservedRunningTime="2025-12-05 07:18:17.982929705 +0000 UTC m=+1398.511836966" Dec 05 07:18:18 crc kubenswrapper[4997]: I1205 07:18:18.002136 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-94df-account-create-update-gsv4j" podStartSLOduration=4.002110683 podStartE2EDuration="4.002110683s" podCreationTimestamp="2025-12-05 07:18:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:18:17.949479243 +0000 UTC m=+1398.478386524" watchObservedRunningTime="2025-12-05 07:18:18.002110683 +0000 UTC m=+1398.531017954" Dec 05 07:18:18 crc kubenswrapper[4997]: I1205 07:18:18.034060 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-ht85n"] Dec 05 07:18:18 crc kubenswrapper[4997]: I1205 07:18:18.043590 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-ht85n"] Dec 05 07:18:18 crc kubenswrapper[4997]: I1205 07:18:18.893453 4997 generic.go:334] "Generic (PLEG): container finished" podID="16042c4a-2200-4f8e-b899-6b75e2b1709b" containerID="2c9db8fa53e1305d4cb7a6121c6b3932ba2fb53b513d58ec4d556042d1e925ab" exitCode=0 Dec 05 07:18:18 crc kubenswrapper[4997]: I1205 07:18:18.893530 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-94df-account-create-update-gsv4j" event={"ID":"16042c4a-2200-4f8e-b899-6b75e2b1709b","Type":"ContainerDied","Data":"2c9db8fa53e1305d4cb7a6121c6b3932ba2fb53b513d58ec4d556042d1e925ab"} Dec 05 07:18:18 crc kubenswrapper[4997]: I1205 07:18:18.895823 4997 generic.go:334] "Generic (PLEG): container finished" podID="eec54251-6dde-4c1b-a588-3747c8e81fd9" containerID="abdd9ffacada164260fbd3ad6870541d93f6c0272c1c8ff72270c8ed33bcbe2d" exitCode=0 Dec 05 07:18:18 crc kubenswrapper[4997]: I1205 07:18:18.895998 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-6t8lv" event={"ID":"eec54251-6dde-4c1b-a588-3747c8e81fd9","Type":"ContainerDied","Data":"abdd9ffacada164260fbd3ad6870541d93f6c0272c1c8ff72270c8ed33bcbe2d"} Dec 05 07:18:18 crc kubenswrapper[4997]: I1205 07:18:18.900381 4997 generic.go:334] "Generic (PLEG): container finished" podID="6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56" containerID="4d6b87afcee18a48538358bac1dcd98772c8ae50d88aed51036e1d032a49b121" exitCode=0 Dec 05 07:18:18 crc kubenswrapper[4997]: I1205 07:18:18.900449 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-7904-account-create-update-j4c7h" event={"ID":"6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56","Type":"ContainerDied","Data":"4d6b87afcee18a48538358bac1dcd98772c8ae50d88aed51036e1d032a49b121"} Dec 05 07:18:19 crc kubenswrapper[4997]: I1205 07:18:19.267053 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zf87t" Dec 05 07:18:19 crc kubenswrapper[4997]: I1205 07:18:19.464369 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c66780bb-35b5-460f-bf9f-4c30e1a640c5-operator-scripts\") pod \"c66780bb-35b5-460f-bf9f-4c30e1a640c5\" (UID: \"c66780bb-35b5-460f-bf9f-4c30e1a640c5\") " Dec 05 07:18:19 crc kubenswrapper[4997]: I1205 07:18:19.464485 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltcrn\" (UniqueName: \"kubernetes.io/projected/c66780bb-35b5-460f-bf9f-4c30e1a640c5-kube-api-access-ltcrn\") pod \"c66780bb-35b5-460f-bf9f-4c30e1a640c5\" (UID: \"c66780bb-35b5-460f-bf9f-4c30e1a640c5\") " Dec 05 07:18:19 crc kubenswrapper[4997]: I1205 07:18:19.464964 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c66780bb-35b5-460f-bf9f-4c30e1a640c5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c66780bb-35b5-460f-bf9f-4c30e1a640c5" (UID: "c66780bb-35b5-460f-bf9f-4c30e1a640c5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:19 crc kubenswrapper[4997]: I1205 07:18:19.470441 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c66780bb-35b5-460f-bf9f-4c30e1a640c5-kube-api-access-ltcrn" (OuterVolumeSpecName: "kube-api-access-ltcrn") pod "c66780bb-35b5-460f-bf9f-4c30e1a640c5" (UID: "c66780bb-35b5-460f-bf9f-4c30e1a640c5"). InnerVolumeSpecName "kube-api-access-ltcrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:18:19 crc kubenswrapper[4997]: I1205 07:18:19.566799 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c66780bb-35b5-460f-bf9f-4c30e1a640c5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:19 crc kubenswrapper[4997]: I1205 07:18:19.566857 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltcrn\" (UniqueName: \"kubernetes.io/projected/c66780bb-35b5-460f-bf9f-4c30e1a640c5-kube-api-access-ltcrn\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:19 crc kubenswrapper[4997]: I1205 07:18:19.762578 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a78999af-0d52-4d34-9bf1-dda9848049bd" path="/var/lib/kubelet/pods/a78999af-0d52-4d34-9bf1-dda9848049bd/volumes" Dec 05 07:18:19 crc kubenswrapper[4997]: I1205 07:18:19.769814 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:18:19 crc kubenswrapper[4997]: I1205 07:18:19.769888 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:18:19 crc kubenswrapper[4997]: I1205 07:18:19.910972 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zf87t" Dec 05 07:18:19 crc kubenswrapper[4997]: I1205 07:18:19.911403 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zf87t" event={"ID":"c66780bb-35b5-460f-bf9f-4c30e1a640c5","Type":"ContainerDied","Data":"ba18ebfe2fb057fc995ab8d29aab4499a49d2d8bbb3ff2a73e466dc4d944049a"} Dec 05 07:18:19 crc kubenswrapper[4997]: I1205 07:18:19.911597 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba18ebfe2fb057fc995ab8d29aab4499a49d2d8bbb3ff2a73e466dc4d944049a" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.324146 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7904-account-create-update-j4c7h" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.440132 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-6t8lv" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.445754 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-94df-account-create-update-gsv4j" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.482595 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgdtr\" (UniqueName: \"kubernetes.io/projected/6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56-kube-api-access-vgdtr\") pod \"6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56\" (UID: \"6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56\") " Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.482733 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56-operator-scripts\") pod \"6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56\" (UID: \"6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56\") " Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.484001 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56" (UID: "6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.488761 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56-kube-api-access-vgdtr" (OuterVolumeSpecName: "kube-api-access-vgdtr") pod "6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56" (UID: "6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56"). InnerVolumeSpecName "kube-api-access-vgdtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.584740 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxm68\" (UniqueName: \"kubernetes.io/projected/eec54251-6dde-4c1b-a588-3747c8e81fd9-kube-api-access-pxm68\") pod \"eec54251-6dde-4c1b-a588-3747c8e81fd9\" (UID: \"eec54251-6dde-4c1b-a588-3747c8e81fd9\") " Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.584828 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16042c4a-2200-4f8e-b899-6b75e2b1709b-operator-scripts\") pod \"16042c4a-2200-4f8e-b899-6b75e2b1709b\" (UID: \"16042c4a-2200-4f8e-b899-6b75e2b1709b\") " Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.584873 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eec54251-6dde-4c1b-a588-3747c8e81fd9-operator-scripts\") pod \"eec54251-6dde-4c1b-a588-3747c8e81fd9\" (UID: \"eec54251-6dde-4c1b-a588-3747c8e81fd9\") " Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.584924 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fp6m\" (UniqueName: \"kubernetes.io/projected/16042c4a-2200-4f8e-b899-6b75e2b1709b-kube-api-access-9fp6m\") pod \"16042c4a-2200-4f8e-b899-6b75e2b1709b\" (UID: \"16042c4a-2200-4f8e-b899-6b75e2b1709b\") " Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.585470 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eec54251-6dde-4c1b-a588-3747c8e81fd9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eec54251-6dde-4c1b-a588-3747c8e81fd9" (UID: "eec54251-6dde-4c1b-a588-3747c8e81fd9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.586049 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eec54251-6dde-4c1b-a588-3747c8e81fd9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.586078 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgdtr\" (UniqueName: \"kubernetes.io/projected/6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56-kube-api-access-vgdtr\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.586095 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.586130 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16042c4a-2200-4f8e-b899-6b75e2b1709b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "16042c4a-2200-4f8e-b899-6b75e2b1709b" (UID: "16042c4a-2200-4f8e-b899-6b75e2b1709b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.601522 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eec54251-6dde-4c1b-a588-3747c8e81fd9-kube-api-access-pxm68" (OuterVolumeSpecName: "kube-api-access-pxm68") pod "eec54251-6dde-4c1b-a588-3747c8e81fd9" (UID: "eec54251-6dde-4c1b-a588-3747c8e81fd9"). InnerVolumeSpecName "kube-api-access-pxm68". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.601579 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16042c4a-2200-4f8e-b899-6b75e2b1709b-kube-api-access-9fp6m" (OuterVolumeSpecName: "kube-api-access-9fp6m") pod "16042c4a-2200-4f8e-b899-6b75e2b1709b" (UID: "16042c4a-2200-4f8e-b899-6b75e2b1709b"). InnerVolumeSpecName "kube-api-access-9fp6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.687323 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxm68\" (UniqueName: \"kubernetes.io/projected/eec54251-6dde-4c1b-a588-3747c8e81fd9-kube-api-access-pxm68\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.687348 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16042c4a-2200-4f8e-b899-6b75e2b1709b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.687358 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fp6m\" (UniqueName: \"kubernetes.io/projected/16042c4a-2200-4f8e-b899-6b75e2b1709b-kube-api-access-9fp6m\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.921695 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-94df-account-create-update-gsv4j" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.921686 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-94df-account-create-update-gsv4j" event={"ID":"16042c4a-2200-4f8e-b899-6b75e2b1709b","Type":"ContainerDied","Data":"6313b7b984f1d75bf0926e33a9d3563872ab91d9f3f87160b723a75d3c079db4"} Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.922282 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6313b7b984f1d75bf0926e33a9d3563872ab91d9f3f87160b723a75d3c079db4" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.924956 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-6t8lv" event={"ID":"eec54251-6dde-4c1b-a588-3747c8e81fd9","Type":"ContainerDied","Data":"dc695b47e5abf676e145f8a356e61db6497786ec0b4319b8c936ad6f13d40743"} Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.925019 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc695b47e5abf676e145f8a356e61db6497786ec0b4319b8c936ad6f13d40743" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.925146 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-6t8lv" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.934328 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-7904-account-create-update-j4c7h" event={"ID":"6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56","Type":"ContainerDied","Data":"38d506fd0662eedb1184cce5f5801d342297e05b26bccf38fcbf36cf8a411dda"} Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.934436 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38d506fd0662eedb1184cce5f5801d342297e05b26bccf38fcbf36cf8a411dda" Dec 05 07:18:20 crc kubenswrapper[4997]: I1205 07:18:20.934646 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7904-account-create-update-j4c7h" Dec 05 07:18:21 crc kubenswrapper[4997]: E1205 07:18:21.080362 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16042c4a_2200_4f8e_b899_6b75e2b1709b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeec54251_6dde_4c1b_a588_3747c8e81fd9.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f5dfa02_1cb3_47ca_a8a3_53ab7980ac56.slice\": RecentStats: unable to find data in memory cache]" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.334981 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vnncj" podUID="5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" containerName="ovn-controller" probeResult="failure" output=< Dec 05 07:18:23 crc kubenswrapper[4997]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 05 07:18:23 crc kubenswrapper[4997]: > Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.640516 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:23 crc kubenswrapper[4997]: E1205 07:18:23.640778 4997 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 07:18:23 crc kubenswrapper[4997]: E1205 07:18:23.640826 4997 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 07:18:23 crc kubenswrapper[4997]: E1205 07:18:23.640920 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift podName:5cc70c62-ad3e-4178-932a-d14571a06daa nodeName:}" failed. No retries permitted until 2025-12-05 07:18:39.640887421 +0000 UTC m=+1420.169794692 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift") pod "swift-storage-0" (UID: "5cc70c62-ad3e-4178-932a-d14571a06daa") : configmap "swift-ring-files" not found Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.734001 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-ktw8k"] Dec 05 07:18:23 crc kubenswrapper[4997]: E1205 07:18:23.734803 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec54251-6dde-4c1b-a588-3747c8e81fd9" containerName="mariadb-database-create" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.734904 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec54251-6dde-4c1b-a588-3747c8e81fd9" containerName="mariadb-database-create" Dec 05 07:18:23 crc kubenswrapper[4997]: E1205 07:18:23.735001 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78999af-0d52-4d34-9bf1-dda9848049bd" containerName="init" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.735121 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78999af-0d52-4d34-9bf1-dda9848049bd" containerName="init" Dec 05 07:18:23 crc kubenswrapper[4997]: E1205 07:18:23.735229 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78999af-0d52-4d34-9bf1-dda9848049bd" containerName="dnsmasq-dns" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.735309 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78999af-0d52-4d34-9bf1-dda9848049bd" containerName="dnsmasq-dns" Dec 05 07:18:23 crc kubenswrapper[4997]: E1205 07:18:23.735402 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16042c4a-2200-4f8e-b899-6b75e2b1709b" containerName="mariadb-account-create-update" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.735467 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="16042c4a-2200-4f8e-b899-6b75e2b1709b" containerName="mariadb-account-create-update" Dec 05 07:18:23 crc kubenswrapper[4997]: E1205 07:18:23.735524 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c66780bb-35b5-460f-bf9f-4c30e1a640c5" containerName="mariadb-database-create" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.735575 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c66780bb-35b5-460f-bf9f-4c30e1a640c5" containerName="mariadb-database-create" Dec 05 07:18:23 crc kubenswrapper[4997]: E1205 07:18:23.735669 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56" containerName="mariadb-account-create-update" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.735734 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56" containerName="mariadb-account-create-update" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.735996 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="16042c4a-2200-4f8e-b899-6b75e2b1709b" containerName="mariadb-account-create-update" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.736103 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="eec54251-6dde-4c1b-a588-3747c8e81fd9" containerName="mariadb-database-create" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.736181 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56" containerName="mariadb-account-create-update" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.736275 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a78999af-0d52-4d34-9bf1-dda9848049bd" containerName="dnsmasq-dns" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.736367 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="c66780bb-35b5-460f-bf9f-4c30e1a640c5" containerName="mariadb-database-create" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.737267 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ktw8k" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.742583 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5hd4\" (UniqueName: \"kubernetes.io/projected/77cf9c93-8871-4190-8522-3cb69ede54dd-kube-api-access-t5hd4\") pod \"keystone-db-create-ktw8k\" (UID: \"77cf9c93-8871-4190-8522-3cb69ede54dd\") " pod="openstack/keystone-db-create-ktw8k" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.742852 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77cf9c93-8871-4190-8522-3cb69ede54dd-operator-scripts\") pod \"keystone-db-create-ktw8k\" (UID: \"77cf9c93-8871-4190-8522-3cb69ede54dd\") " pod="openstack/keystone-db-create-ktw8k" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.772160 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ktw8k"] Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.829861 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-3eec-account-create-update-tckr2"] Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.831414 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3eec-account-create-update-tckr2" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.834938 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.838265 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3eec-account-create-update-tckr2"] Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.857967 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5hd4\" (UniqueName: \"kubernetes.io/projected/77cf9c93-8871-4190-8522-3cb69ede54dd-kube-api-access-t5hd4\") pod \"keystone-db-create-ktw8k\" (UID: \"77cf9c93-8871-4190-8522-3cb69ede54dd\") " pod="openstack/keystone-db-create-ktw8k" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.858514 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6l5x\" (UniqueName: \"kubernetes.io/projected/377ed320-c385-42a5-852e-31360c519c6d-kube-api-access-m6l5x\") pod \"keystone-3eec-account-create-update-tckr2\" (UID: \"377ed320-c385-42a5-852e-31360c519c6d\") " pod="openstack/keystone-3eec-account-create-update-tckr2" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.858734 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77cf9c93-8871-4190-8522-3cb69ede54dd-operator-scripts\") pod \"keystone-db-create-ktw8k\" (UID: \"77cf9c93-8871-4190-8522-3cb69ede54dd\") " pod="openstack/keystone-db-create-ktw8k" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.858848 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/377ed320-c385-42a5-852e-31360c519c6d-operator-scripts\") pod \"keystone-3eec-account-create-update-tckr2\" (UID: \"377ed320-c385-42a5-852e-31360c519c6d\") " pod="openstack/keystone-3eec-account-create-update-tckr2" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.859959 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77cf9c93-8871-4190-8522-3cb69ede54dd-operator-scripts\") pod \"keystone-db-create-ktw8k\" (UID: \"77cf9c93-8871-4190-8522-3cb69ede54dd\") " pod="openstack/keystone-db-create-ktw8k" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.901504 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5hd4\" (UniqueName: \"kubernetes.io/projected/77cf9c93-8871-4190-8522-3cb69ede54dd-kube-api-access-t5hd4\") pod \"keystone-db-create-ktw8k\" (UID: \"77cf9c93-8871-4190-8522-3cb69ede54dd\") " pod="openstack/keystone-db-create-ktw8k" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.960777 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/377ed320-c385-42a5-852e-31360c519c6d-operator-scripts\") pod \"keystone-3eec-account-create-update-tckr2\" (UID: \"377ed320-c385-42a5-852e-31360c519c6d\") " pod="openstack/keystone-3eec-account-create-update-tckr2" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.961274 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6l5x\" (UniqueName: \"kubernetes.io/projected/377ed320-c385-42a5-852e-31360c519c6d-kube-api-access-m6l5x\") pod \"keystone-3eec-account-create-update-tckr2\" (UID: \"377ed320-c385-42a5-852e-31360c519c6d\") " pod="openstack/keystone-3eec-account-create-update-tckr2" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.962210 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/377ed320-c385-42a5-852e-31360c519c6d-operator-scripts\") pod \"keystone-3eec-account-create-update-tckr2\" (UID: \"377ed320-c385-42a5-852e-31360c519c6d\") " pod="openstack/keystone-3eec-account-create-update-tckr2" Dec 05 07:18:23 crc kubenswrapper[4997]: I1205 07:18:23.984375 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6l5x\" (UniqueName: \"kubernetes.io/projected/377ed320-c385-42a5-852e-31360c519c6d-kube-api-access-m6l5x\") pod \"keystone-3eec-account-create-update-tckr2\" (UID: \"377ed320-c385-42a5-852e-31360c519c6d\") " pod="openstack/keystone-3eec-account-create-update-tckr2" Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.069738 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ktw8k" Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.150139 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3eec-account-create-update-tckr2" Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.417878 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ktw8k"] Dec 05 07:18:24 crc kubenswrapper[4997]: W1205 07:18:24.423917 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77cf9c93_8871_4190_8522_3cb69ede54dd.slice/crio-4537589c112cdfd478301e78b4a7c4140dcf39ada77e1b1027dfc49cd4ae32ab WatchSource:0}: Error finding container 4537589c112cdfd478301e78b4a7c4140dcf39ada77e1b1027dfc49cd4ae32ab: Status 404 returned error can't find the container with id 4537589c112cdfd478301e78b4a7c4140dcf39ada77e1b1027dfc49cd4ae32ab Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.474460 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3eec-account-create-update-tckr2"] Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.756900 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-6bdvj"] Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.758492 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6bdvj" Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.761258 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qsb8n" Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.762278 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.767102 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-6bdvj"] Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.884427 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-db-sync-config-data\") pod \"glance-db-sync-6bdvj\" (UID: \"cdf7e311-569b-43a2-bba8-729226e28415\") " pod="openstack/glance-db-sync-6bdvj" Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.884544 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrrqz\" (UniqueName: \"kubernetes.io/projected/cdf7e311-569b-43a2-bba8-729226e28415-kube-api-access-nrrqz\") pod \"glance-db-sync-6bdvj\" (UID: \"cdf7e311-569b-43a2-bba8-729226e28415\") " pod="openstack/glance-db-sync-6bdvj" Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.884843 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-combined-ca-bundle\") pod \"glance-db-sync-6bdvj\" (UID: \"cdf7e311-569b-43a2-bba8-729226e28415\") " pod="openstack/glance-db-sync-6bdvj" Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.884888 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-config-data\") pod \"glance-db-sync-6bdvj\" (UID: \"cdf7e311-569b-43a2-bba8-729226e28415\") " pod="openstack/glance-db-sync-6bdvj" Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.980228 4997 generic.go:334] "Generic (PLEG): container finished" podID="77cf9c93-8871-4190-8522-3cb69ede54dd" containerID="d25c398bc81fe653079c2d969c3c9982a9b528001068fecd4ced24e81fd5d86e" exitCode=0 Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.980296 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ktw8k" event={"ID":"77cf9c93-8871-4190-8522-3cb69ede54dd","Type":"ContainerDied","Data":"d25c398bc81fe653079c2d969c3c9982a9b528001068fecd4ced24e81fd5d86e"} Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.980742 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ktw8k" event={"ID":"77cf9c93-8871-4190-8522-3cb69ede54dd","Type":"ContainerStarted","Data":"4537589c112cdfd478301e78b4a7c4140dcf39ada77e1b1027dfc49cd4ae32ab"} Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.983167 4997 generic.go:334] "Generic (PLEG): container finished" podID="377ed320-c385-42a5-852e-31360c519c6d" containerID="cc518cfb3d28921fd9f53b3004b269e3e22458213667c45944046eb579e9e955" exitCode=0 Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.983212 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3eec-account-create-update-tckr2" event={"ID":"377ed320-c385-42a5-852e-31360c519c6d","Type":"ContainerDied","Data":"cc518cfb3d28921fd9f53b3004b269e3e22458213667c45944046eb579e9e955"} Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.983305 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3eec-account-create-update-tckr2" event={"ID":"377ed320-c385-42a5-852e-31360c519c6d","Type":"ContainerStarted","Data":"b19da1211943c8ef1494ba615c73ed3f2265e9e432001611a76f71541dd500a5"} Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.988954 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-combined-ca-bundle\") pod \"glance-db-sync-6bdvj\" (UID: \"cdf7e311-569b-43a2-bba8-729226e28415\") " pod="openstack/glance-db-sync-6bdvj" Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.989015 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-config-data\") pod \"glance-db-sync-6bdvj\" (UID: \"cdf7e311-569b-43a2-bba8-729226e28415\") " pod="openstack/glance-db-sync-6bdvj" Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.989055 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-db-sync-config-data\") pod \"glance-db-sync-6bdvj\" (UID: \"cdf7e311-569b-43a2-bba8-729226e28415\") " pod="openstack/glance-db-sync-6bdvj" Dec 05 07:18:24 crc kubenswrapper[4997]: I1205 07:18:24.989096 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrrqz\" (UniqueName: \"kubernetes.io/projected/cdf7e311-569b-43a2-bba8-729226e28415-kube-api-access-nrrqz\") pod \"glance-db-sync-6bdvj\" (UID: \"cdf7e311-569b-43a2-bba8-729226e28415\") " pod="openstack/glance-db-sync-6bdvj" Dec 05 07:18:25 crc kubenswrapper[4997]: I1205 07:18:25.002690 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-db-sync-config-data\") pod \"glance-db-sync-6bdvj\" (UID: \"cdf7e311-569b-43a2-bba8-729226e28415\") " pod="openstack/glance-db-sync-6bdvj" Dec 05 07:18:25 crc kubenswrapper[4997]: I1205 07:18:25.002967 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-config-data\") pod \"glance-db-sync-6bdvj\" (UID: \"cdf7e311-569b-43a2-bba8-729226e28415\") " pod="openstack/glance-db-sync-6bdvj" Dec 05 07:18:25 crc kubenswrapper[4997]: I1205 07:18:25.002982 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-combined-ca-bundle\") pod \"glance-db-sync-6bdvj\" (UID: \"cdf7e311-569b-43a2-bba8-729226e28415\") " pod="openstack/glance-db-sync-6bdvj" Dec 05 07:18:25 crc kubenswrapper[4997]: I1205 07:18:25.013252 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrrqz\" (UniqueName: \"kubernetes.io/projected/cdf7e311-569b-43a2-bba8-729226e28415-kube-api-access-nrrqz\") pod \"glance-db-sync-6bdvj\" (UID: \"cdf7e311-569b-43a2-bba8-729226e28415\") " pod="openstack/glance-db-sync-6bdvj" Dec 05 07:18:25 crc kubenswrapper[4997]: I1205 07:18:25.076805 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6bdvj" Dec 05 07:18:25 crc kubenswrapper[4997]: I1205 07:18:25.665956 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-6bdvj"] Dec 05 07:18:25 crc kubenswrapper[4997]: W1205 07:18:25.668329 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdf7e311_569b_43a2_bba8_729226e28415.slice/crio-6e8d450d143a8fda737a6916a193e79a26300ba137ac29950355ad0e49f4618a WatchSource:0}: Error finding container 6e8d450d143a8fda737a6916a193e79a26300ba137ac29950355ad0e49f4618a: Status 404 returned error can't find the container with id 6e8d450d143a8fda737a6916a193e79a26300ba137ac29950355ad0e49f4618a Dec 05 07:18:25 crc kubenswrapper[4997]: I1205 07:18:25.993412 4997 generic.go:334] "Generic (PLEG): container finished" podID="508d4dec-7bf0-4053-a665-b1bf87897981" containerID="5f739def7163631384233edc764a3f8a516103dbcdc15f0624570994d94c96c8" exitCode=0 Dec 05 07:18:25 crc kubenswrapper[4997]: I1205 07:18:25.993484 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jqnrc" event={"ID":"508d4dec-7bf0-4053-a665-b1bf87897981","Type":"ContainerDied","Data":"5f739def7163631384233edc764a3f8a516103dbcdc15f0624570994d94c96c8"} Dec 05 07:18:25 crc kubenswrapper[4997]: I1205 07:18:25.995264 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6bdvj" event={"ID":"cdf7e311-569b-43a2-bba8-729226e28415","Type":"ContainerStarted","Data":"6e8d450d143a8fda737a6916a193e79a26300ba137ac29950355ad0e49f4618a"} Dec 05 07:18:26 crc kubenswrapper[4997]: I1205 07:18:26.369177 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3eec-account-create-update-tckr2" Dec 05 07:18:26 crc kubenswrapper[4997]: I1205 07:18:26.492121 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ktw8k" Dec 05 07:18:26 crc kubenswrapper[4997]: I1205 07:18:26.550633 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6l5x\" (UniqueName: \"kubernetes.io/projected/377ed320-c385-42a5-852e-31360c519c6d-kube-api-access-m6l5x\") pod \"377ed320-c385-42a5-852e-31360c519c6d\" (UID: \"377ed320-c385-42a5-852e-31360c519c6d\") " Dec 05 07:18:26 crc kubenswrapper[4997]: I1205 07:18:26.550883 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/377ed320-c385-42a5-852e-31360c519c6d-operator-scripts\") pod \"377ed320-c385-42a5-852e-31360c519c6d\" (UID: \"377ed320-c385-42a5-852e-31360c519c6d\") " Dec 05 07:18:26 crc kubenswrapper[4997]: I1205 07:18:26.552247 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/377ed320-c385-42a5-852e-31360c519c6d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "377ed320-c385-42a5-852e-31360c519c6d" (UID: "377ed320-c385-42a5-852e-31360c519c6d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:26 crc kubenswrapper[4997]: I1205 07:18:26.561180 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/377ed320-c385-42a5-852e-31360c519c6d-kube-api-access-m6l5x" (OuterVolumeSpecName: "kube-api-access-m6l5x") pod "377ed320-c385-42a5-852e-31360c519c6d" (UID: "377ed320-c385-42a5-852e-31360c519c6d"). InnerVolumeSpecName "kube-api-access-m6l5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:18:26 crc kubenswrapper[4997]: I1205 07:18:26.653077 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77cf9c93-8871-4190-8522-3cb69ede54dd-operator-scripts\") pod \"77cf9c93-8871-4190-8522-3cb69ede54dd\" (UID: \"77cf9c93-8871-4190-8522-3cb69ede54dd\") " Dec 05 07:18:26 crc kubenswrapper[4997]: I1205 07:18:26.653527 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5hd4\" (UniqueName: \"kubernetes.io/projected/77cf9c93-8871-4190-8522-3cb69ede54dd-kube-api-access-t5hd4\") pod \"77cf9c93-8871-4190-8522-3cb69ede54dd\" (UID: \"77cf9c93-8871-4190-8522-3cb69ede54dd\") " Dec 05 07:18:26 crc kubenswrapper[4997]: I1205 07:18:26.653846 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77cf9c93-8871-4190-8522-3cb69ede54dd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "77cf9c93-8871-4190-8522-3cb69ede54dd" (UID: "77cf9c93-8871-4190-8522-3cb69ede54dd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:26 crc kubenswrapper[4997]: I1205 07:18:26.654291 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/377ed320-c385-42a5-852e-31360c519c6d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:26 crc kubenswrapper[4997]: I1205 07:18:26.654385 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77cf9c93-8871-4190-8522-3cb69ede54dd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:26 crc kubenswrapper[4997]: I1205 07:18:26.654455 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6l5x\" (UniqueName: \"kubernetes.io/projected/377ed320-c385-42a5-852e-31360c519c6d-kube-api-access-m6l5x\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:26 crc kubenswrapper[4997]: I1205 07:18:26.656657 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77cf9c93-8871-4190-8522-3cb69ede54dd-kube-api-access-t5hd4" (OuterVolumeSpecName: "kube-api-access-t5hd4") pod "77cf9c93-8871-4190-8522-3cb69ede54dd" (UID: "77cf9c93-8871-4190-8522-3cb69ede54dd"). InnerVolumeSpecName "kube-api-access-t5hd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:18:26 crc kubenswrapper[4997]: I1205 07:18:26.756333 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5hd4\" (UniqueName: \"kubernetes.io/projected/77cf9c93-8871-4190-8522-3cb69ede54dd-kube-api-access-t5hd4\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.026277 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3eec-account-create-update-tckr2" event={"ID":"377ed320-c385-42a5-852e-31360c519c6d","Type":"ContainerDied","Data":"b19da1211943c8ef1494ba615c73ed3f2265e9e432001611a76f71541dd500a5"} Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.026337 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b19da1211943c8ef1494ba615c73ed3f2265e9e432001611a76f71541dd500a5" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.026446 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3eec-account-create-update-tckr2" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.030084 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ktw8k" event={"ID":"77cf9c93-8871-4190-8522-3cb69ede54dd","Type":"ContainerDied","Data":"4537589c112cdfd478301e78b4a7c4140dcf39ada77e1b1027dfc49cd4ae32ab"} Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.030165 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4537589c112cdfd478301e78b4a7c4140dcf39ada77e1b1027dfc49cd4ae32ab" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.030235 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ktw8k" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.447378 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.598292 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nvw8\" (UniqueName: \"kubernetes.io/projected/508d4dec-7bf0-4053-a665-b1bf87897981-kube-api-access-4nvw8\") pod \"508d4dec-7bf0-4053-a665-b1bf87897981\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.598437 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/508d4dec-7bf0-4053-a665-b1bf87897981-etc-swift\") pod \"508d4dec-7bf0-4053-a665-b1bf87897981\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.598491 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-dispersionconf\") pod \"508d4dec-7bf0-4053-a665-b1bf87897981\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.598532 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-combined-ca-bundle\") pod \"508d4dec-7bf0-4053-a665-b1bf87897981\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.598604 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-swiftconf\") pod \"508d4dec-7bf0-4053-a665-b1bf87897981\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.598680 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/508d4dec-7bf0-4053-a665-b1bf87897981-scripts\") pod \"508d4dec-7bf0-4053-a665-b1bf87897981\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.598720 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/508d4dec-7bf0-4053-a665-b1bf87897981-ring-data-devices\") pod \"508d4dec-7bf0-4053-a665-b1bf87897981\" (UID: \"508d4dec-7bf0-4053-a665-b1bf87897981\") " Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.600362 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/508d4dec-7bf0-4053-a665-b1bf87897981-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "508d4dec-7bf0-4053-a665-b1bf87897981" (UID: "508d4dec-7bf0-4053-a665-b1bf87897981"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.600976 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/508d4dec-7bf0-4053-a665-b1bf87897981-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "508d4dec-7bf0-4053-a665-b1bf87897981" (UID: "508d4dec-7bf0-4053-a665-b1bf87897981"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.605266 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/508d4dec-7bf0-4053-a665-b1bf87897981-kube-api-access-4nvw8" (OuterVolumeSpecName: "kube-api-access-4nvw8") pod "508d4dec-7bf0-4053-a665-b1bf87897981" (UID: "508d4dec-7bf0-4053-a665-b1bf87897981"). InnerVolumeSpecName "kube-api-access-4nvw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.610252 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "508d4dec-7bf0-4053-a665-b1bf87897981" (UID: "508d4dec-7bf0-4053-a665-b1bf87897981"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.625359 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/508d4dec-7bf0-4053-a665-b1bf87897981-scripts" (OuterVolumeSpecName: "scripts") pod "508d4dec-7bf0-4053-a665-b1bf87897981" (UID: "508d4dec-7bf0-4053-a665-b1bf87897981"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.626471 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "508d4dec-7bf0-4053-a665-b1bf87897981" (UID: "508d4dec-7bf0-4053-a665-b1bf87897981"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.629637 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "508d4dec-7bf0-4053-a665-b1bf87897981" (UID: "508d4dec-7bf0-4053-a665-b1bf87897981"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.701263 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nvw8\" (UniqueName: \"kubernetes.io/projected/508d4dec-7bf0-4053-a665-b1bf87897981-kube-api-access-4nvw8\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.701310 4997 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/508d4dec-7bf0-4053-a665-b1bf87897981-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.701320 4997 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.701329 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.701339 4997 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/508d4dec-7bf0-4053-a665-b1bf87897981-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.701347 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/508d4dec-7bf0-4053-a665-b1bf87897981-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:27 crc kubenswrapper[4997]: I1205 07:18:27.701355 4997 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/508d4dec-7bf0-4053-a665-b1bf87897981-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:28 crc kubenswrapper[4997]: I1205 07:18:28.039736 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jqnrc" event={"ID":"508d4dec-7bf0-4053-a665-b1bf87897981","Type":"ContainerDied","Data":"b24f60982d9e47f22e03b07911f02809961869fcf3d88cb2d1f637ccd4d37403"} Dec 05 07:18:28 crc kubenswrapper[4997]: I1205 07:18:28.039793 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b24f60982d9e47f22e03b07911f02809961869fcf3d88cb2d1f637ccd4d37403" Dec 05 07:18:28 crc kubenswrapper[4997]: I1205 07:18:28.039814 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jqnrc" Dec 05 07:18:28 crc kubenswrapper[4997]: I1205 07:18:28.328556 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vnncj" podUID="5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" containerName="ovn-controller" probeResult="failure" output=< Dec 05 07:18:28 crc kubenswrapper[4997]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 05 07:18:28 crc kubenswrapper[4997]: > Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.096210 4997 generic.go:334] "Generic (PLEG): container finished" podID="1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" containerID="711f304ea598c5aa2085e471f825cc1a9ccb2da36ded6e28c57872ee1771bbfd" exitCode=0 Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.096271 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79","Type":"ContainerDied","Data":"711f304ea598c5aa2085e471f825cc1a9ccb2da36ded6e28c57872ee1771bbfd"} Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.316174 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vnncj" podUID="5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" containerName="ovn-controller" probeResult="failure" output=< Dec 05 07:18:33 crc kubenswrapper[4997]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 05 07:18:33 crc kubenswrapper[4997]: > Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.336539 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.338969 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.594896 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vnncj-config-nk456"] Dec 05 07:18:33 crc kubenswrapper[4997]: E1205 07:18:33.595446 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77cf9c93-8871-4190-8522-3cb69ede54dd" containerName="mariadb-database-create" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.595467 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="77cf9c93-8871-4190-8522-3cb69ede54dd" containerName="mariadb-database-create" Dec 05 07:18:33 crc kubenswrapper[4997]: E1205 07:18:33.595487 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="508d4dec-7bf0-4053-a665-b1bf87897981" containerName="swift-ring-rebalance" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.595496 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="508d4dec-7bf0-4053-a665-b1bf87897981" containerName="swift-ring-rebalance" Dec 05 07:18:33 crc kubenswrapper[4997]: E1205 07:18:33.595520 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="377ed320-c385-42a5-852e-31360c519c6d" containerName="mariadb-account-create-update" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.595529 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="377ed320-c385-42a5-852e-31360c519c6d" containerName="mariadb-account-create-update" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.595735 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="508d4dec-7bf0-4053-a665-b1bf87897981" containerName="swift-ring-rebalance" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.595763 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="77cf9c93-8871-4190-8522-3cb69ede54dd" containerName="mariadb-database-create" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.595775 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="377ed320-c385-42a5-852e-31360c519c6d" containerName="mariadb-account-create-update" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.596715 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.599096 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.603620 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vnncj-config-nk456"] Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.738220 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-run-ovn\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.738805 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/faef8897-c15d-45a4-9b50-107f653e5a31-additional-scripts\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.738889 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-log-ovn\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.738917 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faef8897-c15d-45a4-9b50-107f653e5a31-scripts\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.738986 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-run\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.739014 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pw2v\" (UniqueName: \"kubernetes.io/projected/faef8897-c15d-45a4-9b50-107f653e5a31-kube-api-access-2pw2v\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.840533 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-run\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.840603 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pw2v\" (UniqueName: \"kubernetes.io/projected/faef8897-c15d-45a4-9b50-107f653e5a31-kube-api-access-2pw2v\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.840739 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-run-ovn\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.840806 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/faef8897-c15d-45a4-9b50-107f653e5a31-additional-scripts\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.840855 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-log-ovn\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.840882 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faef8897-c15d-45a4-9b50-107f653e5a31-scripts\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.840928 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-run\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.842230 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-run-ovn\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.842299 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-log-ovn\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.843315 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faef8897-c15d-45a4-9b50-107f653e5a31-scripts\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.843968 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/faef8897-c15d-45a4-9b50-107f653e5a31-additional-scripts\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.864885 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pw2v\" (UniqueName: \"kubernetes.io/projected/faef8897-c15d-45a4-9b50-107f653e5a31-kube-api-access-2pw2v\") pod \"ovn-controller-vnncj-config-nk456\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:33 crc kubenswrapper[4997]: I1205 07:18:33.920935 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:34 crc kubenswrapper[4997]: I1205 07:18:34.109978 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0f158874-152b-46ce-ac42-d202684853ca","Type":"ContainerDied","Data":"3a4f08d0ae39c150584948dc3f19fc684275263ada8f51f39f56460972b674d2"} Dec 05 07:18:34 crc kubenswrapper[4997]: I1205 07:18:34.109915 4997 generic.go:334] "Generic (PLEG): container finished" podID="0f158874-152b-46ce-ac42-d202684853ca" containerID="3a4f08d0ae39c150584948dc3f19fc684275263ada8f51f39f56460972b674d2" exitCode=0 Dec 05 07:18:38 crc kubenswrapper[4997]: I1205 07:18:38.322409 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vnncj" podUID="5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" containerName="ovn-controller" probeResult="failure" output=< Dec 05 07:18:38 crc kubenswrapper[4997]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 05 07:18:38 crc kubenswrapper[4997]: > Dec 05 07:18:39 crc kubenswrapper[4997]: I1205 07:18:39.695529 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:39 crc kubenswrapper[4997]: I1205 07:18:39.704003 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift\") pod \"swift-storage-0\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " pod="openstack/swift-storage-0" Dec 05 07:18:39 crc kubenswrapper[4997]: I1205 07:18:39.928058 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 05 07:18:43 crc kubenswrapper[4997]: I1205 07:18:43.321630 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vnncj" podUID="5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" containerName="ovn-controller" probeResult="failure" output=< Dec 05 07:18:43 crc kubenswrapper[4997]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 05 07:18:43 crc kubenswrapper[4997]: > Dec 05 07:18:43 crc kubenswrapper[4997]: E1205 07:18:43.631085 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api@sha256:481073ac9deefb38bbd989aaa8dd7aedb4e0af26017f4883f85fce433380bf63" Dec 05 07:18:43 crc kubenswrapper[4997]: E1205 07:18:43.631511 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:481073ac9deefb38bbd989aaa8dd7aedb4e0af26017f4883f85fce433380bf63,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nrrqz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-6bdvj_openstack(cdf7e311-569b-43a2-bba8-729226e28415): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:18:43 crc kubenswrapper[4997]: E1205 07:18:43.633209 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-6bdvj" podUID="cdf7e311-569b-43a2-bba8-729226e28415" Dec 05 07:18:44 crc kubenswrapper[4997]: I1205 07:18:44.168385 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vnncj-config-nk456"] Dec 05 07:18:44 crc kubenswrapper[4997]: W1205 07:18:44.173088 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfaef8897_c15d_45a4_9b50_107f653e5a31.slice/crio-23e1e52cafe2ae070f9152911f8d04abdd6d0f8f62b45b0d3fad60765e9f1e68 WatchSource:0}: Error finding container 23e1e52cafe2ae070f9152911f8d04abdd6d0f8f62b45b0d3fad60765e9f1e68: Status 404 returned error can't find the container with id 23e1e52cafe2ae070f9152911f8d04abdd6d0f8f62b45b0d3fad60765e9f1e68 Dec 05 07:18:44 crc kubenswrapper[4997]: I1205 07:18:44.207647 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 05 07:18:44 crc kubenswrapper[4997]: I1205 07:18:44.249302 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vnncj-config-nk456" event={"ID":"faef8897-c15d-45a4-9b50-107f653e5a31","Type":"ContainerStarted","Data":"23e1e52cafe2ae070f9152911f8d04abdd6d0f8f62b45b0d3fad60765e9f1e68"} Dec 05 07:18:44 crc kubenswrapper[4997]: I1205 07:18:44.252011 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0f158874-152b-46ce-ac42-d202684853ca","Type":"ContainerStarted","Data":"ad9f0fd9866254fcb3d271814a479f3ce3ebdaecf8def2eb7a15d7c6654387e1"} Dec 05 07:18:44 crc kubenswrapper[4997]: I1205 07:18:44.252340 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:18:44 crc kubenswrapper[4997]: I1205 07:18:44.255105 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79","Type":"ContainerStarted","Data":"bd15ae01a5d3986a7e4425b35e0fa45dbfda1aea5cdf404c87ede67fab61b69c"} Dec 05 07:18:44 crc kubenswrapper[4997]: I1205 07:18:44.255566 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 05 07:18:44 crc kubenswrapper[4997]: I1205 07:18:44.262552 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerStarted","Data":"a042f6522375f75c48ce6fb440a64a329021e4a5723014c8733e72171176d91c"} Dec 05 07:18:44 crc kubenswrapper[4997]: E1205 07:18:44.263541 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api@sha256:481073ac9deefb38bbd989aaa8dd7aedb4e0af26017f4883f85fce433380bf63\\\"\"" pod="openstack/glance-db-sync-6bdvj" podUID="cdf7e311-569b-43a2-bba8-729226e28415" Dec 05 07:18:44 crc kubenswrapper[4997]: I1205 07:18:44.283837 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371930.57096 podStartE2EDuration="1m46.283816936s" podCreationTimestamp="2025-12-05 07:16:58 +0000 UTC" firstStartedPulling="2025-12-05 07:17:02.515933965 +0000 UTC m=+1323.044841226" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:18:44.279782307 +0000 UTC m=+1424.808689598" watchObservedRunningTime="2025-12-05 07:18:44.283816936 +0000 UTC m=+1424.812724187" Dec 05 07:18:44 crc kubenswrapper[4997]: I1205 07:18:44.324129 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=48.912394412 podStartE2EDuration="1m45.324108053s" podCreationTimestamp="2025-12-05 07:16:59 +0000 UTC" firstStartedPulling="2025-12-05 07:17:02.753519538 +0000 UTC m=+1323.282426799" lastFinishedPulling="2025-12-05 07:17:59.165233179 +0000 UTC m=+1379.694140440" observedRunningTime="2025-12-05 07:18:44.320743432 +0000 UTC m=+1424.849650713" watchObservedRunningTime="2025-12-05 07:18:44.324108053 +0000 UTC m=+1424.853015314" Dec 05 07:18:45 crc kubenswrapper[4997]: I1205 07:18:45.272303 4997 generic.go:334] "Generic (PLEG): container finished" podID="faef8897-c15d-45a4-9b50-107f653e5a31" containerID="fc45b8c8dbde8e0944b28f364554fb8e350727d01cdcddc568ede9ce3f181336" exitCode=0 Dec 05 07:18:45 crc kubenswrapper[4997]: I1205 07:18:45.273752 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vnncj-config-nk456" event={"ID":"faef8897-c15d-45a4-9b50-107f653e5a31","Type":"ContainerDied","Data":"fc45b8c8dbde8e0944b28f364554fb8e350727d01cdcddc568ede9ce3f181336"} Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.282084 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerStarted","Data":"e62736a4a4741c846c90c4aeb88e01bdf1db4768265d9eb4a765c3853650e11e"} Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.282430 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerStarted","Data":"7ce96074e3871701d5fda41b91a5e0c7fc16c1a429f4da49f7b22d0a58f5bcfa"} Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.282443 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerStarted","Data":"5adcaef44afa1157fae6d04d911a954e618faa7d796126d8888521add2da41db"} Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.282453 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerStarted","Data":"71f731b2d431508653bb17ac618c7642061a78120cb47e7911dfe3ee31b05713"} Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.608451 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.746495 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-run\") pod \"faef8897-c15d-45a4-9b50-107f653e5a31\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.746575 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-log-ovn\") pod \"faef8897-c15d-45a4-9b50-107f653e5a31\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.746653 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faef8897-c15d-45a4-9b50-107f653e5a31-scripts\") pod \"faef8897-c15d-45a4-9b50-107f653e5a31\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.746684 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-run-ovn\") pod \"faef8897-c15d-45a4-9b50-107f653e5a31\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.746747 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pw2v\" (UniqueName: \"kubernetes.io/projected/faef8897-c15d-45a4-9b50-107f653e5a31-kube-api-access-2pw2v\") pod \"faef8897-c15d-45a4-9b50-107f653e5a31\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.746669 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-run" (OuterVolumeSpecName: "var-run") pod "faef8897-c15d-45a4-9b50-107f653e5a31" (UID: "faef8897-c15d-45a4-9b50-107f653e5a31"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.746696 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "faef8897-c15d-45a4-9b50-107f653e5a31" (UID: "faef8897-c15d-45a4-9b50-107f653e5a31"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.746740 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "faef8897-c15d-45a4-9b50-107f653e5a31" (UID: "faef8897-c15d-45a4-9b50-107f653e5a31"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.746857 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/faef8897-c15d-45a4-9b50-107f653e5a31-additional-scripts\") pod \"faef8897-c15d-45a4-9b50-107f653e5a31\" (UID: \"faef8897-c15d-45a4-9b50-107f653e5a31\") " Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.747258 4997 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-run\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.747273 4997 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.747284 4997 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/faef8897-c15d-45a4-9b50-107f653e5a31-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.747880 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faef8897-c15d-45a4-9b50-107f653e5a31-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "faef8897-c15d-45a4-9b50-107f653e5a31" (UID: "faef8897-c15d-45a4-9b50-107f653e5a31"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.747980 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faef8897-c15d-45a4-9b50-107f653e5a31-scripts" (OuterVolumeSpecName: "scripts") pod "faef8897-c15d-45a4-9b50-107f653e5a31" (UID: "faef8897-c15d-45a4-9b50-107f653e5a31"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.753594 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faef8897-c15d-45a4-9b50-107f653e5a31-kube-api-access-2pw2v" (OuterVolumeSpecName: "kube-api-access-2pw2v") pod "faef8897-c15d-45a4-9b50-107f653e5a31" (UID: "faef8897-c15d-45a4-9b50-107f653e5a31"). InnerVolumeSpecName "kube-api-access-2pw2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.848787 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faef8897-c15d-45a4-9b50-107f653e5a31-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.848836 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pw2v\" (UniqueName: \"kubernetes.io/projected/faef8897-c15d-45a4-9b50-107f653e5a31-kube-api-access-2pw2v\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:46 crc kubenswrapper[4997]: I1205 07:18:46.848847 4997 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/faef8897-c15d-45a4-9b50-107f653e5a31-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:18:47 crc kubenswrapper[4997]: I1205 07:18:47.296852 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vnncj-config-nk456" event={"ID":"faef8897-c15d-45a4-9b50-107f653e5a31","Type":"ContainerDied","Data":"23e1e52cafe2ae070f9152911f8d04abdd6d0f8f62b45b0d3fad60765e9f1e68"} Dec 05 07:18:47 crc kubenswrapper[4997]: I1205 07:18:47.296919 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23e1e52cafe2ae070f9152911f8d04abdd6d0f8f62b45b0d3fad60765e9f1e68" Dec 05 07:18:47 crc kubenswrapper[4997]: I1205 07:18:47.296933 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnncj-config-nk456" Dec 05 07:18:47 crc kubenswrapper[4997]: I1205 07:18:47.740839 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-vnncj-config-nk456"] Dec 05 07:18:47 crc kubenswrapper[4997]: I1205 07:18:47.760701 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-vnncj-config-nk456"] Dec 05 07:18:48 crc kubenswrapper[4997]: I1205 07:18:48.346064 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerStarted","Data":"2c3fe897d3a3071e0e6d73643c8249d5b3a509b04f3d1cf2c20850f1110f9dc0"} Dec 05 07:18:48 crc kubenswrapper[4997]: I1205 07:18:48.346138 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerStarted","Data":"ade129587cb10e9f61150cba8b6492aafd796f81df59fc5a8e2e4f2b826b4574"} Dec 05 07:18:48 crc kubenswrapper[4997]: I1205 07:18:48.346152 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerStarted","Data":"e982dc1e734d65f84320cc1a7e6b801c7989b49043d99b850c0c838b530b2638"} Dec 05 07:18:48 crc kubenswrapper[4997]: I1205 07:18:48.378326 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-vnncj" Dec 05 07:18:49 crc kubenswrapper[4997]: I1205 07:18:49.360140 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerStarted","Data":"732523b307a12a50490d6b3563dce9bc973897fbfa53f0e26ffcb66918c5cef4"} Dec 05 07:18:49 crc kubenswrapper[4997]: I1205 07:18:49.765690 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faef8897-c15d-45a4-9b50-107f653e5a31" path="/var/lib/kubelet/pods/faef8897-c15d-45a4-9b50-107f653e5a31/volumes" Dec 05 07:18:49 crc kubenswrapper[4997]: I1205 07:18:49.770290 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:18:49 crc kubenswrapper[4997]: I1205 07:18:49.770361 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:18:50 crc kubenswrapper[4997]: I1205 07:18:50.376646 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerStarted","Data":"f483369c789ec0f4e5bb24b517217914add2901c3b684a8154f738f83e6fa37b"} Dec 05 07:18:50 crc kubenswrapper[4997]: I1205 07:18:50.377133 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerStarted","Data":"b87c752a856aa32bd9d97eb8af0b3d65d2a9e9d6f3d911ba2d99662c5e3deef3"} Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.391047 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerStarted","Data":"4fb495cc6962ca168a43dc4d03761c144cd6fa4532a0aa11050db4ee869e4f5e"} Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.391389 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerStarted","Data":"d09e7dc200cd7427ec542e14f38b16b200be9b00f5c8bcc72aa1988da898ee42"} Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.391408 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerStarted","Data":"82fa232b03602f1c8aad22b6c099f8b323fcdf24fc81e0fba8dc78d929f37d29"} Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.391421 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerStarted","Data":"6916fdc8a79724bc4f616151198996d4533bcdfd1af49ce395a01c0843637cf3"} Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.391433 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerStarted","Data":"7f5cbf566dd19f6211cf0d1dae49e71d2d0c93c0e3cac1ccb8cc390c300b5e33"} Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.425259 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=39.772413057 podStartE2EDuration="45.425241844s" podCreationTimestamp="2025-12-05 07:18:06 +0000 UTC" firstStartedPulling="2025-12-05 07:18:44.218442644 +0000 UTC m=+1424.747349905" lastFinishedPulling="2025-12-05 07:18:49.871271431 +0000 UTC m=+1430.400178692" observedRunningTime="2025-12-05 07:18:51.422155201 +0000 UTC m=+1431.951062462" watchObservedRunningTime="2025-12-05 07:18:51.425241844 +0000 UTC m=+1431.954149105" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.743432 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-vz755"] Dec 05 07:18:51 crc kubenswrapper[4997]: E1205 07:18:51.743948 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faef8897-c15d-45a4-9b50-107f653e5a31" containerName="ovn-config" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.743975 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="faef8897-c15d-45a4-9b50-107f653e5a31" containerName="ovn-config" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.744241 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="faef8897-c15d-45a4-9b50-107f653e5a31" containerName="ovn-config" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.747083 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.753003 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.811865 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-vz755"] Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.846150 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9zdv\" (UniqueName: \"kubernetes.io/projected/bf7639fd-80e4-4f37-8f92-9b9f73a01557-kube-api-access-m9zdv\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.846213 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-dns-svc\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.846240 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-ovsdbserver-nb\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.846316 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-ovsdbserver-sb\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.846478 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-dns-swift-storage-0\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.846516 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-config\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.947960 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-dns-swift-storage-0\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.948026 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-config\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.948952 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-dns-swift-storage-0\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.949059 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-config\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.949204 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9zdv\" (UniqueName: \"kubernetes.io/projected/bf7639fd-80e4-4f37-8f92-9b9f73a01557-kube-api-access-m9zdv\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.949235 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-dns-svc\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.949251 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-ovsdbserver-nb\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.949276 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-ovsdbserver-sb\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.949856 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-ovsdbserver-sb\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.950702 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-dns-svc\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.951199 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-ovsdbserver-nb\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:51 crc kubenswrapper[4997]: I1205 07:18:51.969107 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9zdv\" (UniqueName: \"kubernetes.io/projected/bf7639fd-80e4-4f37-8f92-9b9f73a01557-kube-api-access-m9zdv\") pod \"dnsmasq-dns-779c5847bc-vz755\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:52 crc kubenswrapper[4997]: I1205 07:18:52.119548 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:52 crc kubenswrapper[4997]: I1205 07:18:52.384708 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-vz755"] Dec 05 07:18:52 crc kubenswrapper[4997]: I1205 07:18:52.401875 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-779c5847bc-vz755" event={"ID":"bf7639fd-80e4-4f37-8f92-9b9f73a01557","Type":"ContainerStarted","Data":"f618dc8b8a8b99a60ec86a1d35977485f7a87f6a19bf19a722ef73bdd34d79be"} Dec 05 07:18:53 crc kubenswrapper[4997]: I1205 07:18:53.426742 4997 generic.go:334] "Generic (PLEG): container finished" podID="bf7639fd-80e4-4f37-8f92-9b9f73a01557" containerID="92b3b79042a60525a1139389eba2978e16aefef7e8b713c4e759357afa9f7028" exitCode=0 Dec 05 07:18:53 crc kubenswrapper[4997]: I1205 07:18:53.426916 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-779c5847bc-vz755" event={"ID":"bf7639fd-80e4-4f37-8f92-9b9f73a01557","Type":"ContainerDied","Data":"92b3b79042a60525a1139389eba2978e16aefef7e8b713c4e759357afa9f7028"} Dec 05 07:18:54 crc kubenswrapper[4997]: I1205 07:18:54.437839 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-779c5847bc-vz755" event={"ID":"bf7639fd-80e4-4f37-8f92-9b9f73a01557","Type":"ContainerStarted","Data":"d59ccec3588a99dd3d41c828fb0ed5ed9b64b9c1c576b222b7fac17148e89628"} Dec 05 07:18:54 crc kubenswrapper[4997]: I1205 07:18:54.438340 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:18:54 crc kubenswrapper[4997]: I1205 07:18:54.457890 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-779c5847bc-vz755" podStartSLOduration=3.457865638 podStartE2EDuration="3.457865638s" podCreationTimestamp="2025-12-05 07:18:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:18:54.456700067 +0000 UTC m=+1434.985607398" watchObservedRunningTime="2025-12-05 07:18:54.457865638 +0000 UTC m=+1434.986772899" Dec 05 07:19:00 crc kubenswrapper[4997]: I1205 07:19:00.491341 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6bdvj" event={"ID":"cdf7e311-569b-43a2-bba8-729226e28415","Type":"ContainerStarted","Data":"5e883daffbf9da9e1ebaee0593862fba2d57d7337e7c8104926342e03ff1d1d6"} Dec 05 07:19:00 crc kubenswrapper[4997]: I1205 07:19:00.809868 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:19:00 crc kubenswrapper[4997]: I1205 07:19:00.845341 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-6bdvj" podStartSLOduration=3.16098683 podStartE2EDuration="36.845314005s" podCreationTimestamp="2025-12-05 07:18:24 +0000 UTC" firstStartedPulling="2025-12-05 07:18:25.672892434 +0000 UTC m=+1406.201799735" lastFinishedPulling="2025-12-05 07:18:59.357219649 +0000 UTC m=+1439.886126910" observedRunningTime="2025-12-05 07:19:00.512637754 +0000 UTC m=+1441.041545035" watchObservedRunningTime="2025-12-05 07:19:00.845314005 +0000 UTC m=+1441.374221276" Dec 05 07:19:01 crc kubenswrapper[4997]: I1205 07:19:01.890906 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.124801 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.201192 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-bp85j"] Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.201778 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" podUID="b0f3255a-0cbc-4d2a-bd39-526cbf142fa9" containerName="dnsmasq-dns" containerID="cri-o://792bc50db82e23371bcb60395d2f639459b0e1762fba29595170fd9dc47ea22e" gracePeriod=10 Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.514859 4997 generic.go:334] "Generic (PLEG): container finished" podID="b0f3255a-0cbc-4d2a-bd39-526cbf142fa9" containerID="792bc50db82e23371bcb60395d2f639459b0e1762fba29595170fd9dc47ea22e" exitCode=0 Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.514912 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" event={"ID":"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9","Type":"ContainerDied","Data":"792bc50db82e23371bcb60395d2f639459b0e1762fba29595170fd9dc47ea22e"} Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.724395 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.857927 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-dns-svc\") pod \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.858394 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-ovsdbserver-nb\") pod \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.858434 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shdnn\" (UniqueName: \"kubernetes.io/projected/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-kube-api-access-shdnn\") pod \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.858503 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-ovsdbserver-sb\") pod \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.858538 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-config\") pod \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\" (UID: \"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9\") " Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.879978 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-kube-api-access-shdnn" (OuterVolumeSpecName: "kube-api-access-shdnn") pod "b0f3255a-0cbc-4d2a-bd39-526cbf142fa9" (UID: "b0f3255a-0cbc-4d2a-bd39-526cbf142fa9"). InnerVolumeSpecName "kube-api-access-shdnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.932683 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-config" (OuterVolumeSpecName: "config") pod "b0f3255a-0cbc-4d2a-bd39-526cbf142fa9" (UID: "b0f3255a-0cbc-4d2a-bd39-526cbf142fa9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.949759 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b0f3255a-0cbc-4d2a-bd39-526cbf142fa9" (UID: "b0f3255a-0cbc-4d2a-bd39-526cbf142fa9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.952591 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b0f3255a-0cbc-4d2a-bd39-526cbf142fa9" (UID: "b0f3255a-0cbc-4d2a-bd39-526cbf142fa9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.961027 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.961066 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shdnn\" (UniqueName: \"kubernetes.io/projected/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-kube-api-access-shdnn\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.961079 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.961089 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:02 crc kubenswrapper[4997]: I1205 07:19:02.961679 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b0f3255a-0cbc-4d2a-bd39-526cbf142fa9" (UID: "b0f3255a-0cbc-4d2a-bd39-526cbf142fa9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.062722 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.076049 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-vj5mk"] Dec 05 07:19:03 crc kubenswrapper[4997]: E1205 07:19:03.076569 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0f3255a-0cbc-4d2a-bd39-526cbf142fa9" containerName="dnsmasq-dns" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.076593 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0f3255a-0cbc-4d2a-bd39-526cbf142fa9" containerName="dnsmasq-dns" Dec 05 07:19:03 crc kubenswrapper[4997]: E1205 07:19:03.076630 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0f3255a-0cbc-4d2a-bd39-526cbf142fa9" containerName="init" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.076638 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0f3255a-0cbc-4d2a-bd39-526cbf142fa9" containerName="init" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.076851 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0f3255a-0cbc-4d2a-bd39-526cbf142fa9" containerName="dnsmasq-dns" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.077598 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vj5mk" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.083407 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-a458-account-create-update-7cfrj"] Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.084679 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a458-account-create-update-7cfrj" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.088537 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.105574 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-a458-account-create-update-7cfrj"] Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.164045 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3-operator-scripts\") pod \"cinder-a458-account-create-update-7cfrj\" (UID: \"4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3\") " pod="openstack/cinder-a458-account-create-update-7cfrj" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.164091 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbc5v\" (UniqueName: \"kubernetes.io/projected/4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3-kube-api-access-mbc5v\") pod \"cinder-a458-account-create-update-7cfrj\" (UID: \"4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3\") " pod="openstack/cinder-a458-account-create-update-7cfrj" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.164125 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f0b9769-3e80-4475-b8d3-5fe6c5a265e7-operator-scripts\") pod \"cinder-db-create-vj5mk\" (UID: \"6f0b9769-3e80-4475-b8d3-5fe6c5a265e7\") " pod="openstack/cinder-db-create-vj5mk" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.164147 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-667kb\" (UniqueName: \"kubernetes.io/projected/6f0b9769-3e80-4475-b8d3-5fe6c5a265e7-kube-api-access-667kb\") pod \"cinder-db-create-vj5mk\" (UID: \"6f0b9769-3e80-4475-b8d3-5fe6c5a265e7\") " pod="openstack/cinder-db-create-vj5mk" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.180892 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vj5mk"] Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.188189 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-zmvrc"] Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.189337 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-zmvrc" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.201359 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-zmvrc"] Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.265600 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f0b9769-3e80-4475-b8d3-5fe6c5a265e7-operator-scripts\") pod \"cinder-db-create-vj5mk\" (UID: \"6f0b9769-3e80-4475-b8d3-5fe6c5a265e7\") " pod="openstack/cinder-db-create-vj5mk" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.265662 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-667kb\" (UniqueName: \"kubernetes.io/projected/6f0b9769-3e80-4475-b8d3-5fe6c5a265e7-kube-api-access-667kb\") pod \"cinder-db-create-vj5mk\" (UID: \"6f0b9769-3e80-4475-b8d3-5fe6c5a265e7\") " pod="openstack/cinder-db-create-vj5mk" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.265740 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d28a1824-1cc9-41be-b4af-1957b982a294-operator-scripts\") pod \"barbican-db-create-zmvrc\" (UID: \"d28a1824-1cc9-41be-b4af-1957b982a294\") " pod="openstack/barbican-db-create-zmvrc" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.265784 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95np7\" (UniqueName: \"kubernetes.io/projected/d28a1824-1cc9-41be-b4af-1957b982a294-kube-api-access-95np7\") pod \"barbican-db-create-zmvrc\" (UID: \"d28a1824-1cc9-41be-b4af-1957b982a294\") " pod="openstack/barbican-db-create-zmvrc" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.265853 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3-operator-scripts\") pod \"cinder-a458-account-create-update-7cfrj\" (UID: \"4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3\") " pod="openstack/cinder-a458-account-create-update-7cfrj" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.265899 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbc5v\" (UniqueName: \"kubernetes.io/projected/4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3-kube-api-access-mbc5v\") pod \"cinder-a458-account-create-update-7cfrj\" (UID: \"4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3\") " pod="openstack/cinder-a458-account-create-update-7cfrj" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.266973 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3-operator-scripts\") pod \"cinder-a458-account-create-update-7cfrj\" (UID: \"4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3\") " pod="openstack/cinder-a458-account-create-update-7cfrj" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.267489 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f0b9769-3e80-4475-b8d3-5fe6c5a265e7-operator-scripts\") pod \"cinder-db-create-vj5mk\" (UID: \"6f0b9769-3e80-4475-b8d3-5fe6c5a265e7\") " pod="openstack/cinder-db-create-vj5mk" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.283338 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-69df-account-create-update-2w6f2"] Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.285217 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-69df-account-create-update-2w6f2" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.291071 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-69df-account-create-update-2w6f2"] Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.292063 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.305606 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbc5v\" (UniqueName: \"kubernetes.io/projected/4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3-kube-api-access-mbc5v\") pod \"cinder-a458-account-create-update-7cfrj\" (UID: \"4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3\") " pod="openstack/cinder-a458-account-create-update-7cfrj" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.309883 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-667kb\" (UniqueName: \"kubernetes.io/projected/6f0b9769-3e80-4475-b8d3-5fe6c5a265e7-kube-api-access-667kb\") pod \"cinder-db-create-vj5mk\" (UID: \"6f0b9769-3e80-4475-b8d3-5fe6c5a265e7\") " pod="openstack/cinder-db-create-vj5mk" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.369227 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8-operator-scripts\") pod \"barbican-69df-account-create-update-2w6f2\" (UID: \"b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8\") " pod="openstack/barbican-69df-account-create-update-2w6f2" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.369339 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d28a1824-1cc9-41be-b4af-1957b982a294-operator-scripts\") pod \"barbican-db-create-zmvrc\" (UID: \"d28a1824-1cc9-41be-b4af-1957b982a294\") " pod="openstack/barbican-db-create-zmvrc" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.369390 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95np7\" (UniqueName: \"kubernetes.io/projected/d28a1824-1cc9-41be-b4af-1957b982a294-kube-api-access-95np7\") pod \"barbican-db-create-zmvrc\" (UID: \"d28a1824-1cc9-41be-b4af-1957b982a294\") " pod="openstack/barbican-db-create-zmvrc" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.371809 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkfvr\" (UniqueName: \"kubernetes.io/projected/b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8-kube-api-access-mkfvr\") pod \"barbican-69df-account-create-update-2w6f2\" (UID: \"b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8\") " pod="openstack/barbican-69df-account-create-update-2w6f2" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.381158 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d28a1824-1cc9-41be-b4af-1957b982a294-operator-scripts\") pod \"barbican-db-create-zmvrc\" (UID: \"d28a1824-1cc9-41be-b4af-1957b982a294\") " pod="openstack/barbican-db-create-zmvrc" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.389691 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-9cfz7"] Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.392231 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9cfz7" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.398354 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vj5mk" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.402251 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95np7\" (UniqueName: \"kubernetes.io/projected/d28a1824-1cc9-41be-b4af-1957b982a294-kube-api-access-95np7\") pod \"barbican-db-create-zmvrc\" (UID: \"d28a1824-1cc9-41be-b4af-1957b982a294\") " pod="openstack/barbican-db-create-zmvrc" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.403197 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9cfz7"] Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.405712 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a458-account-create-update-7cfrj" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.473455 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkfvr\" (UniqueName: \"kubernetes.io/projected/b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8-kube-api-access-mkfvr\") pod \"barbican-69df-account-create-update-2w6f2\" (UID: \"b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8\") " pod="openstack/barbican-69df-account-create-update-2w6f2" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.473538 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8-operator-scripts\") pod \"barbican-69df-account-create-update-2w6f2\" (UID: \"b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8\") " pod="openstack/barbican-69df-account-create-update-2w6f2" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.473601 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e06bc74-47af-4636-95c6-5c326c783d9e-operator-scripts\") pod \"neutron-db-create-9cfz7\" (UID: \"8e06bc74-47af-4636-95c6-5c326c783d9e\") " pod="openstack/neutron-db-create-9cfz7" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.473712 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktq9h\" (UniqueName: \"kubernetes.io/projected/8e06bc74-47af-4636-95c6-5c326c783d9e-kube-api-access-ktq9h\") pod \"neutron-db-create-9cfz7\" (UID: \"8e06bc74-47af-4636-95c6-5c326c783d9e\") " pod="openstack/neutron-db-create-9cfz7" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.474884 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8-operator-scripts\") pod \"barbican-69df-account-create-update-2w6f2\" (UID: \"b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8\") " pod="openstack/barbican-69df-account-create-update-2w6f2" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.488845 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-f69dn"] Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.490533 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-f69dn" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.494452 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.494932 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-856gn" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.495132 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.495313 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.504497 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkfvr\" (UniqueName: \"kubernetes.io/projected/b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8-kube-api-access-mkfvr\") pod \"barbican-69df-account-create-update-2w6f2\" (UID: \"b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8\") " pod="openstack/barbican-69df-account-create-update-2w6f2" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.507414 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-84af-account-create-update-7hbg6"] Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.509080 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84af-account-create-update-7hbg6" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.513815 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.518815 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-84af-account-create-update-7hbg6"] Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.529068 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-f69dn"] Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.532198 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-zmvrc" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.551539 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" event={"ID":"b0f3255a-0cbc-4d2a-bd39-526cbf142fa9","Type":"ContainerDied","Data":"edf8782ebc37c005e6ed6cd2bff03e1d279e7ad9bd4cdcab8c227f055645846f"} Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.551602 4997 scope.go:117] "RemoveContainer" containerID="792bc50db82e23371bcb60395d2f639459b0e1762fba29595170fd9dc47ea22e" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.551800 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5fbdd8c-bp85j" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.575141 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50b8f9d6-7eeb-41ea-8746-76c636b2d9d1-operator-scripts\") pod \"neutron-84af-account-create-update-7hbg6\" (UID: \"50b8f9d6-7eeb-41ea-8746-76c636b2d9d1\") " pod="openstack/neutron-84af-account-create-update-7hbg6" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.575225 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8c2c30d-a149-48f0-9995-6ef963e5f50c-config-data\") pod \"keystone-db-sync-f69dn\" (UID: \"c8c2c30d-a149-48f0-9995-6ef963e5f50c\") " pod="openstack/keystone-db-sync-f69dn" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.575252 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c2c30d-a149-48f0-9995-6ef963e5f50c-combined-ca-bundle\") pod \"keystone-db-sync-f69dn\" (UID: \"c8c2c30d-a149-48f0-9995-6ef963e5f50c\") " pod="openstack/keystone-db-sync-f69dn" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.575309 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nchgn\" (UniqueName: \"kubernetes.io/projected/c8c2c30d-a149-48f0-9995-6ef963e5f50c-kube-api-access-nchgn\") pod \"keystone-db-sync-f69dn\" (UID: \"c8c2c30d-a149-48f0-9995-6ef963e5f50c\") " pod="openstack/keystone-db-sync-f69dn" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.575353 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpnc9\" (UniqueName: \"kubernetes.io/projected/50b8f9d6-7eeb-41ea-8746-76c636b2d9d1-kube-api-access-wpnc9\") pod \"neutron-84af-account-create-update-7hbg6\" (UID: \"50b8f9d6-7eeb-41ea-8746-76c636b2d9d1\") " pod="openstack/neutron-84af-account-create-update-7hbg6" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.575418 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e06bc74-47af-4636-95c6-5c326c783d9e-operator-scripts\") pod \"neutron-db-create-9cfz7\" (UID: \"8e06bc74-47af-4636-95c6-5c326c783d9e\") " pod="openstack/neutron-db-create-9cfz7" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.575454 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktq9h\" (UniqueName: \"kubernetes.io/projected/8e06bc74-47af-4636-95c6-5c326c783d9e-kube-api-access-ktq9h\") pod \"neutron-db-create-9cfz7\" (UID: \"8e06bc74-47af-4636-95c6-5c326c783d9e\") " pod="openstack/neutron-db-create-9cfz7" Dec 05 07:19:03 crc kubenswrapper[4997]: I1205 07:19:03.577180 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e06bc74-47af-4636-95c6-5c326c783d9e-operator-scripts\") pod \"neutron-db-create-9cfz7\" (UID: \"8e06bc74-47af-4636-95c6-5c326c783d9e\") " pod="openstack/neutron-db-create-9cfz7" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.610912 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktq9h\" (UniqueName: \"kubernetes.io/projected/8e06bc74-47af-4636-95c6-5c326c783d9e-kube-api-access-ktq9h\") pod \"neutron-db-create-9cfz7\" (UID: \"8e06bc74-47af-4636-95c6-5c326c783d9e\") " pod="openstack/neutron-db-create-9cfz7" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.612573 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-69df-account-create-update-2w6f2" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.616457 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9cfz7" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.643556 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-bp85j"] Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.644919 4997 scope.go:117] "RemoveContainer" containerID="34d43de0b1e528710751b761feea30add9f37a9a1041cf3c2f563ea7891de52a" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.651804 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-bp85j"] Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.679798 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8c2c30d-a149-48f0-9995-6ef963e5f50c-config-data\") pod \"keystone-db-sync-f69dn\" (UID: \"c8c2c30d-a149-48f0-9995-6ef963e5f50c\") " pod="openstack/keystone-db-sync-f69dn" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.679840 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c2c30d-a149-48f0-9995-6ef963e5f50c-combined-ca-bundle\") pod \"keystone-db-sync-f69dn\" (UID: \"c8c2c30d-a149-48f0-9995-6ef963e5f50c\") " pod="openstack/keystone-db-sync-f69dn" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.679890 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nchgn\" (UniqueName: \"kubernetes.io/projected/c8c2c30d-a149-48f0-9995-6ef963e5f50c-kube-api-access-nchgn\") pod \"keystone-db-sync-f69dn\" (UID: \"c8c2c30d-a149-48f0-9995-6ef963e5f50c\") " pod="openstack/keystone-db-sync-f69dn" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.679922 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpnc9\" (UniqueName: \"kubernetes.io/projected/50b8f9d6-7eeb-41ea-8746-76c636b2d9d1-kube-api-access-wpnc9\") pod \"neutron-84af-account-create-update-7hbg6\" (UID: \"50b8f9d6-7eeb-41ea-8746-76c636b2d9d1\") " pod="openstack/neutron-84af-account-create-update-7hbg6" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.680026 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50b8f9d6-7eeb-41ea-8746-76c636b2d9d1-operator-scripts\") pod \"neutron-84af-account-create-update-7hbg6\" (UID: \"50b8f9d6-7eeb-41ea-8746-76c636b2d9d1\") " pod="openstack/neutron-84af-account-create-update-7hbg6" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.682494 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50b8f9d6-7eeb-41ea-8746-76c636b2d9d1-operator-scripts\") pod \"neutron-84af-account-create-update-7hbg6\" (UID: \"50b8f9d6-7eeb-41ea-8746-76c636b2d9d1\") " pod="openstack/neutron-84af-account-create-update-7hbg6" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.685296 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c2c30d-a149-48f0-9995-6ef963e5f50c-combined-ca-bundle\") pod \"keystone-db-sync-f69dn\" (UID: \"c8c2c30d-a149-48f0-9995-6ef963e5f50c\") " pod="openstack/keystone-db-sync-f69dn" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.685323 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8c2c30d-a149-48f0-9995-6ef963e5f50c-config-data\") pod \"keystone-db-sync-f69dn\" (UID: \"c8c2c30d-a149-48f0-9995-6ef963e5f50c\") " pod="openstack/keystone-db-sync-f69dn" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.698034 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nchgn\" (UniqueName: \"kubernetes.io/projected/c8c2c30d-a149-48f0-9995-6ef963e5f50c-kube-api-access-nchgn\") pod \"keystone-db-sync-f69dn\" (UID: \"c8c2c30d-a149-48f0-9995-6ef963e5f50c\") " pod="openstack/keystone-db-sync-f69dn" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.704879 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpnc9\" (UniqueName: \"kubernetes.io/projected/50b8f9d6-7eeb-41ea-8746-76c636b2d9d1-kube-api-access-wpnc9\") pod \"neutron-84af-account-create-update-7hbg6\" (UID: \"50b8f9d6-7eeb-41ea-8746-76c636b2d9d1\") " pod="openstack/neutron-84af-account-create-update-7hbg6" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.768052 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0f3255a-0cbc-4d2a-bd39-526cbf142fa9" path="/var/lib/kubelet/pods/b0f3255a-0cbc-4d2a-bd39-526cbf142fa9/volumes" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.927853 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-f69dn" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:03.939767 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84af-account-create-update-7hbg6" Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:04.725156 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-zmvrc"] Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:04.744795 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9cfz7"] Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:04.758455 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-69df-account-create-update-2w6f2"] Dec 05 07:19:04 crc kubenswrapper[4997]: W1205 07:19:04.771156 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd28a1824_1cc9_41be_b4af_1957b982a294.slice/crio-d2014520d36b5e477025bd2e5f8a49fd11656c3f32f51dcb54874caeb46bf9ca WatchSource:0}: Error finding container d2014520d36b5e477025bd2e5f8a49fd11656c3f32f51dcb54874caeb46bf9ca: Status 404 returned error can't find the container with id d2014520d36b5e477025bd2e5f8a49fd11656c3f32f51dcb54874caeb46bf9ca Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:04.776932 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vj5mk"] Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:04.784755 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-a458-account-create-update-7cfrj"] Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:04.792179 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-f69dn"] Dec 05 07:19:04 crc kubenswrapper[4997]: I1205 07:19:04.799160 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-84af-account-create-update-7hbg6"] Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.591604 4997 generic.go:334] "Generic (PLEG): container finished" podID="d28a1824-1cc9-41be-b4af-1957b982a294" containerID="c9360bc2fe85b0fb890b926fa67438af8aa992628a1b4a88b705b4c8f00b90ea" exitCode=0 Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.592232 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-zmvrc" event={"ID":"d28a1824-1cc9-41be-b4af-1957b982a294","Type":"ContainerDied","Data":"c9360bc2fe85b0fb890b926fa67438af8aa992628a1b4a88b705b4c8f00b90ea"} Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.592271 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-zmvrc" event={"ID":"d28a1824-1cc9-41be-b4af-1957b982a294","Type":"ContainerStarted","Data":"d2014520d36b5e477025bd2e5f8a49fd11656c3f32f51dcb54874caeb46bf9ca"} Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.602821 4997 generic.go:334] "Generic (PLEG): container finished" podID="b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8" containerID="4ec3032847e09b12c6c9fedc21149ebbebb9600b8ed37fb546deca7c3f8f58c7" exitCode=0 Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.602875 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-69df-account-create-update-2w6f2" event={"ID":"b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8","Type":"ContainerDied","Data":"4ec3032847e09b12c6c9fedc21149ebbebb9600b8ed37fb546deca7c3f8f58c7"} Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.603032 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-69df-account-create-update-2w6f2" event={"ID":"b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8","Type":"ContainerStarted","Data":"4115db523b5b23eca8baff8841c5a97162206865584dd6e1f60ecc081b284e3b"} Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.606279 4997 generic.go:334] "Generic (PLEG): container finished" podID="8e06bc74-47af-4636-95c6-5c326c783d9e" containerID="7e3740ba6a3041381dab8f67d89402db7bd8d90cf3554259e64bd2cbc184a290" exitCode=0 Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.606564 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9cfz7" event={"ID":"8e06bc74-47af-4636-95c6-5c326c783d9e","Type":"ContainerDied","Data":"7e3740ba6a3041381dab8f67d89402db7bd8d90cf3554259e64bd2cbc184a290"} Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.606606 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9cfz7" event={"ID":"8e06bc74-47af-4636-95c6-5c326c783d9e","Type":"ContainerStarted","Data":"061ddc62dce11d002ffb449d2fcb4ea677d57aeb8707d3c8fe4eb9e5d6bab645"} Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.609766 4997 generic.go:334] "Generic (PLEG): container finished" podID="6f0b9769-3e80-4475-b8d3-5fe6c5a265e7" containerID="ebb950a103ad8b9026183650a563fd260355285301d70a1c13b4d592f483a882" exitCode=0 Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.609820 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vj5mk" event={"ID":"6f0b9769-3e80-4475-b8d3-5fe6c5a265e7","Type":"ContainerDied","Data":"ebb950a103ad8b9026183650a563fd260355285301d70a1c13b4d592f483a882"} Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.609840 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vj5mk" event={"ID":"6f0b9769-3e80-4475-b8d3-5fe6c5a265e7","Type":"ContainerStarted","Data":"fd766d96194c9a4e98563c9d955f63dfeb39aaabc826d70eec3d9c5756b07e3e"} Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.616066 4997 generic.go:334] "Generic (PLEG): container finished" podID="4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3" containerID="643884fd8f6b17c03dc7c861985fbb559e530326ef051b82964745eadd0773de" exitCode=0 Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.616419 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-a458-account-create-update-7cfrj" event={"ID":"4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3","Type":"ContainerDied","Data":"643884fd8f6b17c03dc7c861985fbb559e530326ef051b82964745eadd0773de"} Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.616459 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-a458-account-create-update-7cfrj" event={"ID":"4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3","Type":"ContainerStarted","Data":"27dc41becaac9a0171a636e216ad2d964481e22d1c0ac091f53e3dc197df241b"} Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.619117 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-f69dn" event={"ID":"c8c2c30d-a149-48f0-9995-6ef963e5f50c","Type":"ContainerStarted","Data":"445ae07424276767f4467d6cfa8a1e4cee1a2b715e047d2048c25d58673063d5"} Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.635025 4997 generic.go:334] "Generic (PLEG): container finished" podID="50b8f9d6-7eeb-41ea-8746-76c636b2d9d1" containerID="4f26581bd5f07af968fd994ea8a57ae49455eb186f958e5e52364d251398e21c" exitCode=0 Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.635115 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84af-account-create-update-7hbg6" event={"ID":"50b8f9d6-7eeb-41ea-8746-76c636b2d9d1","Type":"ContainerDied","Data":"4f26581bd5f07af968fd994ea8a57ae49455eb186f958e5e52364d251398e21c"} Dec 05 07:19:05 crc kubenswrapper[4997]: I1205 07:19:05.635191 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84af-account-create-update-7hbg6" event={"ID":"50b8f9d6-7eeb-41ea-8746-76c636b2d9d1","Type":"ContainerStarted","Data":"4c769b4ac4e00ffca484623f9a785e91ad164d6e392ac6854ee4f8f54c3a7159"} Dec 05 07:19:07 crc kubenswrapper[4997]: I1205 07:19:07.664386 4997 generic.go:334] "Generic (PLEG): container finished" podID="cdf7e311-569b-43a2-bba8-729226e28415" containerID="5e883daffbf9da9e1ebaee0593862fba2d57d7337e7c8104926342e03ff1d1d6" exitCode=0 Dec 05 07:19:07 crc kubenswrapper[4997]: I1205 07:19:07.664424 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6bdvj" event={"ID":"cdf7e311-569b-43a2-bba8-729226e28415","Type":"ContainerDied","Data":"5e883daffbf9da9e1ebaee0593862fba2d57d7337e7c8104926342e03ff1d1d6"} Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.391088 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84af-account-create-update-7hbg6" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.401104 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vj5mk" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.439392 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a458-account-create-update-7cfrj" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.447042 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-zmvrc" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.467336 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9cfz7" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.469408 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6bdvj" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.482067 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-69df-account-create-update-2w6f2" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.509845 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50b8f9d6-7eeb-41ea-8746-76c636b2d9d1-operator-scripts\") pod \"50b8f9d6-7eeb-41ea-8746-76c636b2d9d1\" (UID: \"50b8f9d6-7eeb-41ea-8746-76c636b2d9d1\") " Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.510078 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpnc9\" (UniqueName: \"kubernetes.io/projected/50b8f9d6-7eeb-41ea-8746-76c636b2d9d1-kube-api-access-wpnc9\") pod \"50b8f9d6-7eeb-41ea-8746-76c636b2d9d1\" (UID: \"50b8f9d6-7eeb-41ea-8746-76c636b2d9d1\") " Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.516817 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50b8f9d6-7eeb-41ea-8746-76c636b2d9d1-kube-api-access-wpnc9" (OuterVolumeSpecName: "kube-api-access-wpnc9") pod "50b8f9d6-7eeb-41ea-8746-76c636b2d9d1" (UID: "50b8f9d6-7eeb-41ea-8746-76c636b2d9d1"). InnerVolumeSpecName "kube-api-access-wpnc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.524710 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50b8f9d6-7eeb-41ea-8746-76c636b2d9d1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "50b8f9d6-7eeb-41ea-8746-76c636b2d9d1" (UID: "50b8f9d6-7eeb-41ea-8746-76c636b2d9d1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.612208 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95np7\" (UniqueName: \"kubernetes.io/projected/d28a1824-1cc9-41be-b4af-1957b982a294-kube-api-access-95np7\") pod \"d28a1824-1cc9-41be-b4af-1957b982a294\" (UID: \"d28a1824-1cc9-41be-b4af-1957b982a294\") " Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.612265 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-config-data\") pod \"cdf7e311-569b-43a2-bba8-729226e28415\" (UID: \"cdf7e311-569b-43a2-bba8-729226e28415\") " Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.612309 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3-operator-scripts\") pod \"4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3\" (UID: \"4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3\") " Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.612332 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f0b9769-3e80-4475-b8d3-5fe6c5a265e7-operator-scripts\") pod \"6f0b9769-3e80-4475-b8d3-5fe6c5a265e7\" (UID: \"6f0b9769-3e80-4475-b8d3-5fe6c5a265e7\") " Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.612429 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-db-sync-config-data\") pod \"cdf7e311-569b-43a2-bba8-729226e28415\" (UID: \"cdf7e311-569b-43a2-bba8-729226e28415\") " Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.612466 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8-operator-scripts\") pod \"b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8\" (UID: \"b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8\") " Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.612504 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e06bc74-47af-4636-95c6-5c326c783d9e-operator-scripts\") pod \"8e06bc74-47af-4636-95c6-5c326c783d9e\" (UID: \"8e06bc74-47af-4636-95c6-5c326c783d9e\") " Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.612552 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-combined-ca-bundle\") pod \"cdf7e311-569b-43a2-bba8-729226e28415\" (UID: \"cdf7e311-569b-43a2-bba8-729226e28415\") " Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.612575 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-667kb\" (UniqueName: \"kubernetes.io/projected/6f0b9769-3e80-4475-b8d3-5fe6c5a265e7-kube-api-access-667kb\") pod \"6f0b9769-3e80-4475-b8d3-5fe6c5a265e7\" (UID: \"6f0b9769-3e80-4475-b8d3-5fe6c5a265e7\") " Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.612600 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d28a1824-1cc9-41be-b4af-1957b982a294-operator-scripts\") pod \"d28a1824-1cc9-41be-b4af-1957b982a294\" (UID: \"d28a1824-1cc9-41be-b4af-1957b982a294\") " Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.612675 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrrqz\" (UniqueName: \"kubernetes.io/projected/cdf7e311-569b-43a2-bba8-729226e28415-kube-api-access-nrrqz\") pod \"cdf7e311-569b-43a2-bba8-729226e28415\" (UID: \"cdf7e311-569b-43a2-bba8-729226e28415\") " Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.612729 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbc5v\" (UniqueName: \"kubernetes.io/projected/4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3-kube-api-access-mbc5v\") pod \"4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3\" (UID: \"4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3\") " Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.612762 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktq9h\" (UniqueName: \"kubernetes.io/projected/8e06bc74-47af-4636-95c6-5c326c783d9e-kube-api-access-ktq9h\") pod \"8e06bc74-47af-4636-95c6-5c326c783d9e\" (UID: \"8e06bc74-47af-4636-95c6-5c326c783d9e\") " Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.612811 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkfvr\" (UniqueName: \"kubernetes.io/projected/b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8-kube-api-access-mkfvr\") pod \"b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8\" (UID: \"b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8\") " Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.613182 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50b8f9d6-7eeb-41ea-8746-76c636b2d9d1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.613202 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpnc9\" (UniqueName: \"kubernetes.io/projected/50b8f9d6-7eeb-41ea-8746-76c636b2d9d1-kube-api-access-wpnc9\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.613404 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8" (UID: "b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.613478 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f0b9769-3e80-4475-b8d3-5fe6c5a265e7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6f0b9769-3e80-4475-b8d3-5fe6c5a265e7" (UID: "6f0b9769-3e80-4475-b8d3-5fe6c5a265e7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.614103 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d28a1824-1cc9-41be-b4af-1957b982a294-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d28a1824-1cc9-41be-b4af-1957b982a294" (UID: "d28a1824-1cc9-41be-b4af-1957b982a294"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.614368 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3" (UID: "4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.614843 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e06bc74-47af-4636-95c6-5c326c783d9e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8e06bc74-47af-4636-95c6-5c326c783d9e" (UID: "8e06bc74-47af-4636-95c6-5c326c783d9e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.617725 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3-kube-api-access-mbc5v" (OuterVolumeSpecName: "kube-api-access-mbc5v") pod "4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3" (UID: "4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3"). InnerVolumeSpecName "kube-api-access-mbc5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.617721 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "cdf7e311-569b-43a2-bba8-729226e28415" (UID: "cdf7e311-569b-43a2-bba8-729226e28415"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.618036 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d28a1824-1cc9-41be-b4af-1957b982a294-kube-api-access-95np7" (OuterVolumeSpecName: "kube-api-access-95np7") pod "d28a1824-1cc9-41be-b4af-1957b982a294" (UID: "d28a1824-1cc9-41be-b4af-1957b982a294"). InnerVolumeSpecName "kube-api-access-95np7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.619262 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e06bc74-47af-4636-95c6-5c326c783d9e-kube-api-access-ktq9h" (OuterVolumeSpecName: "kube-api-access-ktq9h") pod "8e06bc74-47af-4636-95c6-5c326c783d9e" (UID: "8e06bc74-47af-4636-95c6-5c326c783d9e"). InnerVolumeSpecName "kube-api-access-ktq9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.620582 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdf7e311-569b-43a2-bba8-729226e28415-kube-api-access-nrrqz" (OuterVolumeSpecName: "kube-api-access-nrrqz") pod "cdf7e311-569b-43a2-bba8-729226e28415" (UID: "cdf7e311-569b-43a2-bba8-729226e28415"). InnerVolumeSpecName "kube-api-access-nrrqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.620646 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8-kube-api-access-mkfvr" (OuterVolumeSpecName: "kube-api-access-mkfvr") pod "b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8" (UID: "b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8"). InnerVolumeSpecName "kube-api-access-mkfvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.621266 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f0b9769-3e80-4475-b8d3-5fe6c5a265e7-kube-api-access-667kb" (OuterVolumeSpecName: "kube-api-access-667kb") pod "6f0b9769-3e80-4475-b8d3-5fe6c5a265e7" (UID: "6f0b9769-3e80-4475-b8d3-5fe6c5a265e7"). InnerVolumeSpecName "kube-api-access-667kb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.641963 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cdf7e311-569b-43a2-bba8-729226e28415" (UID: "cdf7e311-569b-43a2-bba8-729226e28415"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.663444 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-config-data" (OuterVolumeSpecName: "config-data") pod "cdf7e311-569b-43a2-bba8-729226e28415" (UID: "cdf7e311-569b-43a2-bba8-729226e28415"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.706885 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9cfz7" event={"ID":"8e06bc74-47af-4636-95c6-5c326c783d9e","Type":"ContainerDied","Data":"061ddc62dce11d002ffb449d2fcb4ea677d57aeb8707d3c8fe4eb9e5d6bab645"} Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.706956 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="061ddc62dce11d002ffb449d2fcb4ea677d57aeb8707d3c8fe4eb9e5d6bab645" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.706916 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9cfz7" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.714144 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrrqz\" (UniqueName: \"kubernetes.io/projected/cdf7e311-569b-43a2-bba8-729226e28415-kube-api-access-nrrqz\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.714178 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbc5v\" (UniqueName: \"kubernetes.io/projected/4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3-kube-api-access-mbc5v\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.714191 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktq9h\" (UniqueName: \"kubernetes.io/projected/8e06bc74-47af-4636-95c6-5c326c783d9e-kube-api-access-ktq9h\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.714204 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkfvr\" (UniqueName: \"kubernetes.io/projected/b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8-kube-api-access-mkfvr\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.714215 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95np7\" (UniqueName: \"kubernetes.io/projected/d28a1824-1cc9-41be-b4af-1957b982a294-kube-api-access-95np7\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.714227 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.714239 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.714251 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f0b9769-3e80-4475-b8d3-5fe6c5a265e7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.714262 4997 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.714362 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.714373 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e06bc74-47af-4636-95c6-5c326c783d9e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.714385 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf7e311-569b-43a2-bba8-729226e28415-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.714396 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-667kb\" (UniqueName: \"kubernetes.io/projected/6f0b9769-3e80-4475-b8d3-5fe6c5a265e7-kube-api-access-667kb\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.714408 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d28a1824-1cc9-41be-b4af-1957b982a294-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.716886 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vj5mk" event={"ID":"6f0b9769-3e80-4475-b8d3-5fe6c5a265e7","Type":"ContainerDied","Data":"fd766d96194c9a4e98563c9d955f63dfeb39aaabc826d70eec3d9c5756b07e3e"} Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.716936 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vj5mk" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.716962 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd766d96194c9a4e98563c9d955f63dfeb39aaabc826d70eec3d9c5756b07e3e" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.724858 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-a458-account-create-update-7cfrj" event={"ID":"4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3","Type":"ContainerDied","Data":"27dc41becaac9a0171a636e216ad2d964481e22d1c0ac091f53e3dc197df241b"} Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.724904 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a458-account-create-update-7cfrj" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.724960 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27dc41becaac9a0171a636e216ad2d964481e22d1c0ac091f53e3dc197df241b" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.728671 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-f69dn" event={"ID":"c8c2c30d-a149-48f0-9995-6ef963e5f50c","Type":"ContainerStarted","Data":"0f86635af561a3a610b8f8d117854e4f529405ad3fb7080d7a0116d8b95d0876"} Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.730404 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84af-account-create-update-7hbg6" event={"ID":"50b8f9d6-7eeb-41ea-8746-76c636b2d9d1","Type":"ContainerDied","Data":"4c769b4ac4e00ffca484623f9a785e91ad164d6e392ac6854ee4f8f54c3a7159"} Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.730468 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c769b4ac4e00ffca484623f9a785e91ad164d6e392ac6854ee4f8f54c3a7159" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.730496 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84af-account-create-update-7hbg6" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.732656 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6bdvj" event={"ID":"cdf7e311-569b-43a2-bba8-729226e28415","Type":"ContainerDied","Data":"6e8d450d143a8fda737a6916a193e79a26300ba137ac29950355ad0e49f4618a"} Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.732683 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e8d450d143a8fda737a6916a193e79a26300ba137ac29950355ad0e49f4618a" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.732735 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6bdvj" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.734095 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-zmvrc" event={"ID":"d28a1824-1cc9-41be-b4af-1957b982a294","Type":"ContainerDied","Data":"d2014520d36b5e477025bd2e5f8a49fd11656c3f32f51dcb54874caeb46bf9ca"} Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.734144 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2014520d36b5e477025bd2e5f8a49fd11656c3f32f51dcb54874caeb46bf9ca" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.734224 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-zmvrc" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.735149 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-69df-account-create-update-2w6f2" event={"ID":"b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8","Type":"ContainerDied","Data":"4115db523b5b23eca8baff8841c5a97162206865584dd6e1f60ecc081b284e3b"} Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.735166 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4115db523b5b23eca8baff8841c5a97162206865584dd6e1f60ecc081b284e3b" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.735224 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-69df-account-create-update-2w6f2" Dec 05 07:19:09 crc kubenswrapper[4997]: I1205 07:19:09.749630 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-f69dn" podStartSLOduration=2.299649604 podStartE2EDuration="6.749582617s" podCreationTimestamp="2025-12-05 07:19:03 +0000 UTC" firstStartedPulling="2025-12-05 07:19:04.800342391 +0000 UTC m=+1445.329249662" lastFinishedPulling="2025-12-05 07:19:09.250275414 +0000 UTC m=+1449.779182675" observedRunningTime="2025-12-05 07:19:09.748956041 +0000 UTC m=+1450.277863312" watchObservedRunningTime="2025-12-05 07:19:09.749582617 +0000 UTC m=+1450.278489878" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.189965 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d84db446f-m9lt2"] Dec 05 07:19:10 crc kubenswrapper[4997]: E1205 07:19:10.190683 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8" containerName="mariadb-account-create-update" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.190712 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8" containerName="mariadb-account-create-update" Dec 05 07:19:10 crc kubenswrapper[4997]: E1205 07:19:10.190735 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b8f9d6-7eeb-41ea-8746-76c636b2d9d1" containerName="mariadb-account-create-update" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.190744 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b8f9d6-7eeb-41ea-8746-76c636b2d9d1" containerName="mariadb-account-create-update" Dec 05 07:19:10 crc kubenswrapper[4997]: E1205 07:19:10.190753 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f0b9769-3e80-4475-b8d3-5fe6c5a265e7" containerName="mariadb-database-create" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.190761 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f0b9769-3e80-4475-b8d3-5fe6c5a265e7" containerName="mariadb-database-create" Dec 05 07:19:10 crc kubenswrapper[4997]: E1205 07:19:10.190780 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3" containerName="mariadb-account-create-update" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.190790 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3" containerName="mariadb-account-create-update" Dec 05 07:19:10 crc kubenswrapper[4997]: E1205 07:19:10.190816 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdf7e311-569b-43a2-bba8-729226e28415" containerName="glance-db-sync" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.190825 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdf7e311-569b-43a2-bba8-729226e28415" containerName="glance-db-sync" Dec 05 07:19:10 crc kubenswrapper[4997]: E1205 07:19:10.190847 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e06bc74-47af-4636-95c6-5c326c783d9e" containerName="mariadb-database-create" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.190855 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e06bc74-47af-4636-95c6-5c326c783d9e" containerName="mariadb-database-create" Dec 05 07:19:10 crc kubenswrapper[4997]: E1205 07:19:10.190878 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d28a1824-1cc9-41be-b4af-1957b982a294" containerName="mariadb-database-create" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.190894 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d28a1824-1cc9-41be-b4af-1957b982a294" containerName="mariadb-database-create" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.191129 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdf7e311-569b-43a2-bba8-729226e28415" containerName="glance-db-sync" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.191147 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="d28a1824-1cc9-41be-b4af-1957b982a294" containerName="mariadb-database-create" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.191165 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8" containerName="mariadb-account-create-update" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.191186 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f0b9769-3e80-4475-b8d3-5fe6c5a265e7" containerName="mariadb-database-create" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.191199 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b8f9d6-7eeb-41ea-8746-76c636b2d9d1" containerName="mariadb-account-create-update" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.191214 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3" containerName="mariadb-account-create-update" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.191228 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e06bc74-47af-4636-95c6-5c326c783d9e" containerName="mariadb-database-create" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.192524 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.198335 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84db446f-m9lt2"] Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.326950 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.327067 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.327113 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-dns-svc\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.327137 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbtl2\" (UniqueName: \"kubernetes.io/projected/7f350f7f-80f2-4329-8b61-cac9cdfeb676-kube-api-access-qbtl2\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.327171 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-config\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.327216 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.428986 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.430773 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.431304 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.432115 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.432374 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.432476 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-dns-svc\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.432576 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbtl2\" (UniqueName: \"kubernetes.io/projected/7f350f7f-80f2-4329-8b61-cac9cdfeb676-kube-api-access-qbtl2\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.432724 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-config\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.433367 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.433580 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-config\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.433673 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-dns-svc\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.465930 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbtl2\" (UniqueName: \"kubernetes.io/projected/7f350f7f-80f2-4329-8b61-cac9cdfeb676-kube-api-access-qbtl2\") pod \"dnsmasq-dns-7d84db446f-m9lt2\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:10 crc kubenswrapper[4997]: I1205 07:19:10.564034 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:11 crc kubenswrapper[4997]: I1205 07:19:11.046322 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84db446f-m9lt2"] Dec 05 07:19:11 crc kubenswrapper[4997]: I1205 07:19:11.757881 4997 generic.go:334] "Generic (PLEG): container finished" podID="7f350f7f-80f2-4329-8b61-cac9cdfeb676" containerID="6097c98b4bc5bbf756341e4b34764fca582ce25906d164d635e716e0334eabc6" exitCode=0 Dec 05 07:19:11 crc kubenswrapper[4997]: I1205 07:19:11.763344 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" event={"ID":"7f350f7f-80f2-4329-8b61-cac9cdfeb676","Type":"ContainerDied","Data":"6097c98b4bc5bbf756341e4b34764fca582ce25906d164d635e716e0334eabc6"} Dec 05 07:19:11 crc kubenswrapper[4997]: I1205 07:19:11.763399 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" event={"ID":"7f350f7f-80f2-4329-8b61-cac9cdfeb676","Type":"ContainerStarted","Data":"83c2b0a20b78bfe6f81db74ec9f1d79c1fc895512468380ead3b2b944f99cee0"} Dec 05 07:19:12 crc kubenswrapper[4997]: I1205 07:19:12.771213 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" event={"ID":"7f350f7f-80f2-4329-8b61-cac9cdfeb676","Type":"ContainerStarted","Data":"ec17cb996774103234482c6620417bc620c0143ead85111450d1cd12dd354eff"} Dec 05 07:19:12 crc kubenswrapper[4997]: I1205 07:19:12.771935 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:12 crc kubenswrapper[4997]: I1205 07:19:12.772875 4997 generic.go:334] "Generic (PLEG): container finished" podID="c8c2c30d-a149-48f0-9995-6ef963e5f50c" containerID="0f86635af561a3a610b8f8d117854e4f529405ad3fb7080d7a0116d8b95d0876" exitCode=0 Dec 05 07:19:12 crc kubenswrapper[4997]: I1205 07:19:12.772908 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-f69dn" event={"ID":"c8c2c30d-a149-48f0-9995-6ef963e5f50c","Type":"ContainerDied","Data":"0f86635af561a3a610b8f8d117854e4f529405ad3fb7080d7a0116d8b95d0876"} Dec 05 07:19:12 crc kubenswrapper[4997]: I1205 07:19:12.805114 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" podStartSLOduration=2.805086469 podStartE2EDuration="2.805086469s" podCreationTimestamp="2025-12-05 07:19:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:19:12.797570696 +0000 UTC m=+1453.326477957" watchObservedRunningTime="2025-12-05 07:19:12.805086469 +0000 UTC m=+1453.333993740" Dec 05 07:19:14 crc kubenswrapper[4997]: I1205 07:19:14.117358 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-f69dn" Dec 05 07:19:14 crc kubenswrapper[4997]: I1205 07:19:14.243433 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c2c30d-a149-48f0-9995-6ef963e5f50c-combined-ca-bundle\") pod \"c8c2c30d-a149-48f0-9995-6ef963e5f50c\" (UID: \"c8c2c30d-a149-48f0-9995-6ef963e5f50c\") " Dec 05 07:19:14 crc kubenswrapper[4997]: I1205 07:19:14.243511 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nchgn\" (UniqueName: \"kubernetes.io/projected/c8c2c30d-a149-48f0-9995-6ef963e5f50c-kube-api-access-nchgn\") pod \"c8c2c30d-a149-48f0-9995-6ef963e5f50c\" (UID: \"c8c2c30d-a149-48f0-9995-6ef963e5f50c\") " Dec 05 07:19:14 crc kubenswrapper[4997]: I1205 07:19:14.243708 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8c2c30d-a149-48f0-9995-6ef963e5f50c-config-data\") pod \"c8c2c30d-a149-48f0-9995-6ef963e5f50c\" (UID: \"c8c2c30d-a149-48f0-9995-6ef963e5f50c\") " Dec 05 07:19:14 crc kubenswrapper[4997]: I1205 07:19:14.253897 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8c2c30d-a149-48f0-9995-6ef963e5f50c-kube-api-access-nchgn" (OuterVolumeSpecName: "kube-api-access-nchgn") pod "c8c2c30d-a149-48f0-9995-6ef963e5f50c" (UID: "c8c2c30d-a149-48f0-9995-6ef963e5f50c"). InnerVolumeSpecName "kube-api-access-nchgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:14 crc kubenswrapper[4997]: I1205 07:19:14.288738 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8c2c30d-a149-48f0-9995-6ef963e5f50c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8c2c30d-a149-48f0-9995-6ef963e5f50c" (UID: "c8c2c30d-a149-48f0-9995-6ef963e5f50c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:14 crc kubenswrapper[4997]: I1205 07:19:14.340843 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8c2c30d-a149-48f0-9995-6ef963e5f50c-config-data" (OuterVolumeSpecName: "config-data") pod "c8c2c30d-a149-48f0-9995-6ef963e5f50c" (UID: "c8c2c30d-a149-48f0-9995-6ef963e5f50c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:14 crc kubenswrapper[4997]: I1205 07:19:14.349092 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c2c30d-a149-48f0-9995-6ef963e5f50c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:14 crc kubenswrapper[4997]: I1205 07:19:14.349140 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nchgn\" (UniqueName: \"kubernetes.io/projected/c8c2c30d-a149-48f0-9995-6ef963e5f50c-kube-api-access-nchgn\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:14 crc kubenswrapper[4997]: I1205 07:19:14.349153 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8c2c30d-a149-48f0-9995-6ef963e5f50c-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:14 crc kubenswrapper[4997]: I1205 07:19:14.810058 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-f69dn" event={"ID":"c8c2c30d-a149-48f0-9995-6ef963e5f50c","Type":"ContainerDied","Data":"445ae07424276767f4467d6cfa8a1e4cee1a2b715e047d2048c25d58673063d5"} Dec 05 07:19:14 crc kubenswrapper[4997]: I1205 07:19:14.810184 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="445ae07424276767f4467d6cfa8a1e4cee1a2b715e047d2048c25d58673063d5" Dec 05 07:19:14 crc kubenswrapper[4997]: I1205 07:19:14.810355 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-f69dn" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.137854 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84db446f-m9lt2"] Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.138181 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" podUID="7f350f7f-80f2-4329-8b61-cac9cdfeb676" containerName="dnsmasq-dns" containerID="cri-o://ec17cb996774103234482c6620417bc620c0143ead85111450d1cd12dd354eff" gracePeriod=10 Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.155677 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-srgk8"] Dec 05 07:19:15 crc kubenswrapper[4997]: E1205 07:19:15.156964 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8c2c30d-a149-48f0-9995-6ef963e5f50c" containerName="keystone-db-sync" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.156990 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8c2c30d-a149-48f0-9995-6ef963e5f50c" containerName="keystone-db-sync" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.157273 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8c2c30d-a149-48f0-9995-6ef963e5f50c" containerName="keystone-db-sync" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.158044 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.164053 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-srgk8"] Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.167403 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.167439 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.167479 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.168534 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-856gn" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.168686 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.229191 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76cc8cc9bc-cdcn9"] Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.254350 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76cc8cc9bc-cdcn9"] Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.254504 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.281694 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-fernet-keys\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.281756 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-config-data\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.281787 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qslqt\" (UniqueName: \"kubernetes.io/projected/42a735ac-395f-4357-9db3-1aff67aef6fa-kube-api-access-qslqt\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.281842 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-combined-ca-bundle\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.281877 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-credential-keys\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.281901 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-scripts\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.384302 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-ovsdbserver-sb\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.384366 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-combined-ca-bundle\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.384408 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-credential-keys\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.384441 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-scripts\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.384521 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-config\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.384547 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-dns-svc\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.384730 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-fernet-keys\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.384761 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-ovsdbserver-nb\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.384782 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-dns-swift-storage-0\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.384800 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-config-data\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.384825 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qslqt\" (UniqueName: \"kubernetes.io/projected/42a735ac-395f-4357-9db3-1aff67aef6fa-kube-api-access-qslqt\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.384861 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvlms\" (UniqueName: \"kubernetes.io/projected/fe20e521-eeb9-4f15-84d1-925c0f34126d-kube-api-access-lvlms\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.411190 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-scripts\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.411229 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-combined-ca-bundle\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.413974 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-fernet-keys\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.428843 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-92zcp"] Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.430267 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-config-data\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.439057 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-credential-keys\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.448267 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-92zcp" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.452684 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.452906 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-lgbp8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.454359 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qslqt\" (UniqueName: \"kubernetes.io/projected/42a735ac-395f-4357-9db3-1aff67aef6fa-kube-api-access-qslqt\") pod \"keystone-bootstrap-srgk8\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.456220 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.469785 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-92zcp"] Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.477299 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-k8cjm"] Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.478971 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.481780 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.483092 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.489661 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-ovsdbserver-sb\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.489804 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-config\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.489837 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-dns-svc\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.489888 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-dns-swift-storage-0\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.489905 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-ovsdbserver-nb\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.489941 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvlms\" (UniqueName: \"kubernetes.io/projected/fe20e521-eeb9-4f15-84d1-925c0f34126d-kube-api-access-lvlms\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.491331 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-ovsdbserver-sb\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.491930 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-config\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.493356 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-k8cjm"] Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.493471 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-dns-swift-storage-0\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.494065 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-dns-svc\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.494228 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-ovsdbserver-nb\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.498039 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-4jgwm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.498425 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.535481 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvlms\" (UniqueName: \"kubernetes.io/projected/fe20e521-eeb9-4f15-84d1-925c0f34126d-kube-api-access-lvlms\") pod \"dnsmasq-dns-76cc8cc9bc-cdcn9\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.575776 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76cc8cc9bc-cdcn9"] Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.576997 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.591433 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-combined-ca-bundle\") pod \"neutron-db-sync-92zcp\" (UID: \"1e522d4c-e094-4518-bdb5-8ad8e3eccc97\") " pod="openstack/neutron-db-sync-92zcp" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.591527 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-scripts\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.591588 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m75ld\" (UniqueName: \"kubernetes.io/projected/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-kube-api-access-m75ld\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.591672 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-combined-ca-bundle\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.597633 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-etc-machine-id\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.597703 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-db-sync-config-data\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.597724 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-config\") pod \"neutron-db-sync-92zcp\" (UID: \"1e522d4c-e094-4518-bdb5-8ad8e3eccc97\") " pod="openstack/neutron-db-sync-92zcp" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.597752 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-config-data\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.597903 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvhpv\" (UniqueName: \"kubernetes.io/projected/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-kube-api-access-jvhpv\") pod \"neutron-db-sync-92zcp\" (UID: \"1e522d4c-e094-4518-bdb5-8ad8e3eccc97\") " pod="openstack/neutron-db-sync-92zcp" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.614698 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-svg9r"] Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.616358 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.629279 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-svg9r"] Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.640111 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-h79p5"] Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.641530 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-h79p5" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.643341 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.647786 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-jmntt" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.650770 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-h79p5"] Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.682975 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-6lm2t"] Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.684948 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.689119 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.689349 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.689511 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-86x2z" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.702734 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m75ld\" (UniqueName: \"kubernetes.io/projected/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-kube-api-access-m75ld\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.702807 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-combined-ca-bundle\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.702856 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-etc-machine-id\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.702878 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-db-sync-config-data\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.702896 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-config\") pod \"neutron-db-sync-92zcp\" (UID: \"1e522d4c-e094-4518-bdb5-8ad8e3eccc97\") " pod="openstack/neutron-db-sync-92zcp" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.702932 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-config-data\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.703001 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvhpv\" (UniqueName: \"kubernetes.io/projected/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-kube-api-access-jvhpv\") pod \"neutron-db-sync-92zcp\" (UID: \"1e522d4c-e094-4518-bdb5-8ad8e3eccc97\") " pod="openstack/neutron-db-sync-92zcp" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.703028 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-ovsdbserver-nb\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.703059 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-combined-ca-bundle\") pod \"neutron-db-sync-92zcp\" (UID: \"1e522d4c-e094-4518-bdb5-8ad8e3eccc97\") " pod="openstack/neutron-db-sync-92zcp" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.704365 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-ovsdbserver-sb\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.704413 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-dns-swift-storage-0\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.704666 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-scripts\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.704826 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs9ck\" (UniqueName: \"kubernetes.io/projected/337e2c92-f622-4781-8a26-81d358d7baea-kube-api-access-rs9ck\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.704881 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-dns-svc\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.704909 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-config\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.710293 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-etc-machine-id\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.720579 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-combined-ca-bundle\") pod \"neutron-db-sync-92zcp\" (UID: \"1e522d4c-e094-4518-bdb5-8ad8e3eccc97\") " pod="openstack/neutron-db-sync-92zcp" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.721097 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6lm2t"] Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.722721 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-combined-ca-bundle\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.722788 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-config\") pod \"neutron-db-sync-92zcp\" (UID: \"1e522d4c-e094-4518-bdb5-8ad8e3eccc97\") " pod="openstack/neutron-db-sync-92zcp" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.722828 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-scripts\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.723944 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-db-sync-config-data\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.724595 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-config-data\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.739386 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvhpv\" (UniqueName: \"kubernetes.io/projected/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-kube-api-access-jvhpv\") pod \"neutron-db-sync-92zcp\" (UID: \"1e522d4c-e094-4518-bdb5-8ad8e3eccc97\") " pod="openstack/neutron-db-sync-92zcp" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.741497 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m75ld\" (UniqueName: \"kubernetes.io/projected/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-kube-api-access-m75ld\") pod \"cinder-db-sync-k8cjm\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.810557 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl4dm\" (UniqueName: \"kubernetes.io/projected/3cf40633-ef49-416c-afcc-88c33fa76750-kube-api-access-zl4dm\") pod \"barbican-db-sync-h79p5\" (UID: \"3cf40633-ef49-416c-afcc-88c33fa76750\") " pod="openstack/barbican-db-sync-h79p5" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.810628 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-ovsdbserver-sb\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.810651 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-dns-swift-storage-0\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.810674 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-combined-ca-bundle\") pod \"placement-db-sync-6lm2t\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.810708 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-config-data\") pod \"placement-db-sync-6lm2t\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.810731 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3cf40633-ef49-416c-afcc-88c33fa76750-db-sync-config-data\") pod \"barbican-db-sync-h79p5\" (UID: \"3cf40633-ef49-416c-afcc-88c33fa76750\") " pod="openstack/barbican-db-sync-h79p5" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.810761 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs9ck\" (UniqueName: \"kubernetes.io/projected/337e2c92-f622-4781-8a26-81d358d7baea-kube-api-access-rs9ck\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.810796 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-dns-svc\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.810817 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-config\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.811055 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3a4d04e-bcc8-450c-830f-154f9637d513-logs\") pod \"placement-db-sync-6lm2t\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.811137 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpw4z\" (UniqueName: \"kubernetes.io/projected/d3a4d04e-bcc8-450c-830f-154f9637d513-kube-api-access-hpw4z\") pod \"placement-db-sync-6lm2t\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.811261 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf40633-ef49-416c-afcc-88c33fa76750-combined-ca-bundle\") pod \"barbican-db-sync-h79p5\" (UID: \"3cf40633-ef49-416c-afcc-88c33fa76750\") " pod="openstack/barbican-db-sync-h79p5" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.811340 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-ovsdbserver-nb\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.811359 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-scripts\") pod \"placement-db-sync-6lm2t\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.811672 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-config\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.812181 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-ovsdbserver-sb\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.813057 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-dns-svc\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.814173 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-dns-swift-storage-0\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.814191 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-ovsdbserver-nb\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.845284 4997 generic.go:334] "Generic (PLEG): container finished" podID="7f350f7f-80f2-4329-8b61-cac9cdfeb676" containerID="ec17cb996774103234482c6620417bc620c0143ead85111450d1cd12dd354eff" exitCode=0 Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.845340 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" event={"ID":"7f350f7f-80f2-4329-8b61-cac9cdfeb676","Type":"ContainerDied","Data":"ec17cb996774103234482c6620417bc620c0143ead85111450d1cd12dd354eff"} Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.846448 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs9ck\" (UniqueName: \"kubernetes.io/projected/337e2c92-f622-4781-8a26-81d358d7baea-kube-api-access-rs9ck\") pod \"dnsmasq-dns-7884648fd9-svg9r\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.905570 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.906083 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:19:15 crc kubenswrapper[4997]: E1205 07:19:15.906472 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f350f7f-80f2-4329-8b61-cac9cdfeb676" containerName="init" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.906485 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f350f7f-80f2-4329-8b61-cac9cdfeb676" containerName="init" Dec 05 07:19:15 crc kubenswrapper[4997]: E1205 07:19:15.906497 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f350f7f-80f2-4329-8b61-cac9cdfeb676" containerName="dnsmasq-dns" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.906504 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f350f7f-80f2-4329-8b61-cac9cdfeb676" containerName="dnsmasq-dns" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.906708 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f350f7f-80f2-4329-8b61-cac9cdfeb676" containerName="dnsmasq-dns" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.923836 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3a4d04e-bcc8-450c-830f-154f9637d513-logs\") pod \"placement-db-sync-6lm2t\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.923920 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpw4z\" (UniqueName: \"kubernetes.io/projected/d3a4d04e-bcc8-450c-830f-154f9637d513-kube-api-access-hpw4z\") pod \"placement-db-sync-6lm2t\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.924004 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf40633-ef49-416c-afcc-88c33fa76750-combined-ca-bundle\") pod \"barbican-db-sync-h79p5\" (UID: \"3cf40633-ef49-416c-afcc-88c33fa76750\") " pod="openstack/barbican-db-sync-h79p5" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.924047 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.924068 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-scripts\") pod \"placement-db-sync-6lm2t\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.924387 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl4dm\" (UniqueName: \"kubernetes.io/projected/3cf40633-ef49-416c-afcc-88c33fa76750-kube-api-access-zl4dm\") pod \"barbican-db-sync-h79p5\" (UID: \"3cf40633-ef49-416c-afcc-88c33fa76750\") " pod="openstack/barbican-db-sync-h79p5" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.924466 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-combined-ca-bundle\") pod \"placement-db-sync-6lm2t\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.924578 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-config-data\") pod \"placement-db-sync-6lm2t\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.924640 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3cf40633-ef49-416c-afcc-88c33fa76750-db-sync-config-data\") pod \"barbican-db-sync-h79p5\" (UID: \"3cf40633-ef49-416c-afcc-88c33fa76750\") " pod="openstack/barbican-db-sync-h79p5" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.927720 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.929279 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3a4d04e-bcc8-450c-830f-154f9637d513-logs\") pod \"placement-db-sync-6lm2t\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.931860 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-scripts\") pod \"placement-db-sync-6lm2t\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.936436 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.940010 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-config-data\") pod \"placement-db-sync-6lm2t\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.944945 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-combined-ca-bundle\") pod \"placement-db-sync-6lm2t\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.945487 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3cf40633-ef49-416c-afcc-88c33fa76750-db-sync-config-data\") pod \"barbican-db-sync-h79p5\" (UID: \"3cf40633-ef49-416c-afcc-88c33fa76750\") " pod="openstack/barbican-db-sync-h79p5" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.946076 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf40633-ef49-416c-afcc-88c33fa76750-combined-ca-bundle\") pod \"barbican-db-sync-h79p5\" (UID: \"3cf40633-ef49-416c-afcc-88c33fa76750\") " pod="openstack/barbican-db-sync-h79p5" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.952780 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpw4z\" (UniqueName: \"kubernetes.io/projected/d3a4d04e-bcc8-450c-830f-154f9637d513-kube-api-access-hpw4z\") pod \"placement-db-sync-6lm2t\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.953901 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-92zcp" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.963521 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl4dm\" (UniqueName: \"kubernetes.io/projected/3cf40633-ef49-416c-afcc-88c33fa76750-kube-api-access-zl4dm\") pod \"barbican-db-sync-h79p5\" (UID: \"3cf40633-ef49-416c-afcc-88c33fa76750\") " pod="openstack/barbican-db-sync-h79p5" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.972733 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.974359 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:19:15 crc kubenswrapper[4997]: I1205 07:19:15.985672 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.017118 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-h79p5" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.019787 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.026239 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-ovsdbserver-nb\") pod \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.026419 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-ovsdbserver-sb\") pod \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.026534 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbtl2\" (UniqueName: \"kubernetes.io/projected/7f350f7f-80f2-4329-8b61-cac9cdfeb676-kube-api-access-qbtl2\") pod \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.026586 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-dns-svc\") pod \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.026639 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-dns-swift-storage-0\") pod \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.026671 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-config\") pod \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\" (UID: \"7f350f7f-80f2-4329-8b61-cac9cdfeb676\") " Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.026937 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe5443e-726b-4976-a7a3-c226bd3ec481-log-httpd\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.026997 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-config-data\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.027025 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-scripts\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.027053 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.027070 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxr7q\" (UniqueName: \"kubernetes.io/projected/dbe5443e-726b-4976-a7a3-c226bd3ec481-kube-api-access-hxr7q\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.027117 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.027133 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe5443e-726b-4976-a7a3-c226bd3ec481-run-httpd\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.037877 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f350f7f-80f2-4329-8b61-cac9cdfeb676-kube-api-access-qbtl2" (OuterVolumeSpecName: "kube-api-access-qbtl2") pod "7f350f7f-80f2-4329-8b61-cac9cdfeb676" (UID: "7f350f7f-80f2-4329-8b61-cac9cdfeb676"). InnerVolumeSpecName "kube-api-access-qbtl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.112277 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7f350f7f-80f2-4329-8b61-cac9cdfeb676" (UID: "7f350f7f-80f2-4329-8b61-cac9cdfeb676"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.130165 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-scripts\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.130232 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.130256 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxr7q\" (UniqueName: \"kubernetes.io/projected/dbe5443e-726b-4976-a7a3-c226bd3ec481-kube-api-access-hxr7q\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.130300 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe5443e-726b-4976-a7a3-c226bd3ec481-run-httpd\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.130319 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.130378 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe5443e-726b-4976-a7a3-c226bd3ec481-log-httpd\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.130422 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-config-data\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.130476 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbtl2\" (UniqueName: \"kubernetes.io/projected/7f350f7f-80f2-4329-8b61-cac9cdfeb676-kube-api-access-qbtl2\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.130492 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.133544 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe5443e-726b-4976-a7a3-c226bd3ec481-run-httpd\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.133551 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe5443e-726b-4976-a7a3-c226bd3ec481-log-httpd\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.147353 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-config-data\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.148467 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.148784 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.151569 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxr7q\" (UniqueName: \"kubernetes.io/projected/dbe5443e-726b-4976-a7a3-c226bd3ec481-kube-api-access-hxr7q\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.160359 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-scripts\") pod \"ceilometer-0\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.168719 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7f350f7f-80f2-4329-8b61-cac9cdfeb676" (UID: "7f350f7f-80f2-4329-8b61-cac9cdfeb676"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.171389 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7f350f7f-80f2-4329-8b61-cac9cdfeb676" (UID: "7f350f7f-80f2-4329-8b61-cac9cdfeb676"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.181203 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7f350f7f-80f2-4329-8b61-cac9cdfeb676" (UID: "7f350f7f-80f2-4329-8b61-cac9cdfeb676"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.203060 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-config" (OuterVolumeSpecName: "config") pod "7f350f7f-80f2-4329-8b61-cac9cdfeb676" (UID: "7f350f7f-80f2-4329-8b61-cac9cdfeb676"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.232559 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.234111 4997 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.234241 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.234325 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f350f7f-80f2-4329-8b61-cac9cdfeb676-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.281916 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.311195 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-srgk8"] Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.327203 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76cc8cc9bc-cdcn9"] Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.370606 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.372876 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.380283 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qsb8n" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.380568 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.382164 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.382913 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.384582 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.412733 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.426296 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.427417 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.434547 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.434932 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.454175 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.454257 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-config-data\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.454288 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.454308 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4jd9\" (UniqueName: \"kubernetes.io/projected/7098f936-a364-48df-922c-8202d2f4a80e-kube-api-access-z4jd9\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.454355 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7098f936-a364-48df-922c-8202d2f4a80e-logs\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.454381 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-scripts\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.454410 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.454485 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7098f936-a364-48df-922c-8202d2f4a80e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.557752 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.557808 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.557826 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.557848 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.557894 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7098f936-a364-48df-922c-8202d2f4a80e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.557935 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.557957 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-logs\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.558001 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-config-data\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.558073 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.558094 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4jd9\" (UniqueName: \"kubernetes.io/projected/7098f936-a364-48df-922c-8202d2f4a80e-kube-api-access-z4jd9\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.558114 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.558153 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.558176 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7098f936-a364-48df-922c-8202d2f4a80e-logs\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.558209 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-scripts\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.558249 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.558297 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7lc6\" (UniqueName: \"kubernetes.io/projected/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-kube-api-access-q7lc6\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.559767 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.567805 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7098f936-a364-48df-922c-8202d2f4a80e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.567317 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7098f936-a364-48df-922c-8202d2f4a80e-logs\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.571800 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-scripts\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.572065 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.575768 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-config-data\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.598740 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4jd9\" (UniqueName: \"kubernetes.io/projected/7098f936-a364-48df-922c-8202d2f4a80e-kube-api-access-z4jd9\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.608073 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-92zcp"] Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.608877 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.661180 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-logs\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.661255 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.661293 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.661355 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7lc6\" (UniqueName: \"kubernetes.io/projected/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-kube-api-access-q7lc6\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.661381 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.661409 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.661427 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.661444 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.663924 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-logs\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.664146 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.665743 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.670866 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.678356 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.679572 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.683373 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.687113 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7lc6\" (UniqueName: \"kubernetes.io/projected/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-kube-api-access-q7lc6\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.686393 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.713418 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.805231 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-k8cjm"] Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.825680 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-svg9r"] Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.873996 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-92zcp" event={"ID":"1e522d4c-e094-4518-bdb5-8ad8e3eccc97","Type":"ContainerStarted","Data":"fba9c55e23f58ada6acdc1dd2897927d67ae151a7e37e18f0949c102217318ed"} Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.882933 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" event={"ID":"7f350f7f-80f2-4329-8b61-cac9cdfeb676","Type":"ContainerDied","Data":"83c2b0a20b78bfe6f81db74ec9f1d79c1fc895512468380ead3b2b944f99cee0"} Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.883010 4997 scope.go:117] "RemoveContainer" containerID="ec17cb996774103234482c6620417bc620c0143ead85111450d1cd12dd354eff" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.883196 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84db446f-m9lt2" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.890891 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" event={"ID":"fe20e521-eeb9-4f15-84d1-925c0f34126d","Type":"ContainerStarted","Data":"bc65a4cf34802a127013828e11816b46434c21965bea30e73f81819e58f810fc"} Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.916698 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7884648fd9-svg9r" event={"ID":"337e2c92-f622-4781-8a26-81d358d7baea","Type":"ContainerStarted","Data":"f611c5363a1ca2b236c2eace6c685f3ddd3d473a869c4efd55ea04a162edefdb"} Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.922461 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-k8cjm" event={"ID":"78c3f4c1-158a-4d14-96c2-f9b6a663fd45","Type":"ContainerStarted","Data":"dcbe5a58304eddda0bd0d21df8f7798731d65a144530c50d8acddfc444a5795f"} Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.929282 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-srgk8" event={"ID":"42a735ac-395f-4357-9db3-1aff67aef6fa","Type":"ContainerStarted","Data":"e959b0558c0b44a2e65e83af63a216e610db0b7448ec2a4f7fe2ff61d78e4421"} Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.939320 4997 scope.go:117] "RemoveContainer" containerID="6097c98b4bc5bbf756341e4b34764fca582ce25906d164d635e716e0334eabc6" Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.943540 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84db446f-m9lt2"] Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.970575 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d84db446f-m9lt2"] Dec 05 07:19:16 crc kubenswrapper[4997]: I1205 07:19:16.985663 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 07:19:17 crc kubenswrapper[4997]: I1205 07:19:17.000086 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-h79p5"] Dec 05 07:19:17 crc kubenswrapper[4997]: I1205 07:19:17.009104 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 07:19:17 crc kubenswrapper[4997]: I1205 07:19:17.027234 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6lm2t"] Dec 05 07:19:17 crc kubenswrapper[4997]: I1205 07:19:17.108713 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:19:17 crc kubenswrapper[4997]: W1205 07:19:17.118014 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbe5443e_726b_4976_a7a3_c226bd3ec481.slice/crio-b07e02d6dbb72e64ea73e58b4669858e343f72aae7b1d6391e5a98ab1c793e71 WatchSource:0}: Error finding container b07e02d6dbb72e64ea73e58b4669858e343f72aae7b1d6391e5a98ab1c793e71: Status 404 returned error can't find the container with id b07e02d6dbb72e64ea73e58b4669858e343f72aae7b1d6391e5a98ab1c793e71 Dec 05 07:19:17 crc kubenswrapper[4997]: I1205 07:19:17.616167 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:19:17 crc kubenswrapper[4997]: I1205 07:19:17.832988 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f350f7f-80f2-4329-8b61-cac9cdfeb676" path="/var/lib/kubelet/pods/7f350f7f-80f2-4329-8b61-cac9cdfeb676/volumes" Dec 05 07:19:17 crc kubenswrapper[4997]: I1205 07:19:17.835503 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:19:17 crc kubenswrapper[4997]: I1205 07:19:17.941889 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.019001 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6lm2t" event={"ID":"d3a4d04e-bcc8-450c-830f-154f9637d513","Type":"ContainerStarted","Data":"8fd8513dd403b29aca71ca24141cdaeba873404f796acf8c160b410cb6ce66f5"} Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.042830 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8","Type":"ContainerStarted","Data":"11cc1953573411538efd483295538acfbd5e4a06f7c00e5aa18441289c34d7f4"} Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.052669 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-92zcp" event={"ID":"1e522d4c-e094-4518-bdb5-8ad8e3eccc97","Type":"ContainerStarted","Data":"2d0cd2f22990c11fe1120f60bcd9583801b4644625f0d7fb00ad15a10ace632b"} Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.093294 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-srgk8" event={"ID":"42a735ac-395f-4357-9db3-1aff67aef6fa","Type":"ContainerStarted","Data":"f82f717242295d98f71aec9f7fc80f2c5f943a9da4e85ec393c1fd2c9ccd37ab"} Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.110566 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-h79p5" event={"ID":"3cf40633-ef49-416c-afcc-88c33fa76750","Type":"ContainerStarted","Data":"9436cb2bb18829ab26ddd4408fbf547030e5faa667d4d1940682f0e87ed35eeb"} Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.115061 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-92zcp" podStartSLOduration=3.115051171 podStartE2EDuration="3.115051171s" podCreationTimestamp="2025-12-05 07:19:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:19:18.111233058 +0000 UTC m=+1458.640140339" watchObservedRunningTime="2025-12-05 07:19:18.115051171 +0000 UTC m=+1458.643958432" Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.164800 4997 generic.go:334] "Generic (PLEG): container finished" podID="fe20e521-eeb9-4f15-84d1-925c0f34126d" containerID="7637207a94d3955c0085d27d935b57a74343e3153921161c8424970bd10a0967" exitCode=0 Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.164904 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" event={"ID":"fe20e521-eeb9-4f15-84d1-925c0f34126d","Type":"ContainerDied","Data":"7637207a94d3955c0085d27d935b57a74343e3153921161c8424970bd10a0967"} Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.193808 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-srgk8" podStartSLOduration=3.193767183 podStartE2EDuration="3.193767183s" podCreationTimestamp="2025-12-05 07:19:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:19:18.156879179 +0000 UTC m=+1458.685786440" watchObservedRunningTime="2025-12-05 07:19:18.193767183 +0000 UTC m=+1458.722674464" Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.243152 4997 generic.go:334] "Generic (PLEG): container finished" podID="337e2c92-f622-4781-8a26-81d358d7baea" containerID="8471674046ec6cb711d5c75b3b5646e1ecf75702d27e3f7b07f8200edf153d13" exitCode=0 Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.244164 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7884648fd9-svg9r" event={"ID":"337e2c92-f622-4781-8a26-81d358d7baea","Type":"ContainerDied","Data":"8471674046ec6cb711d5c75b3b5646e1ecf75702d27e3f7b07f8200edf153d13"} Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.257804 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.258392 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7098f936-a364-48df-922c-8202d2f4a80e","Type":"ContainerStarted","Data":"16ac764d8694b6ec708506770c93eea4f1e5f02b1aeb024a35646d412c268106"} Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.273501 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe5443e-726b-4976-a7a3-c226bd3ec481","Type":"ContainerStarted","Data":"b07e02d6dbb72e64ea73e58b4669858e343f72aae7b1d6391e5a98ab1c793e71"} Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.285274 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.863559 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.962654 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-dns-swift-storage-0\") pod \"fe20e521-eeb9-4f15-84d1-925c0f34126d\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.962739 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-ovsdbserver-nb\") pod \"fe20e521-eeb9-4f15-84d1-925c0f34126d\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.962849 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-dns-svc\") pod \"fe20e521-eeb9-4f15-84d1-925c0f34126d\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.962902 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-config\") pod \"fe20e521-eeb9-4f15-84d1-925c0f34126d\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.962957 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvlms\" (UniqueName: \"kubernetes.io/projected/fe20e521-eeb9-4f15-84d1-925c0f34126d-kube-api-access-lvlms\") pod \"fe20e521-eeb9-4f15-84d1-925c0f34126d\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " Dec 05 07:19:18 crc kubenswrapper[4997]: I1205 07:19:18.963123 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-ovsdbserver-sb\") pod \"fe20e521-eeb9-4f15-84d1-925c0f34126d\" (UID: \"fe20e521-eeb9-4f15-84d1-925c0f34126d\") " Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:18.999368 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe20e521-eeb9-4f15-84d1-925c0f34126d-kube-api-access-lvlms" (OuterVolumeSpecName: "kube-api-access-lvlms") pod "fe20e521-eeb9-4f15-84d1-925c0f34126d" (UID: "fe20e521-eeb9-4f15-84d1-925c0f34126d"). InnerVolumeSpecName "kube-api-access-lvlms". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.003488 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fe20e521-eeb9-4f15-84d1-925c0f34126d" (UID: "fe20e521-eeb9-4f15-84d1-925c0f34126d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.008858 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fe20e521-eeb9-4f15-84d1-925c0f34126d" (UID: "fe20e521-eeb9-4f15-84d1-925c0f34126d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.010865 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-config" (OuterVolumeSpecName: "config") pod "fe20e521-eeb9-4f15-84d1-925c0f34126d" (UID: "fe20e521-eeb9-4f15-84d1-925c0f34126d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.025182 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fe20e521-eeb9-4f15-84d1-925c0f34126d" (UID: "fe20e521-eeb9-4f15-84d1-925c0f34126d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.026211 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fe20e521-eeb9-4f15-84d1-925c0f34126d" (UID: "fe20e521-eeb9-4f15-84d1-925c0f34126d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.065197 4997 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.065231 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.065241 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.065252 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.065267 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvlms\" (UniqueName: \"kubernetes.io/projected/fe20e521-eeb9-4f15-84d1-925c0f34126d-kube-api-access-lvlms\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.065280 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe20e521-eeb9-4f15-84d1-925c0f34126d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.289128 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7098f936-a364-48df-922c-8202d2f4a80e","Type":"ContainerStarted","Data":"8ebc72dc3d50298609d59b368904d6395e889e7151f99e044fa28283008de46e"} Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.296142 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" event={"ID":"fe20e521-eeb9-4f15-84d1-925c0f34126d","Type":"ContainerDied","Data":"bc65a4cf34802a127013828e11816b46434c21965bea30e73f81819e58f810fc"} Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.296706 4997 scope.go:117] "RemoveContainer" containerID="7637207a94d3955c0085d27d935b57a74343e3153921161c8424970bd10a0967" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.296202 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76cc8cc9bc-cdcn9" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.311195 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7884648fd9-svg9r" event={"ID":"337e2c92-f622-4781-8a26-81d358d7baea","Type":"ContainerStarted","Data":"28d20545adfcfd72e1756fb0dc1443224afda6a7f0dc28db51dd56d5b989674c"} Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.311861 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.338215 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7884648fd9-svg9r" podStartSLOduration=4.338192973 podStartE2EDuration="4.338192973s" podCreationTimestamp="2025-12-05 07:19:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:19:19.336878088 +0000 UTC m=+1459.865785359" watchObservedRunningTime="2025-12-05 07:19:19.338192973 +0000 UTC m=+1459.867100224" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.633100 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76cc8cc9bc-cdcn9"] Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.662926 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76cc8cc9bc-cdcn9"] Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.770085 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe20e521-eeb9-4f15-84d1-925c0f34126d" path="/var/lib/kubelet/pods/fe20e521-eeb9-4f15-84d1-925c0f34126d/volumes" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.771321 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.771373 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.771419 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.772213 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1b8d25609ffa42f5e50251e3eb59f7240e05fe03de714e77c29a98e4dc7b4510"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 07:19:19 crc kubenswrapper[4997]: I1205 07:19:19.772281 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://1b8d25609ffa42f5e50251e3eb59f7240e05fe03de714e77c29a98e4dc7b4510" gracePeriod=600 Dec 05 07:19:20 crc kubenswrapper[4997]: I1205 07:19:20.392789 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="1b8d25609ffa42f5e50251e3eb59f7240e05fe03de714e77c29a98e4dc7b4510" exitCode=0 Dec 05 07:19:20 crc kubenswrapper[4997]: I1205 07:19:20.394052 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"1b8d25609ffa42f5e50251e3eb59f7240e05fe03de714e77c29a98e4dc7b4510"} Dec 05 07:19:20 crc kubenswrapper[4997]: I1205 07:19:20.394811 4997 scope.go:117] "RemoveContainer" containerID="6cb7be4205d2b6d3befa7784153570d8d5c45d7400acdc190e0c1608d52679b2" Dec 05 07:19:20 crc kubenswrapper[4997]: I1205 07:19:20.414261 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8","Type":"ContainerStarted","Data":"2c33de8cc56e9f49bb17a4c545e112e59d5b817d9e70ca54e9b46c79c79014ba"} Dec 05 07:19:21 crc kubenswrapper[4997]: I1205 07:19:21.430429 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7098f936-a364-48df-922c-8202d2f4a80e","Type":"ContainerStarted","Data":"697cbd204f70ac1631f244a00c193d1f66b8976e7fb561cf00fe1afe93eb3bb0"} Dec 05 07:19:21 crc kubenswrapper[4997]: I1205 07:19:21.430656 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7098f936-a364-48df-922c-8202d2f4a80e" containerName="glance-log" containerID="cri-o://8ebc72dc3d50298609d59b368904d6395e889e7151f99e044fa28283008de46e" gracePeriod=30 Dec 05 07:19:21 crc kubenswrapper[4997]: I1205 07:19:21.430689 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7098f936-a364-48df-922c-8202d2f4a80e" containerName="glance-httpd" containerID="cri-o://697cbd204f70ac1631f244a00c193d1f66b8976e7fb561cf00fe1afe93eb3bb0" gracePeriod=30 Dec 05 07:19:21 crc kubenswrapper[4997]: I1205 07:19:21.436205 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08"} Dec 05 07:19:21 crc kubenswrapper[4997]: I1205 07:19:21.441498 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8","Type":"ContainerStarted","Data":"512330c0d21f3df416e8eea2323fc059b60383128779e8e043be108be9438d25"} Dec 05 07:19:21 crc kubenswrapper[4997]: I1205 07:19:21.441738 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" containerName="glance-log" containerID="cri-o://2c33de8cc56e9f49bb17a4c545e112e59d5b817d9e70ca54e9b46c79c79014ba" gracePeriod=30 Dec 05 07:19:21 crc kubenswrapper[4997]: I1205 07:19:21.442112 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" containerName="glance-httpd" containerID="cri-o://512330c0d21f3df416e8eea2323fc059b60383128779e8e043be108be9438d25" gracePeriod=30 Dec 05 07:19:21 crc kubenswrapper[4997]: I1205 07:19:21.469737 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.469711729 podStartE2EDuration="6.469711729s" podCreationTimestamp="2025-12-05 07:19:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:19:21.456915954 +0000 UTC m=+1461.985823225" watchObservedRunningTime="2025-12-05 07:19:21.469711729 +0000 UTC m=+1461.998619020" Dec 05 07:19:21 crc kubenswrapper[4997]: I1205 07:19:21.532200 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.532170714 podStartE2EDuration="6.532170714s" podCreationTimestamp="2025-12-05 07:19:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:19:21.500677134 +0000 UTC m=+1462.029584415" watchObservedRunningTime="2025-12-05 07:19:21.532170714 +0000 UTC m=+1462.061077975" Dec 05 07:19:22 crc kubenswrapper[4997]: I1205 07:19:22.461160 4997 generic.go:334] "Generic (PLEG): container finished" podID="7098f936-a364-48df-922c-8202d2f4a80e" containerID="697cbd204f70ac1631f244a00c193d1f66b8976e7fb561cf00fe1afe93eb3bb0" exitCode=0 Dec 05 07:19:22 crc kubenswrapper[4997]: I1205 07:19:22.461770 4997 generic.go:334] "Generic (PLEG): container finished" podID="7098f936-a364-48df-922c-8202d2f4a80e" containerID="8ebc72dc3d50298609d59b368904d6395e889e7151f99e044fa28283008de46e" exitCode=143 Dec 05 07:19:22 crc kubenswrapper[4997]: I1205 07:19:22.461876 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7098f936-a364-48df-922c-8202d2f4a80e","Type":"ContainerDied","Data":"697cbd204f70ac1631f244a00c193d1f66b8976e7fb561cf00fe1afe93eb3bb0"} Dec 05 07:19:22 crc kubenswrapper[4997]: I1205 07:19:22.461912 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7098f936-a364-48df-922c-8202d2f4a80e","Type":"ContainerDied","Data":"8ebc72dc3d50298609d59b368904d6395e889e7151f99e044fa28283008de46e"} Dec 05 07:19:22 crc kubenswrapper[4997]: I1205 07:19:22.467979 4997 generic.go:334] "Generic (PLEG): container finished" podID="0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" containerID="512330c0d21f3df416e8eea2323fc059b60383128779e8e043be108be9438d25" exitCode=0 Dec 05 07:19:22 crc kubenswrapper[4997]: I1205 07:19:22.468008 4997 generic.go:334] "Generic (PLEG): container finished" podID="0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" containerID="2c33de8cc56e9f49bb17a4c545e112e59d5b817d9e70ca54e9b46c79c79014ba" exitCode=143 Dec 05 07:19:22 crc kubenswrapper[4997]: I1205 07:19:22.469158 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8","Type":"ContainerDied","Data":"512330c0d21f3df416e8eea2323fc059b60383128779e8e043be108be9438d25"} Dec 05 07:19:22 crc kubenswrapper[4997]: I1205 07:19:22.469182 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8","Type":"ContainerDied","Data":"2c33de8cc56e9f49bb17a4c545e112e59d5b817d9e70ca54e9b46c79c79014ba"} Dec 05 07:19:23 crc kubenswrapper[4997]: I1205 07:19:23.482052 4997 generic.go:334] "Generic (PLEG): container finished" podID="42a735ac-395f-4357-9db3-1aff67aef6fa" containerID="f82f717242295d98f71aec9f7fc80f2c5f943a9da4e85ec393c1fd2c9ccd37ab" exitCode=0 Dec 05 07:19:23 crc kubenswrapper[4997]: I1205 07:19:23.482255 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-srgk8" event={"ID":"42a735ac-395f-4357-9db3-1aff67aef6fa","Type":"ContainerDied","Data":"f82f717242295d98f71aec9f7fc80f2c5f943a9da4e85ec393c1fd2c9ccd37ab"} Dec 05 07:19:25 crc kubenswrapper[4997]: I1205 07:19:25.988056 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:26 crc kubenswrapper[4997]: I1205 07:19:26.060362 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-vz755"] Dec 05 07:19:26 crc kubenswrapper[4997]: I1205 07:19:26.061204 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-779c5847bc-vz755" podUID="bf7639fd-80e4-4f37-8f92-9b9f73a01557" containerName="dnsmasq-dns" containerID="cri-o://d59ccec3588a99dd3d41c828fb0ed5ed9b64b9c1c576b222b7fac17148e89628" gracePeriod=10 Dec 05 07:19:26 crc kubenswrapper[4997]: I1205 07:19:26.525362 4997 generic.go:334] "Generic (PLEG): container finished" podID="bf7639fd-80e4-4f37-8f92-9b9f73a01557" containerID="d59ccec3588a99dd3d41c828fb0ed5ed9b64b9c1c576b222b7fac17148e89628" exitCode=0 Dec 05 07:19:26 crc kubenswrapper[4997]: I1205 07:19:26.525417 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-779c5847bc-vz755" event={"ID":"bf7639fd-80e4-4f37-8f92-9b9f73a01557","Type":"ContainerDied","Data":"d59ccec3588a99dd3d41c828fb0ed5ed9b64b9c1c576b222b7fac17148e89628"} Dec 05 07:19:27 crc kubenswrapper[4997]: I1205 07:19:27.120899 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-779c5847bc-vz755" podUID="bf7639fd-80e4-4f37-8f92-9b9f73a01557" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: connect: connection refused" Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.303309 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.444120 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-scripts\") pod \"42a735ac-395f-4357-9db3-1aff67aef6fa\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.444174 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-config-data\") pod \"42a735ac-395f-4357-9db3-1aff67aef6fa\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.444202 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-fernet-keys\") pod \"42a735ac-395f-4357-9db3-1aff67aef6fa\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.444261 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-credential-keys\") pod \"42a735ac-395f-4357-9db3-1aff67aef6fa\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.444471 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-combined-ca-bundle\") pod \"42a735ac-395f-4357-9db3-1aff67aef6fa\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.444627 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qslqt\" (UniqueName: \"kubernetes.io/projected/42a735ac-395f-4357-9db3-1aff67aef6fa-kube-api-access-qslqt\") pod \"42a735ac-395f-4357-9db3-1aff67aef6fa\" (UID: \"42a735ac-395f-4357-9db3-1aff67aef6fa\") " Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.455193 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-scripts" (OuterVolumeSpecName: "scripts") pod "42a735ac-395f-4357-9db3-1aff67aef6fa" (UID: "42a735ac-395f-4357-9db3-1aff67aef6fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.456372 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42a735ac-395f-4357-9db3-1aff67aef6fa-kube-api-access-qslqt" (OuterVolumeSpecName: "kube-api-access-qslqt") pod "42a735ac-395f-4357-9db3-1aff67aef6fa" (UID: "42a735ac-395f-4357-9db3-1aff67aef6fa"). InnerVolumeSpecName "kube-api-access-qslqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.459090 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "42a735ac-395f-4357-9db3-1aff67aef6fa" (UID: "42a735ac-395f-4357-9db3-1aff67aef6fa"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.462055 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "42a735ac-395f-4357-9db3-1aff67aef6fa" (UID: "42a735ac-395f-4357-9db3-1aff67aef6fa"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.520847 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42a735ac-395f-4357-9db3-1aff67aef6fa" (UID: "42a735ac-395f-4357-9db3-1aff67aef6fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.524749 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-config-data" (OuterVolumeSpecName: "config-data") pod "42a735ac-395f-4357-9db3-1aff67aef6fa" (UID: "42a735ac-395f-4357-9db3-1aff67aef6fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.547090 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qslqt\" (UniqueName: \"kubernetes.io/projected/42a735ac-395f-4357-9db3-1aff67aef6fa-kube-api-access-qslqt\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.547143 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.547156 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.547170 4997 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.547186 4997 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.547199 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42a735ac-395f-4357-9db3-1aff67aef6fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.570885 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-srgk8" event={"ID":"42a735ac-395f-4357-9db3-1aff67aef6fa","Type":"ContainerDied","Data":"e959b0558c0b44a2e65e83af63a216e610db0b7448ec2a4f7fe2ff61d78e4421"} Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.570942 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e959b0558c0b44a2e65e83af63a216e610db0b7448ec2a4f7fe2ff61d78e4421" Dec 05 07:19:30 crc kubenswrapper[4997]: I1205 07:19:30.571012 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-srgk8" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.411248 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-srgk8"] Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.422005 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-srgk8"] Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.504717 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-7dc7z"] Dec 05 07:19:31 crc kubenswrapper[4997]: E1205 07:19:31.505377 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42a735ac-395f-4357-9db3-1aff67aef6fa" containerName="keystone-bootstrap" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.505402 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="42a735ac-395f-4357-9db3-1aff67aef6fa" containerName="keystone-bootstrap" Dec 05 07:19:31 crc kubenswrapper[4997]: E1205 07:19:31.505447 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe20e521-eeb9-4f15-84d1-925c0f34126d" containerName="init" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.505457 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe20e521-eeb9-4f15-84d1-925c0f34126d" containerName="init" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.505760 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe20e521-eeb9-4f15-84d1-925c0f34126d" containerName="init" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.505815 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="42a735ac-395f-4357-9db3-1aff67aef6fa" containerName="keystone-bootstrap" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.506969 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.510395 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.510638 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.510775 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.511040 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-856gn" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.511201 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.516944 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7dc7z"] Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.674079 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-fernet-keys\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.674172 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwnxq\" (UniqueName: \"kubernetes.io/projected/145b182b-23c5-444d-864f-0cbd2c46902d-kube-api-access-xwnxq\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.674261 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-scripts\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.674805 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-combined-ca-bundle\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.674936 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-config-data\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.675024 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-credential-keys\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.762311 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42a735ac-395f-4357-9db3-1aff67aef6fa" path="/var/lib/kubelet/pods/42a735ac-395f-4357-9db3-1aff67aef6fa/volumes" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.776220 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-scripts\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.776302 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-combined-ca-bundle\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.776348 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-config-data\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.776371 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-credential-keys\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.776399 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-fernet-keys\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.776430 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwnxq\" (UniqueName: \"kubernetes.io/projected/145b182b-23c5-444d-864f-0cbd2c46902d-kube-api-access-xwnxq\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.783420 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-combined-ca-bundle\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.784417 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-credential-keys\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.789051 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-scripts\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.793263 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-fernet-keys\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.795469 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwnxq\" (UniqueName: \"kubernetes.io/projected/145b182b-23c5-444d-864f-0cbd2c46902d-kube-api-access-xwnxq\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.802019 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-config-data\") pod \"keystone-bootstrap-7dc7z\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:31 crc kubenswrapper[4997]: I1205 07:19:31.840850 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:37 crc kubenswrapper[4997]: I1205 07:19:37.120475 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-779c5847bc-vz755" podUID="bf7639fd-80e4-4f37-8f92-9b9f73a01557" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: i/o timeout" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.655554 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-779c5847bc-vz755" event={"ID":"bf7639fd-80e4-4f37-8f92-9b9f73a01557","Type":"ContainerDied","Data":"f618dc8b8a8b99a60ec86a1d35977485f7a87f6a19bf19a722ef73bdd34d79be"} Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.656079 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f618dc8b8a8b99a60ec86a1d35977485f7a87f6a19bf19a722ef73bdd34d79be" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.660167 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8","Type":"ContainerDied","Data":"11cc1953573411538efd483295538acfbd5e4a06f7c00e5aa18441289c34d7f4"} Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.660206 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11cc1953573411538efd483295538acfbd5e4a06f7c00e5aa18441289c34d7f4" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.663415 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7098f936-a364-48df-922c-8202d2f4a80e","Type":"ContainerDied","Data":"16ac764d8694b6ec708506770c93eea4f1e5f02b1aeb024a35646d412c268106"} Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.663529 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16ac764d8694b6ec708506770c93eea4f1e5f02b1aeb024a35646d412c268106" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.735172 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.749706 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.756137 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.816770 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-dns-swift-storage-0\") pod \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817232 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-config-data\") pod \"7098f936-a364-48df-922c-8202d2f4a80e\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817262 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-config-data\") pod \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817293 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7098f936-a364-48df-922c-8202d2f4a80e-logs\") pod \"7098f936-a364-48df-922c-8202d2f4a80e\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817322 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"7098f936-a364-48df-922c-8202d2f4a80e\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817349 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-public-tls-certs\") pod \"7098f936-a364-48df-922c-8202d2f4a80e\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817369 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4jd9\" (UniqueName: \"kubernetes.io/projected/7098f936-a364-48df-922c-8202d2f4a80e-kube-api-access-z4jd9\") pod \"7098f936-a364-48df-922c-8202d2f4a80e\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817386 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-internal-tls-certs\") pod \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817428 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-combined-ca-bundle\") pod \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817472 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7098f936-a364-48df-922c-8202d2f4a80e-httpd-run\") pod \"7098f936-a364-48df-922c-8202d2f4a80e\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817499 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-scripts\") pod \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817523 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817551 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9zdv\" (UniqueName: \"kubernetes.io/projected/bf7639fd-80e4-4f37-8f92-9b9f73a01557-kube-api-access-m9zdv\") pod \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817576 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-dns-svc\") pod \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817598 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-config\") pod \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817640 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-ovsdbserver-nb\") pod \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817665 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-ovsdbserver-sb\") pod \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\" (UID: \"bf7639fd-80e4-4f37-8f92-9b9f73a01557\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817683 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7lc6\" (UniqueName: \"kubernetes.io/projected/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-kube-api-access-q7lc6\") pod \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817701 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-logs\") pod \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817723 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-combined-ca-bundle\") pod \"7098f936-a364-48df-922c-8202d2f4a80e\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817741 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-httpd-run\") pod \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\" (UID: \"0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.817779 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-scripts\") pod \"7098f936-a364-48df-922c-8202d2f4a80e\" (UID: \"7098f936-a364-48df-922c-8202d2f4a80e\") " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.819984 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-logs" (OuterVolumeSpecName: "logs") pod "0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" (UID: "0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.822343 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" (UID: "0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.841969 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-scripts" (OuterVolumeSpecName: "scripts") pod "0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" (UID: "0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.842283 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7098f936-a364-48df-922c-8202d2f4a80e-logs" (OuterVolumeSpecName: "logs") pod "7098f936-a364-48df-922c-8202d2f4a80e" (UID: "7098f936-a364-48df-922c-8202d2f4a80e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.842301 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7098f936-a364-48df-922c-8202d2f4a80e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7098f936-a364-48df-922c-8202d2f4a80e" (UID: "7098f936-a364-48df-922c-8202d2f4a80e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.876149 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7098f936-a364-48df-922c-8202d2f4a80e-kube-api-access-z4jd9" (OuterVolumeSpecName: "kube-api-access-z4jd9") pod "7098f936-a364-48df-922c-8202d2f4a80e" (UID: "7098f936-a364-48df-922c-8202d2f4a80e"). InnerVolumeSpecName "kube-api-access-z4jd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.876318 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" (UID: "0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.876362 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "7098f936-a364-48df-922c-8202d2f4a80e" (UID: "7098f936-a364-48df-922c-8202d2f4a80e"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.877388 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf7639fd-80e4-4f37-8f92-9b9f73a01557-kube-api-access-m9zdv" (OuterVolumeSpecName: "kube-api-access-m9zdv") pod "bf7639fd-80e4-4f37-8f92-9b9f73a01557" (UID: "bf7639fd-80e4-4f37-8f92-9b9f73a01557"). InnerVolumeSpecName "kube-api-access-m9zdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.879366 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-scripts" (OuterVolumeSpecName: "scripts") pod "7098f936-a364-48df-922c-8202d2f4a80e" (UID: "7098f936-a364-48df-922c-8202d2f4a80e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.880515 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-kube-api-access-q7lc6" (OuterVolumeSpecName: "kube-api-access-q7lc6") pod "0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" (UID: "0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8"). InnerVolumeSpecName "kube-api-access-q7lc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.903504 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7098f936-a364-48df-922c-8202d2f4a80e" (UID: "7098f936-a364-48df-922c-8202d2f4a80e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.905460 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" (UID: "0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.929321 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.930287 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4jd9\" (UniqueName: \"kubernetes.io/projected/7098f936-a364-48df-922c-8202d2f4a80e-kube-api-access-z4jd9\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.930358 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.930373 4997 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7098f936-a364-48df-922c-8202d2f4a80e-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.930387 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.930424 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.930438 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9zdv\" (UniqueName: \"kubernetes.io/projected/bf7639fd-80e4-4f37-8f92-9b9f73a01557-kube-api-access-m9zdv\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.930451 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7lc6\" (UniqueName: \"kubernetes.io/projected/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-kube-api-access-q7lc6\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.930465 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.930476 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.930488 4997 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.930497 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.930507 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7098f936-a364-48df-922c-8202d2f4a80e-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.948675 4997 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.949068 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bf7639fd-80e4-4f37-8f92-9b9f73a01557" (UID: "bf7639fd-80e4-4f37-8f92-9b9f73a01557"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.951732 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-config-data" (OuterVolumeSpecName: "config-data") pod "7098f936-a364-48df-922c-8202d2f4a80e" (UID: "7098f936-a364-48df-922c-8202d2f4a80e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.956183 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-config" (OuterVolumeSpecName: "config") pod "bf7639fd-80e4-4f37-8f92-9b9f73a01557" (UID: "bf7639fd-80e4-4f37-8f92-9b9f73a01557"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.958278 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" (UID: "0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.959598 4997 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.966540 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7098f936-a364-48df-922c-8202d2f4a80e" (UID: "7098f936-a364-48df-922c-8202d2f4a80e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.977212 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bf7639fd-80e4-4f37-8f92-9b9f73a01557" (UID: "bf7639fd-80e4-4f37-8f92-9b9f73a01557"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.985894 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-config-data" (OuterVolumeSpecName: "config-data") pod "0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" (UID: "0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.990882 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bf7639fd-80e4-4f37-8f92-9b9f73a01557" (UID: "bf7639fd-80e4-4f37-8f92-9b9f73a01557"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:38 crc kubenswrapper[4997]: I1205 07:19:38.994997 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bf7639fd-80e4-4f37-8f92-9b9f73a01557" (UID: "bf7639fd-80e4-4f37-8f92-9b9f73a01557"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.032495 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.032555 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.032572 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.032626 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.032638 4997 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf7639fd-80e4-4f37-8f92-9b9f73a01557-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.032651 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.032661 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.032671 4997 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.032681 4997 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7098f936-a364-48df-922c-8202d2f4a80e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.032691 4997 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.032701 4997 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.680878 4997 generic.go:334] "Generic (PLEG): container finished" podID="1e522d4c-e094-4518-bdb5-8ad8e3eccc97" containerID="2d0cd2f22990c11fe1120f60bcd9583801b4644625f0d7fb00ad15a10ace632b" exitCode=0 Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.681493 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.682917 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-779c5847bc-vz755" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.681084 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-92zcp" event={"ID":"1e522d4c-e094-4518-bdb5-8ad8e3eccc97","Type":"ContainerDied","Data":"2d0cd2f22990c11fe1120f60bcd9583801b4644625f0d7fb00ad15a10ace632b"} Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.683211 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.802454 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-vz755"] Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.802499 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-vz755"] Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.815959 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.838702 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.879043 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.905122 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.909039 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:19:39 crc kubenswrapper[4997]: E1205 07:19:39.909596 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7098f936-a364-48df-922c-8202d2f4a80e" containerName="glance-httpd" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.909628 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7098f936-a364-48df-922c-8202d2f4a80e" containerName="glance-httpd" Dec 05 07:19:39 crc kubenswrapper[4997]: E1205 07:19:39.909658 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" containerName="glance-log" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.909666 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" containerName="glance-log" Dec 05 07:19:39 crc kubenswrapper[4997]: E1205 07:19:39.909682 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" containerName="glance-httpd" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.909690 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" containerName="glance-httpd" Dec 05 07:19:39 crc kubenswrapper[4997]: E1205 07:19:39.909709 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf7639fd-80e4-4f37-8f92-9b9f73a01557" containerName="dnsmasq-dns" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.909716 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf7639fd-80e4-4f37-8f92-9b9f73a01557" containerName="dnsmasq-dns" Dec 05 07:19:39 crc kubenswrapper[4997]: E1205 07:19:39.909735 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf7639fd-80e4-4f37-8f92-9b9f73a01557" containerName="init" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.909742 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf7639fd-80e4-4f37-8f92-9b9f73a01557" containerName="init" Dec 05 07:19:39 crc kubenswrapper[4997]: E1205 07:19:39.909755 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7098f936-a364-48df-922c-8202d2f4a80e" containerName="glance-log" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.909763 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7098f936-a364-48df-922c-8202d2f4a80e" containerName="glance-log" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.909969 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf7639fd-80e4-4f37-8f92-9b9f73a01557" containerName="dnsmasq-dns" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.909994 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" containerName="glance-log" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.910004 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" containerName="glance-httpd" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.910027 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="7098f936-a364-48df-922c-8202d2f4a80e" containerName="glance-httpd" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.910041 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="7098f936-a364-48df-922c-8202d2f4a80e" containerName="glance-log" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.911431 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.913515 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qsb8n" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.913862 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.913862 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.914061 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.926603 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.929226 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.931706 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.932434 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.956935 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:19:39 crc kubenswrapper[4997]: I1205 07:19:39.969501 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.051994 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.052085 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.052178 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z54hg\" (UniqueName: \"kubernetes.io/projected/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-kube-api-access-z54hg\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.052219 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-config-data\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.052248 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/494f1331-dcea-4745-add0-50794d9e38c9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.052281 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttgzh\" (UniqueName: \"kubernetes.io/projected/494f1331-dcea-4745-add0-50794d9e38c9-kube-api-access-ttgzh\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.052303 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.052363 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.052401 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.052441 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.052861 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.053033 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-scripts\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.053102 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.053251 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.053733 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-logs\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.053782 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/494f1331-dcea-4745-add0-50794d9e38c9-logs\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.157162 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.157240 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.157276 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z54hg\" (UniqueName: \"kubernetes.io/projected/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-kube-api-access-z54hg\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.157310 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-config-data\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.157341 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/494f1331-dcea-4745-add0-50794d9e38c9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.157371 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttgzh\" (UniqueName: \"kubernetes.io/projected/494f1331-dcea-4745-add0-50794d9e38c9-kube-api-access-ttgzh\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.157393 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.157426 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.157459 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.157492 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.157563 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.157663 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-scripts\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.157699 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.157724 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.157749 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-logs\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.157774 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/494f1331-dcea-4745-add0-50794d9e38c9-logs\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.157866 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.158737 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/494f1331-dcea-4745-add0-50794d9e38c9-logs\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.159635 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.160099 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/494f1331-dcea-4745-add0-50794d9e38c9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.160272 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-logs\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.159635 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.178716 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-scripts\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.178761 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.179084 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.183808 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.187750 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.187766 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.188025 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-config-data\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.195580 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z54hg\" (UniqueName: \"kubernetes.io/projected/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-kube-api-access-z54hg\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.195924 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.197120 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttgzh\" (UniqueName: \"kubernetes.io/projected/494f1331-dcea-4745-add0-50794d9e38c9-kube-api-access-ttgzh\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.249988 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.268498 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: E1205 07:19:40.273848 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2" Dec 05 07:19:40 crc kubenswrapper[4997]: E1205 07:19:40.274137 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m75ld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-k8cjm_openstack(78c3f4c1-158a-4d14-96c2-f9b6a663fd45): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 07:19:40 crc kubenswrapper[4997]: E1205 07:19:40.275935 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-k8cjm" podUID="78c3f4c1-158a-4d14-96c2-f9b6a663fd45" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.530428 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.561580 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.757166 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe5443e-726b-4976-a7a3-c226bd3ec481","Type":"ContainerStarted","Data":"3f9c3f41c902d84f5da3f081f73987894ad168e0e10bf5cdaeb82c27e1353e2a"} Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.763706 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-h79p5" event={"ID":"3cf40633-ef49-416c-afcc-88c33fa76750","Type":"ContainerStarted","Data":"ecc9c395cc572c039778c1c4acd19fee3798c7ced60d872d9daba5d55a266d23"} Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.771241 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7dc7z"] Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.772767 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6lm2t" event={"ID":"d3a4d04e-bcc8-450c-830f-154f9637d513","Type":"ContainerStarted","Data":"161dca13d9cb4be2264cec96bcf12ea766cba1f66ad53b3908409e4612cc7f59"} Dec 05 07:19:40 crc kubenswrapper[4997]: E1205 07:19:40.781455 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2\\\"\"" pod="openstack/cinder-db-sync-k8cjm" podUID="78c3f4c1-158a-4d14-96c2-f9b6a663fd45" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.797054 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-h79p5" podStartSLOduration=2.646797875 podStartE2EDuration="25.797032718s" podCreationTimestamp="2025-12-05 07:19:15 +0000 UTC" firstStartedPulling="2025-12-05 07:19:17.037855425 +0000 UTC m=+1457.566762686" lastFinishedPulling="2025-12-05 07:19:40.188090268 +0000 UTC m=+1480.716997529" observedRunningTime="2025-12-05 07:19:40.790766828 +0000 UTC m=+1481.319674089" watchObservedRunningTime="2025-12-05 07:19:40.797032718 +0000 UTC m=+1481.325939989" Dec 05 07:19:40 crc kubenswrapper[4997]: I1205 07:19:40.846915 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-6lm2t" podStartSLOduration=2.696498083 podStartE2EDuration="25.846877861s" podCreationTimestamp="2025-12-05 07:19:15 +0000 UTC" firstStartedPulling="2025-12-05 07:19:17.037524645 +0000 UTC m=+1457.566431906" lastFinishedPulling="2025-12-05 07:19:40.187904423 +0000 UTC m=+1480.716811684" observedRunningTime="2025-12-05 07:19:40.814981222 +0000 UTC m=+1481.343888483" watchObservedRunningTime="2025-12-05 07:19:40.846877861 +0000 UTC m=+1481.375785122" Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.328161 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-92zcp" Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.364365 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.425643 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvhpv\" (UniqueName: \"kubernetes.io/projected/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-kube-api-access-jvhpv\") pod \"1e522d4c-e094-4518-bdb5-8ad8e3eccc97\" (UID: \"1e522d4c-e094-4518-bdb5-8ad8e3eccc97\") " Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.425769 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-combined-ca-bundle\") pod \"1e522d4c-e094-4518-bdb5-8ad8e3eccc97\" (UID: \"1e522d4c-e094-4518-bdb5-8ad8e3eccc97\") " Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.425914 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-config\") pod \"1e522d4c-e094-4518-bdb5-8ad8e3eccc97\" (UID: \"1e522d4c-e094-4518-bdb5-8ad8e3eccc97\") " Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.430863 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-kube-api-access-jvhpv" (OuterVolumeSpecName: "kube-api-access-jvhpv") pod "1e522d4c-e094-4518-bdb5-8ad8e3eccc97" (UID: "1e522d4c-e094-4518-bdb5-8ad8e3eccc97"). InnerVolumeSpecName "kube-api-access-jvhpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.462069 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e522d4c-e094-4518-bdb5-8ad8e3eccc97" (UID: "1e522d4c-e094-4518-bdb5-8ad8e3eccc97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.466835 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-config" (OuterVolumeSpecName: "config") pod "1e522d4c-e094-4518-bdb5-8ad8e3eccc97" (UID: "1e522d4c-e094-4518-bdb5-8ad8e3eccc97"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.531166 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.531203 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvhpv\" (UniqueName: \"kubernetes.io/projected/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-kube-api-access-jvhpv\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.531261 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e522d4c-e094-4518-bdb5-8ad8e3eccc97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.544862 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:19:41 crc kubenswrapper[4997]: W1205 07:19:41.553582 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ed6d9d7_6364_4dc5_b017_b9ba5e07f4e4.slice/crio-bf48eed91f67604c5ca7387bd16157d7eb043065e6137abaa35f1e5fd434d20f WatchSource:0}: Error finding container bf48eed91f67604c5ca7387bd16157d7eb043065e6137abaa35f1e5fd434d20f: Status 404 returned error can't find the container with id bf48eed91f67604c5ca7387bd16157d7eb043065e6137abaa35f1e5fd434d20f Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.838383 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8" path="/var/lib/kubelet/pods/0e79c8f4-0ffd-46c3-8d6d-ed67858fa2b8/volumes" Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.839833 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7098f936-a364-48df-922c-8202d2f4a80e" path="/var/lib/kubelet/pods/7098f936-a364-48df-922c-8202d2f4a80e/volumes" Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.840506 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf7639fd-80e4-4f37-8f92-9b9f73a01557" path="/var/lib/kubelet/pods/bf7639fd-80e4-4f37-8f92-9b9f73a01557/volumes" Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.910444 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-vbzgc"] Dec 05 07:19:41 crc kubenswrapper[4997]: E1205 07:19:41.912008 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e522d4c-e094-4518-bdb5-8ad8e3eccc97" containerName="neutron-db-sync" Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.912027 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e522d4c-e094-4518-bdb5-8ad8e3eccc97" containerName="neutron-db-sync" Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.912330 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e522d4c-e094-4518-bdb5-8ad8e3eccc97" containerName="neutron-db-sync" Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.915140 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.919014 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7dc7z" event={"ID":"145b182b-23c5-444d-864f-0cbd2c46902d","Type":"ContainerStarted","Data":"4130ec716a6c5c2d01f70c3a03e4d4ebc152a16367604a994406cb6eda7060ce"} Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.919103 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7dc7z" event={"ID":"145b182b-23c5-444d-864f-0cbd2c46902d","Type":"ContainerStarted","Data":"96f4181f5c75f9b04f7270aa36af33c57a28fd0e7f2dbcaa601802bd17139478"} Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.923907 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-vbzgc"] Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.926461 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4","Type":"ContainerStarted","Data":"bf48eed91f67604c5ca7387bd16157d7eb043065e6137abaa35f1e5fd434d20f"} Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.959145 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-92zcp" event={"ID":"1e522d4c-e094-4518-bdb5-8ad8e3eccc97","Type":"ContainerDied","Data":"fba9c55e23f58ada6acdc1dd2897927d67ae151a7e37e18f0949c102217318ed"} Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.959192 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fba9c55e23f58ada6acdc1dd2897927d67ae151a7e37e18f0949c102217318ed" Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.959254 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-92zcp" Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.985234 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"494f1331-dcea-4745-add0-50794d9e38c9","Type":"ContainerStarted","Data":"a5d9bd50afec7230b300e10c9fb8e9dccf422550391720d17984c22fb562aebe"} Dec 05 07:19:41 crc kubenswrapper[4997]: I1205 07:19:41.995889 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-7dc7z" podStartSLOduration=10.995863514 podStartE2EDuration="10.995863514s" podCreationTimestamp="2025-12-05 07:19:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:19:41.985879395 +0000 UTC m=+1482.514786656" watchObservedRunningTime="2025-12-05 07:19:41.995863514 +0000 UTC m=+1482.524770795" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.047065 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-64fc5c4f5b-mnq4n"] Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.049279 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.056160 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.056516 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.056721 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.056733 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-lgbp8" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.068045 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xwql\" (UniqueName: \"kubernetes.io/projected/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-kube-api-access-6xwql\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.068127 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-combined-ca-bundle\") pod \"neutron-64fc5c4f5b-mnq4n\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.068213 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-config\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.069097 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk8jl\" (UniqueName: \"kubernetes.io/projected/edf49641-b44d-4dba-af43-9a9da06eb55d-kube-api-access-wk8jl\") pod \"neutron-64fc5c4f5b-mnq4n\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.069153 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-dns-swift-storage-0\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.069237 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-ovndb-tls-certs\") pod \"neutron-64fc5c4f5b-mnq4n\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.069274 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-httpd-config\") pod \"neutron-64fc5c4f5b-mnq4n\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.069316 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-ovsdbserver-nb\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.069349 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-config\") pod \"neutron-64fc5c4f5b-mnq4n\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.069379 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-ovsdbserver-sb\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.069404 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-dns-svc\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.074318 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-64fc5c4f5b-mnq4n"] Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.122186 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-779c5847bc-vz755" podUID="bf7639fd-80e4-4f37-8f92-9b9f73a01557" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: i/o timeout" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.171580 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xwql\" (UniqueName: \"kubernetes.io/projected/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-kube-api-access-6xwql\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.171671 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-combined-ca-bundle\") pod \"neutron-64fc5c4f5b-mnq4n\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.171726 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-config\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.171768 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk8jl\" (UniqueName: \"kubernetes.io/projected/edf49641-b44d-4dba-af43-9a9da06eb55d-kube-api-access-wk8jl\") pod \"neutron-64fc5c4f5b-mnq4n\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.171805 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-dns-swift-storage-0\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.171855 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-ovndb-tls-certs\") pod \"neutron-64fc5c4f5b-mnq4n\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.171878 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-httpd-config\") pod \"neutron-64fc5c4f5b-mnq4n\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.171923 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-ovsdbserver-nb\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.171950 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-config\") pod \"neutron-64fc5c4f5b-mnq4n\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.171984 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-ovsdbserver-sb\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.172019 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-dns-svc\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.173138 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-dns-svc\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.177737 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-ovsdbserver-nb\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.181184 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-ovndb-tls-certs\") pod \"neutron-64fc5c4f5b-mnq4n\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.182057 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-config\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.183105 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-ovsdbserver-sb\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.189390 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-combined-ca-bundle\") pod \"neutron-64fc5c4f5b-mnq4n\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.190304 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-dns-swift-storage-0\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.197579 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-httpd-config\") pod \"neutron-64fc5c4f5b-mnq4n\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.200952 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk8jl\" (UniqueName: \"kubernetes.io/projected/edf49641-b44d-4dba-af43-9a9da06eb55d-kube-api-access-wk8jl\") pod \"neutron-64fc5c4f5b-mnq4n\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.201079 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xwql\" (UniqueName: \"kubernetes.io/projected/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-kube-api-access-6xwql\") pod \"dnsmasq-dns-77f55878d5-vbzgc\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.206772 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-config\") pod \"neutron-64fc5c4f5b-mnq4n\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.239634 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:42 crc kubenswrapper[4997]: I1205 07:19:42.392217 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:43 crc kubenswrapper[4997]: I1205 07:19:43.022081 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4","Type":"ContainerStarted","Data":"919328b940a3d4e6e28f76104075700a77242e6a1b78fc6e92aac2aa62da5236"} Dec 05 07:19:43 crc kubenswrapper[4997]: I1205 07:19:43.028305 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"494f1331-dcea-4745-add0-50794d9e38c9","Type":"ContainerStarted","Data":"60b0bdb40b4513ec19380aa4ea11f57950298ef219df1a43ffc7f22f094a77b2"} Dec 05 07:19:43 crc kubenswrapper[4997]: I1205 07:19:43.255415 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-vbzgc"] Dec 05 07:19:43 crc kubenswrapper[4997]: I1205 07:19:43.561651 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-64fc5c4f5b-mnq4n"] Dec 05 07:19:43 crc kubenswrapper[4997]: W1205 07:19:43.570706 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedf49641_b44d_4dba_af43_9a9da06eb55d.slice/crio-5baf1bf21c042acc4112da5f4d2072fb8d5c38aa4ba01691dc519354750c00ba WatchSource:0}: Error finding container 5baf1bf21c042acc4112da5f4d2072fb8d5c38aa4ba01691dc519354750c00ba: Status 404 returned error can't find the container with id 5baf1bf21c042acc4112da5f4d2072fb8d5c38aa4ba01691dc519354750c00ba Dec 05 07:19:44 crc kubenswrapper[4997]: I1205 07:19:44.042457 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64fc5c4f5b-mnq4n" event={"ID":"edf49641-b44d-4dba-af43-9a9da06eb55d","Type":"ContainerStarted","Data":"7a4dbab571ceebe87a020f8c86c1c678db709ecca7bb391c41b9909b0ab0648f"} Dec 05 07:19:44 crc kubenswrapper[4997]: I1205 07:19:44.044064 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64fc5c4f5b-mnq4n" event={"ID":"edf49641-b44d-4dba-af43-9a9da06eb55d","Type":"ContainerStarted","Data":"5baf1bf21c042acc4112da5f4d2072fb8d5c38aa4ba01691dc519354750c00ba"} Dec 05 07:19:44 crc kubenswrapper[4997]: I1205 07:19:44.048099 4997 generic.go:334] "Generic (PLEG): container finished" podID="ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e" containerID="8738a1e9499085a2414139ca014a0ee0f5ad8ed806324ba328c6a1fb7006b6b6" exitCode=0 Dec 05 07:19:44 crc kubenswrapper[4997]: I1205 07:19:44.048257 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" event={"ID":"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e","Type":"ContainerDied","Data":"8738a1e9499085a2414139ca014a0ee0f5ad8ed806324ba328c6a1fb7006b6b6"} Dec 05 07:19:44 crc kubenswrapper[4997]: I1205 07:19:44.048304 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" event={"ID":"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e","Type":"ContainerStarted","Data":"f9aa9b19a4539ead367b8e31efe2226124a659adbe7e0fe5dccde26730fcda2a"} Dec 05 07:19:44 crc kubenswrapper[4997]: I1205 07:19:44.066484 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4","Type":"ContainerStarted","Data":"48adac2190d73ab746eea173c0da78a53022e28b46d2c8de7580024fb3ddb353"} Dec 05 07:19:44 crc kubenswrapper[4997]: I1205 07:19:44.085462 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe5443e-726b-4976-a7a3-c226bd3ec481","Type":"ContainerStarted","Data":"6b46eaec9f50401eed6987b13e75bc96a2424eaf209614937ac23a0c2cc3c2a0"} Dec 05 07:19:44 crc kubenswrapper[4997]: I1205 07:19:44.094893 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"494f1331-dcea-4745-add0-50794d9e38c9","Type":"ContainerStarted","Data":"6d255e415e3ec2bac137bdf03a4ca5835d173b925ab6f3c048d7f48d26a75cae"} Dec 05 07:19:44 crc kubenswrapper[4997]: I1205 07:19:44.100370 4997 generic.go:334] "Generic (PLEG): container finished" podID="d3a4d04e-bcc8-450c-830f-154f9637d513" containerID="161dca13d9cb4be2264cec96bcf12ea766cba1f66ad53b3908409e4612cc7f59" exitCode=0 Dec 05 07:19:44 crc kubenswrapper[4997]: I1205 07:19:44.100452 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6lm2t" event={"ID":"d3a4d04e-bcc8-450c-830f-154f9637d513","Type":"ContainerDied","Data":"161dca13d9cb4be2264cec96bcf12ea766cba1f66ad53b3908409e4612cc7f59"} Dec 05 07:19:44 crc kubenswrapper[4997]: I1205 07:19:44.125688 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.125660503 podStartE2EDuration="5.125660503s" podCreationTimestamp="2025-12-05 07:19:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:19:44.114699828 +0000 UTC m=+1484.643607109" watchObservedRunningTime="2025-12-05 07:19:44.125660503 +0000 UTC m=+1484.654567754" Dec 05 07:19:44 crc kubenswrapper[4997]: I1205 07:19:44.187321 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.187292145 podStartE2EDuration="5.187292145s" podCreationTimestamp="2025-12-05 07:19:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:19:44.144380598 +0000 UTC m=+1484.673287859" watchObservedRunningTime="2025-12-05 07:19:44.187292145 +0000 UTC m=+1484.716199406" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.138873 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64fc5c4f5b-mnq4n" event={"ID":"edf49641-b44d-4dba-af43-9a9da06eb55d","Type":"ContainerStarted","Data":"d179214c82ee33c9b885acffd78f36623afd0e09d1efcc0df4a6be02aeb8e158"} Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.140301 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.160528 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" event={"ID":"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e","Type":"ContainerStarted","Data":"8a9193193aa72784b1caede2e5cb4ba050873269ad92e24cf4a5a1853ef29f74"} Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.160592 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.171814 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-64fc5c4f5b-mnq4n" podStartSLOduration=4.171779112 podStartE2EDuration="4.171779112s" podCreationTimestamp="2025-12-05 07:19:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:19:45.164384333 +0000 UTC m=+1485.693291614" watchObservedRunningTime="2025-12-05 07:19:45.171779112 +0000 UTC m=+1485.700686373" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.555475 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" podStartSLOduration=4.555445987 podStartE2EDuration="4.555445987s" podCreationTimestamp="2025-12-05 07:19:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:19:45.191512614 +0000 UTC m=+1485.720419895" watchObservedRunningTime="2025-12-05 07:19:45.555445987 +0000 UTC m=+1486.084353248" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.566771 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-74f5b8d45c-jbkcz"] Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.570579 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.574563 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.574818 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.602233 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-74f5b8d45c-jbkcz"] Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.608049 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.668821 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-combined-ca-bundle\") pod \"d3a4d04e-bcc8-450c-830f-154f9637d513\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.668929 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpw4z\" (UniqueName: \"kubernetes.io/projected/d3a4d04e-bcc8-450c-830f-154f9637d513-kube-api-access-hpw4z\") pod \"d3a4d04e-bcc8-450c-830f-154f9637d513\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.668989 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3a4d04e-bcc8-450c-830f-154f9637d513-logs\") pod \"d3a4d04e-bcc8-450c-830f-154f9637d513\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.669018 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-config-data\") pod \"d3a4d04e-bcc8-450c-830f-154f9637d513\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.669121 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-scripts\") pod \"d3a4d04e-bcc8-450c-830f-154f9637d513\" (UID: \"d3a4d04e-bcc8-450c-830f-154f9637d513\") " Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.669737 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-public-tls-certs\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.669823 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-httpd-config\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.669859 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gplnq\" (UniqueName: \"kubernetes.io/projected/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-kube-api-access-gplnq\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.669887 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-config\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.669924 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-ovndb-tls-certs\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.670003 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-internal-tls-certs\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.670029 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-combined-ca-bundle\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.671391 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3a4d04e-bcc8-450c-830f-154f9637d513-logs" (OuterVolumeSpecName: "logs") pod "d3a4d04e-bcc8-450c-830f-154f9637d513" (UID: "d3a4d04e-bcc8-450c-830f-154f9637d513"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.687497 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3a4d04e-bcc8-450c-830f-154f9637d513-kube-api-access-hpw4z" (OuterVolumeSpecName: "kube-api-access-hpw4z") pod "d3a4d04e-bcc8-450c-830f-154f9637d513" (UID: "d3a4d04e-bcc8-450c-830f-154f9637d513"). InnerVolumeSpecName "kube-api-access-hpw4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.712504 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-scripts" (OuterVolumeSpecName: "scripts") pod "d3a4d04e-bcc8-450c-830f-154f9637d513" (UID: "d3a4d04e-bcc8-450c-830f-154f9637d513"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.721888 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3a4d04e-bcc8-450c-830f-154f9637d513" (UID: "d3a4d04e-bcc8-450c-830f-154f9637d513"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.727912 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-config-data" (OuterVolumeSpecName: "config-data") pod "d3a4d04e-bcc8-450c-830f-154f9637d513" (UID: "d3a4d04e-bcc8-450c-830f-154f9637d513"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.809755 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-httpd-config\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.809841 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gplnq\" (UniqueName: \"kubernetes.io/projected/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-kube-api-access-gplnq\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.809880 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-config\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.809931 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-ovndb-tls-certs\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.810052 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-internal-tls-certs\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.810086 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-combined-ca-bundle\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.810353 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-public-tls-certs\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.815809 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3a4d04e-bcc8-450c-830f-154f9637d513-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.821747 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-public-tls-certs\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.823399 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-combined-ca-bundle\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.823488 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.823530 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.823546 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a4d04e-bcc8-450c-830f-154f9637d513-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.823570 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpw4z\" (UniqueName: \"kubernetes.io/projected/d3a4d04e-bcc8-450c-830f-154f9637d513-kube-api-access-hpw4z\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.826667 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-httpd-config\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.828163 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-ovndb-tls-certs\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.828455 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-config\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.847711 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-internal-tls-certs\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.849941 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gplnq\" (UniqueName: \"kubernetes.io/projected/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-kube-api-access-gplnq\") pod \"neutron-74f5b8d45c-jbkcz\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:45 crc kubenswrapper[4997]: I1205 07:19:45.935731 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.183559 4997 generic.go:334] "Generic (PLEG): container finished" podID="145b182b-23c5-444d-864f-0cbd2c46902d" containerID="4130ec716a6c5c2d01f70c3a03e4d4ebc152a16367604a994406cb6eda7060ce" exitCode=0 Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.184066 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7dc7z" event={"ID":"145b182b-23c5-444d-864f-0cbd2c46902d","Type":"ContainerDied","Data":"4130ec716a6c5c2d01f70c3a03e4d4ebc152a16367604a994406cb6eda7060ce"} Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.191664 4997 generic.go:334] "Generic (PLEG): container finished" podID="3cf40633-ef49-416c-afcc-88c33fa76750" containerID="ecc9c395cc572c039778c1c4acd19fee3798c7ced60d872d9daba5d55a266d23" exitCode=0 Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.191832 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-h79p5" event={"ID":"3cf40633-ef49-416c-afcc-88c33fa76750","Type":"ContainerDied","Data":"ecc9c395cc572c039778c1c4acd19fee3798c7ced60d872d9daba5d55a266d23"} Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.213671 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6lm2t" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.214036 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6lm2t" event={"ID":"d3a4d04e-bcc8-450c-830f-154f9637d513","Type":"ContainerDied","Data":"8fd8513dd403b29aca71ca24141cdaeba873404f796acf8c160b410cb6ce66f5"} Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.214109 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fd8513dd403b29aca71ca24141cdaeba873404f796acf8c160b410cb6ce66f5" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.309742 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-54ff4c8496-8z64m"] Dec 05 07:19:46 crc kubenswrapper[4997]: E1205 07:19:46.310410 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a4d04e-bcc8-450c-830f-154f9637d513" containerName="placement-db-sync" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.310428 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a4d04e-bcc8-450c-830f-154f9637d513" containerName="placement-db-sync" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.310716 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3a4d04e-bcc8-450c-830f-154f9637d513" containerName="placement-db-sync" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.312124 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.316063 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-54ff4c8496-8z64m"] Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.319073 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.319561 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-86x2z" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.319842 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.320057 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.321593 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.443921 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-74f5b8d45c-jbkcz"] Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.449375 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkptj\" (UniqueName: \"kubernetes.io/projected/2a733d85-ad49-41b8-a75b-842dce56e85c-kube-api-access-vkptj\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.449440 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-internal-tls-certs\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.449498 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-config-data\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.449601 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-combined-ca-bundle\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.449947 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a733d85-ad49-41b8-a75b-842dce56e85c-logs\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.450024 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-scripts\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.450061 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-public-tls-certs\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: W1205 07:19:46.456496 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc835ac5d_bd54_46d4_aab3_3ef7682f50b7.slice/crio-bd761217d241bd21cfb00057e7c3e3d35d519e845af67b85fb45144e1f373129 WatchSource:0}: Error finding container bd761217d241bd21cfb00057e7c3e3d35d519e845af67b85fb45144e1f373129: Status 404 returned error can't find the container with id bd761217d241bd21cfb00057e7c3e3d35d519e845af67b85fb45144e1f373129 Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.552691 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkptj\" (UniqueName: \"kubernetes.io/projected/2a733d85-ad49-41b8-a75b-842dce56e85c-kube-api-access-vkptj\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.552746 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-internal-tls-certs\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.552769 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-config-data\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.552837 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-combined-ca-bundle\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.552878 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a733d85-ad49-41b8-a75b-842dce56e85c-logs\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.552899 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-scripts\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.552959 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-public-tls-certs\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.554375 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a733d85-ad49-41b8-a75b-842dce56e85c-logs\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.558161 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-internal-tls-certs\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.558929 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-combined-ca-bundle\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.560361 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-scripts\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.560400 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-public-tls-certs\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.561532 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-config-data\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.571963 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkptj\" (UniqueName: \"kubernetes.io/projected/2a733d85-ad49-41b8-a75b-842dce56e85c-kube-api-access-vkptj\") pod \"placement-54ff4c8496-8z64m\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:46 crc kubenswrapper[4997]: I1205 07:19:46.676386 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:47 crc kubenswrapper[4997]: I1205 07:19:47.260737 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74f5b8d45c-jbkcz" event={"ID":"c835ac5d-bd54-46d4-aab3-3ef7682f50b7","Type":"ContainerStarted","Data":"bd761217d241bd21cfb00057e7c3e3d35d519e845af67b85fb45144e1f373129"} Dec 05 07:19:50 crc kubenswrapper[4997]: I1205 07:19:50.531280 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 07:19:50 crc kubenswrapper[4997]: I1205 07:19:50.532157 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 07:19:50 crc kubenswrapper[4997]: I1205 07:19:50.562994 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 07:19:50 crc kubenswrapper[4997]: I1205 07:19:50.563458 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 07:19:50 crc kubenswrapper[4997]: I1205 07:19:50.566534 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 07:19:50 crc kubenswrapper[4997]: I1205 07:19:50.581112 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 07:19:50 crc kubenswrapper[4997]: I1205 07:19:50.612154 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 07:19:50 crc kubenswrapper[4997]: I1205 07:19:50.626533 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 07:19:51 crc kubenswrapper[4997]: I1205 07:19:51.321204 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 07:19:51 crc kubenswrapper[4997]: I1205 07:19:51.321245 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 07:19:51 crc kubenswrapper[4997]: I1205 07:19:51.321256 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 07:19:51 crc kubenswrapper[4997]: I1205 07:19:51.321269 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 07:19:52 crc kubenswrapper[4997]: I1205 07:19:52.241978 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:19:52 crc kubenswrapper[4997]: I1205 07:19:52.358530 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-svg9r"] Dec 05 07:19:52 crc kubenswrapper[4997]: I1205 07:19:52.358928 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7884648fd9-svg9r" podUID="337e2c92-f622-4781-8a26-81d358d7baea" containerName="dnsmasq-dns" containerID="cri-o://28d20545adfcfd72e1756fb0dc1443224afda6a7f0dc28db51dd56d5b989674c" gracePeriod=10 Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.176536 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-h79p5" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.219309 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.317401 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3cf40633-ef49-416c-afcc-88c33fa76750-db-sync-config-data\") pod \"3cf40633-ef49-416c-afcc-88c33fa76750\" (UID: \"3cf40633-ef49-416c-afcc-88c33fa76750\") " Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.317868 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-config-data\") pod \"145b182b-23c5-444d-864f-0cbd2c46902d\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.317996 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwnxq\" (UniqueName: \"kubernetes.io/projected/145b182b-23c5-444d-864f-0cbd2c46902d-kube-api-access-xwnxq\") pod \"145b182b-23c5-444d-864f-0cbd2c46902d\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.318042 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-fernet-keys\") pod \"145b182b-23c5-444d-864f-0cbd2c46902d\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.318078 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-scripts\") pod \"145b182b-23c5-444d-864f-0cbd2c46902d\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.318109 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl4dm\" (UniqueName: \"kubernetes.io/projected/3cf40633-ef49-416c-afcc-88c33fa76750-kube-api-access-zl4dm\") pod \"3cf40633-ef49-416c-afcc-88c33fa76750\" (UID: \"3cf40633-ef49-416c-afcc-88c33fa76750\") " Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.318138 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-credential-keys\") pod \"145b182b-23c5-444d-864f-0cbd2c46902d\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.318265 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-combined-ca-bundle\") pod \"145b182b-23c5-444d-864f-0cbd2c46902d\" (UID: \"145b182b-23c5-444d-864f-0cbd2c46902d\") " Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.318370 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf40633-ef49-416c-afcc-88c33fa76750-combined-ca-bundle\") pod \"3cf40633-ef49-416c-afcc-88c33fa76750\" (UID: \"3cf40633-ef49-416c-afcc-88c33fa76750\") " Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.357877 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "145b182b-23c5-444d-864f-0cbd2c46902d" (UID: "145b182b-23c5-444d-864f-0cbd2c46902d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.358011 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cf40633-ef49-416c-afcc-88c33fa76750-kube-api-access-zl4dm" (OuterVolumeSpecName: "kube-api-access-zl4dm") pod "3cf40633-ef49-416c-afcc-88c33fa76750" (UID: "3cf40633-ef49-416c-afcc-88c33fa76750"). InnerVolumeSpecName "kube-api-access-zl4dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.358183 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-scripts" (OuterVolumeSpecName: "scripts") pod "145b182b-23c5-444d-864f-0cbd2c46902d" (UID: "145b182b-23c5-444d-864f-0cbd2c46902d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.358509 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/145b182b-23c5-444d-864f-0cbd2c46902d-kube-api-access-xwnxq" (OuterVolumeSpecName: "kube-api-access-xwnxq") pod "145b182b-23c5-444d-864f-0cbd2c46902d" (UID: "145b182b-23c5-444d-864f-0cbd2c46902d"). InnerVolumeSpecName "kube-api-access-xwnxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.365798 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf40633-ef49-416c-afcc-88c33fa76750-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3cf40633-ef49-416c-afcc-88c33fa76750" (UID: "3cf40633-ef49-416c-afcc-88c33fa76750"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.366278 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "145b182b-23c5-444d-864f-0cbd2c46902d" (UID: "145b182b-23c5-444d-864f-0cbd2c46902d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.396836 4997 generic.go:334] "Generic (PLEG): container finished" podID="337e2c92-f622-4781-8a26-81d358d7baea" containerID="28d20545adfcfd72e1756fb0dc1443224afda6a7f0dc28db51dd56d5b989674c" exitCode=0 Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.396914 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7884648fd9-svg9r" event={"ID":"337e2c92-f622-4781-8a26-81d358d7baea","Type":"ContainerDied","Data":"28d20545adfcfd72e1756fb0dc1443224afda6a7f0dc28db51dd56d5b989674c"} Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.417431 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "145b182b-23c5-444d-864f-0cbd2c46902d" (UID: "145b182b-23c5-444d-864f-0cbd2c46902d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.421314 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.421349 4997 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3cf40633-ef49-416c-afcc-88c33fa76750-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.421360 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwnxq\" (UniqueName: \"kubernetes.io/projected/145b182b-23c5-444d-864f-0cbd2c46902d-kube-api-access-xwnxq\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.421372 4997 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.421381 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.421390 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl4dm\" (UniqueName: \"kubernetes.io/projected/3cf40633-ef49-416c-afcc-88c33fa76750-kube-api-access-zl4dm\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.421398 4997 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.442310 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-54ff4c8496-8z64m"] Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.445341 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-config-data" (OuterVolumeSpecName: "config-data") pod "145b182b-23c5-444d-864f-0cbd2c46902d" (UID: "145b182b-23c5-444d-864f-0cbd2c46902d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.451935 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-h79p5" event={"ID":"3cf40633-ef49-416c-afcc-88c33fa76750","Type":"ContainerDied","Data":"9436cb2bb18829ab26ddd4408fbf547030e5faa667d4d1940682f0e87ed35eeb"} Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.451988 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9436cb2bb18829ab26ddd4408fbf547030e5faa667d4d1940682f0e87ed35eeb" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.452087 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-h79p5" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.466009 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7dc7z" event={"ID":"145b182b-23c5-444d-864f-0cbd2c46902d","Type":"ContainerDied","Data":"96f4181f5c75f9b04f7270aa36af33c57a28fd0e7f2dbcaa601802bd17139478"} Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.466446 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96f4181f5c75f9b04f7270aa36af33c57a28fd0e7f2dbcaa601802bd17139478" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.466531 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7dc7z" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.477246 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.482220 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf40633-ef49-416c-afcc-88c33fa76750-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3cf40633-ef49-416c-afcc-88c33fa76750" (UID: "3cf40633-ef49-416c-afcc-88c33fa76750"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.523315 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf40633-ef49-416c-afcc-88c33fa76750-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.523357 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/145b182b-23c5-444d-864f-0cbd2c46902d-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.624156 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-dns-svc\") pod \"337e2c92-f622-4781-8a26-81d358d7baea\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.624263 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs9ck\" (UniqueName: \"kubernetes.io/projected/337e2c92-f622-4781-8a26-81d358d7baea-kube-api-access-rs9ck\") pod \"337e2c92-f622-4781-8a26-81d358d7baea\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.624344 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-dns-swift-storage-0\") pod \"337e2c92-f622-4781-8a26-81d358d7baea\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.624389 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-ovsdbserver-nb\") pod \"337e2c92-f622-4781-8a26-81d358d7baea\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.624512 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-ovsdbserver-sb\") pod \"337e2c92-f622-4781-8a26-81d358d7baea\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.624576 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-config\") pod \"337e2c92-f622-4781-8a26-81d358d7baea\" (UID: \"337e2c92-f622-4781-8a26-81d358d7baea\") " Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.674383 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/337e2c92-f622-4781-8a26-81d358d7baea-kube-api-access-rs9ck" (OuterVolumeSpecName: "kube-api-access-rs9ck") pod "337e2c92-f622-4781-8a26-81d358d7baea" (UID: "337e2c92-f622-4781-8a26-81d358d7baea"). InnerVolumeSpecName "kube-api-access-rs9ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.729250 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs9ck\" (UniqueName: \"kubernetes.io/projected/337e2c92-f622-4781-8a26-81d358d7baea-kube-api-access-rs9ck\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.729904 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "337e2c92-f622-4781-8a26-81d358d7baea" (UID: "337e2c92-f622-4781-8a26-81d358d7baea"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.733773 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-config" (OuterVolumeSpecName: "config") pod "337e2c92-f622-4781-8a26-81d358d7baea" (UID: "337e2c92-f622-4781-8a26-81d358d7baea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.751629 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "337e2c92-f622-4781-8a26-81d358d7baea" (UID: "337e2c92-f622-4781-8a26-81d358d7baea"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.761055 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "337e2c92-f622-4781-8a26-81d358d7baea" (UID: "337e2c92-f622-4781-8a26-81d358d7baea"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.777237 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "337e2c92-f622-4781-8a26-81d358d7baea" (UID: "337e2c92-f622-4781-8a26-81d358d7baea"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.834473 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.834510 4997 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.834523 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.834534 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:53 crc kubenswrapper[4997]: I1205 07:19:53.834544 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/337e2c92-f622-4781-8a26-81d358d7baea-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:19:53 crc kubenswrapper[4997]: E1205 07:19:53.944500 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod145b182b_23c5_444d_864f_0cbd2c46902d.slice/crio-96f4181f5c75f9b04f7270aa36af33c57a28fd0e7f2dbcaa601802bd17139478\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod145b182b_23c5_444d_864f_0cbd2c46902d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3cf40633_ef49_416c_afcc_88c33fa76750.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3cf40633_ef49_416c_afcc_88c33fa76750.slice/crio-9436cb2bb18829ab26ddd4408fbf547030e5faa667d4d1940682f0e87ed35eeb\": RecentStats: unable to find data in memory cache]" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.395051 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5477474c5d-zs25t"] Dec 05 07:19:54 crc kubenswrapper[4997]: E1205 07:19:54.395846 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="337e2c92-f622-4781-8a26-81d358d7baea" containerName="dnsmasq-dns" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.395863 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="337e2c92-f622-4781-8a26-81d358d7baea" containerName="dnsmasq-dns" Dec 05 07:19:54 crc kubenswrapper[4997]: E1205 07:19:54.395896 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cf40633-ef49-416c-afcc-88c33fa76750" containerName="barbican-db-sync" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.395904 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cf40633-ef49-416c-afcc-88c33fa76750" containerName="barbican-db-sync" Dec 05 07:19:54 crc kubenswrapper[4997]: E1205 07:19:54.395915 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="337e2c92-f622-4781-8a26-81d358d7baea" containerName="init" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.395924 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="337e2c92-f622-4781-8a26-81d358d7baea" containerName="init" Dec 05 07:19:54 crc kubenswrapper[4997]: E1205 07:19:54.395931 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="145b182b-23c5-444d-864f-0cbd2c46902d" containerName="keystone-bootstrap" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.395937 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="145b182b-23c5-444d-864f-0cbd2c46902d" containerName="keystone-bootstrap" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.396109 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="145b182b-23c5-444d-864f-0cbd2c46902d" containerName="keystone-bootstrap" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.396127 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cf40633-ef49-416c-afcc-88c33fa76750" containerName="barbican-db-sync" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.396139 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="337e2c92-f622-4781-8a26-81d358d7baea" containerName="dnsmasq-dns" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.396798 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.400154 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.400356 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-856gn" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.403280 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.414902 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.415473 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.415834 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.427919 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5477474c5d-zs25t"] Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.563467 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-credential-keys\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.563550 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-combined-ca-bundle\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.563581 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-internal-tls-certs\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.563626 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2tjr\" (UniqueName: \"kubernetes.io/projected/6783d79d-4874-43c1-92bc-e09133e45989-kube-api-access-b2tjr\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.563687 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-scripts\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.563762 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-fernet-keys\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.563865 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-config-data\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.563916 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-public-tls-certs\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.583810 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74f5b8d45c-jbkcz" event={"ID":"c835ac5d-bd54-46d4-aab3-3ef7682f50b7","Type":"ContainerStarted","Data":"7f4731817ad8f583546eee8efc40ce792a1f83dad3127d8c88136943c5c893be"} Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.583866 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74f5b8d45c-jbkcz" event={"ID":"c835ac5d-bd54-46d4-aab3-3ef7682f50b7","Type":"ContainerStarted","Data":"8ffd311c9d4ec727b6a0cad0412afdc5d1ea879f7894edd49a446d48d71d135f"} Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.584493 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.595508 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-555c8b7b77-44qw4"] Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.597883 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.623006 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-555c8b7b77-44qw4"] Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.623933 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-54ff4c8496-8z64m" event={"ID":"2a733d85-ad49-41b8-a75b-842dce56e85c","Type":"ContainerStarted","Data":"08946f5deeba339b1b854496b842f982af5b2b54721ebee20c813990c70fbe66"} Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.623992 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-54ff4c8496-8z64m" event={"ID":"2a733d85-ad49-41b8-a75b-842dce56e85c","Type":"ContainerStarted","Data":"ba7342d2d5bfb967137e1fd9c51e48f142b7847eec802846b07240462ab579d3"} Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.624003 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-54ff4c8496-8z64m" event={"ID":"2a733d85-ad49-41b8-a75b-842dce56e85c","Type":"ContainerStarted","Data":"954c36632faeae217ead931414bfae62826ed6239cd9550f2be576495871cc20"} Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.624063 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.624143 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.633210 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-jmntt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.633587 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.633781 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.634763 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-db8868644-mchk9"] Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.636598 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.643283 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-db8868644-mchk9"] Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.648529 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7884648fd9-svg9r" event={"ID":"337e2c92-f622-4781-8a26-81d358d7baea","Type":"ContainerDied","Data":"f611c5363a1ca2b236c2eace6c685f3ddd3d473a869c4efd55ea04a162edefdb"} Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.648599 4997 scope.go:117] "RemoveContainer" containerID="28d20545adfcfd72e1756fb0dc1443224afda6a7f0dc28db51dd56d5b989674c" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.648800 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7884648fd9-svg9r" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.658194 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.672364 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-logs\") pod \"barbican-keystone-listener-db8868644-mchk9\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.672442 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-credential-keys\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.672486 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6bfb8b2-df28-4992-8831-2a5060323ddf-logs\") pod \"barbican-worker-555c8b7b77-44qw4\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.672505 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-combined-ca-bundle\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.672525 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-internal-tls-certs\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.672545 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2tjr\" (UniqueName: \"kubernetes.io/projected/6783d79d-4874-43c1-92bc-e09133e45989-kube-api-access-b2tjr\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.672594 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhpd4\" (UniqueName: \"kubernetes.io/projected/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-kube-api-access-vhpd4\") pod \"barbican-keystone-listener-db8868644-mchk9\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.672629 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-config-data\") pod \"barbican-keystone-listener-db8868644-mchk9\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.672652 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-scripts\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.672676 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-combined-ca-bundle\") pod \"barbican-keystone-listener-db8868644-mchk9\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.672695 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-combined-ca-bundle\") pod \"barbican-worker-555c8b7b77-44qw4\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.672727 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-config-data-custom\") pod \"barbican-keystone-listener-db8868644-mchk9\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.672770 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-fernet-keys\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.672807 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-config-data\") pod \"barbican-worker-555c8b7b77-44qw4\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.672873 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-config-data-custom\") pod \"barbican-worker-555c8b7b77-44qw4\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.672953 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-config-data\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.672984 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gn2g\" (UniqueName: \"kubernetes.io/projected/a6bfb8b2-df28-4992-8831-2a5060323ddf-kube-api-access-4gn2g\") pod \"barbican-worker-555c8b7b77-44qw4\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.673006 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-public-tls-certs\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.710876 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-74f5b8d45c-jbkcz" podStartSLOduration=9.71082802 podStartE2EDuration="9.71082802s" podCreationTimestamp="2025-12-05 07:19:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:19:54.673240564 +0000 UTC m=+1495.202147825" watchObservedRunningTime="2025-12-05 07:19:54.71082802 +0000 UTC m=+1495.239735271" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.731712 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe5443e-726b-4976-a7a3-c226bd3ec481","Type":"ContainerStarted","Data":"5a0093fba43766bdbd2468af37a378fd8ab58701871d81188044c42c82c86057"} Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.734385 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-internal-tls-certs\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.734450 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-config-data\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.734988 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-fernet-keys\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.735083 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-public-tls-certs\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.735506 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-combined-ca-bundle\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.742027 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2tjr\" (UniqueName: \"kubernetes.io/projected/6783d79d-4874-43c1-92bc-e09133e45989-kube-api-access-b2tjr\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.747776 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-credential-keys\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.752310 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-scripts\") pod \"keystone-5477474c5d-zs25t\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.771218 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-jz4qt"] Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.773692 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.774561 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-logs\") pod \"barbican-keystone-listener-db8868644-mchk9\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.774661 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6bfb8b2-df28-4992-8831-2a5060323ddf-logs\") pod \"barbican-worker-555c8b7b77-44qw4\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.774702 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhpd4\" (UniqueName: \"kubernetes.io/projected/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-kube-api-access-vhpd4\") pod \"barbican-keystone-listener-db8868644-mchk9\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.774723 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-config-data\") pod \"barbican-keystone-listener-db8868644-mchk9\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.774742 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-combined-ca-bundle\") pod \"barbican-keystone-listener-db8868644-mchk9\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.774764 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-combined-ca-bundle\") pod \"barbican-worker-555c8b7b77-44qw4\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.774785 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-config-data-custom\") pod \"barbican-keystone-listener-db8868644-mchk9\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.774821 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-config-data\") pod \"barbican-worker-555c8b7b77-44qw4\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.774855 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-config-data-custom\") pod \"barbican-worker-555c8b7b77-44qw4\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.774893 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gn2g\" (UniqueName: \"kubernetes.io/projected/a6bfb8b2-df28-4992-8831-2a5060323ddf-kube-api-access-4gn2g\") pod \"barbican-worker-555c8b7b77-44qw4\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.783355 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6bfb8b2-df28-4992-8831-2a5060323ddf-logs\") pod \"barbican-worker-555c8b7b77-44qw4\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.783723 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-logs\") pod \"barbican-keystone-listener-db8868644-mchk9\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.804896 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-config-data\") pod \"barbican-worker-555c8b7b77-44qw4\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.806442 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-combined-ca-bundle\") pod \"barbican-worker-555c8b7b77-44qw4\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.818791 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-config-data-custom\") pod \"barbican-keystone-listener-db8868644-mchk9\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.828021 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-config-data\") pod \"barbican-keystone-listener-db8868644-mchk9\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.846846 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-jz4qt"] Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.848648 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-config-data-custom\") pod \"barbican-worker-555c8b7b77-44qw4\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.848844 4997 scope.go:117] "RemoveContainer" containerID="8471674046ec6cb711d5c75b3b5646e1ecf75702d27e3f7b07f8200edf153d13" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.849405 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-combined-ca-bundle\") pod \"barbican-keystone-listener-db8868644-mchk9\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.853502 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhpd4\" (UniqueName: \"kubernetes.io/projected/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-kube-api-access-vhpd4\") pod \"barbican-keystone-listener-db8868644-mchk9\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.872974 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gn2g\" (UniqueName: \"kubernetes.io/projected/a6bfb8b2-df28-4992-8831-2a5060323ddf-kube-api-access-4gn2g\") pod \"barbican-worker-555c8b7b77-44qw4\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.874130 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-54ff4c8496-8z64m" podStartSLOduration=8.874099393 podStartE2EDuration="8.874099393s" podCreationTimestamp="2025-12-05 07:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:19:54.730389378 +0000 UTC m=+1495.259296649" watchObservedRunningTime="2025-12-05 07:19:54.874099393 +0000 UTC m=+1495.403006654" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.883672 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-dns-swift-storage-0\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.883994 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5lm9\" (UniqueName: \"kubernetes.io/projected/81bdd956-a535-4f79-9188-528dc91b0a1d-kube-api-access-z5lm9\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.884243 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-dns-svc\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.884404 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-config\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.884673 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-ovsdbserver-sb\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.884887 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-ovsdbserver-nb\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.931326 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.945958 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-57bc457fc6-28p4d"] Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.956175 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.959810 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-svg9r"] Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.960023 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.970668 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-svg9r"] Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.984377 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57bc457fc6-28p4d"] Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.986752 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-ovsdbserver-sb\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.986830 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-ovsdbserver-nb\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.986860 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-dns-swift-storage-0\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.986877 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5lm9\" (UniqueName: \"kubernetes.io/projected/81bdd956-a535-4f79-9188-528dc91b0a1d-kube-api-access-z5lm9\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.986930 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-dns-svc\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.986957 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-config\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.989556 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-config\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.989731 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-dns-swift-storage-0\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.990433 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-ovsdbserver-nb\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.990762 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-dns-svc\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:54 crc kubenswrapper[4997]: I1205 07:19:54.990939 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-ovsdbserver-sb\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.012097 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.022401 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5lm9\" (UniqueName: \"kubernetes.io/projected/81bdd956-a535-4f79-9188-528dc91b0a1d-kube-api-access-z5lm9\") pod \"dnsmasq-dns-687dbb56f-jz4qt\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.041965 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.057378 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.057523 4997 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.094978 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-config-data\") pod \"barbican-api-57bc457fc6-28p4d\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.095075 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-combined-ca-bundle\") pod \"barbican-api-57bc457fc6-28p4d\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.095148 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-config-data-custom\") pod \"barbican-api-57bc457fc6-28p4d\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.095203 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnghc\" (UniqueName: \"kubernetes.io/projected/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-kube-api-access-bnghc\") pod \"barbican-api-57bc457fc6-28p4d\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.095277 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-logs\") pod \"barbican-api-57bc457fc6-28p4d\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.204471 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-config-data-custom\") pod \"barbican-api-57bc457fc6-28p4d\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.204557 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnghc\" (UniqueName: \"kubernetes.io/projected/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-kube-api-access-bnghc\") pod \"barbican-api-57bc457fc6-28p4d\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.204600 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-logs\") pod \"barbican-api-57bc457fc6-28p4d\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.204744 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-config-data\") pod \"barbican-api-57bc457fc6-28p4d\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.204800 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-combined-ca-bundle\") pod \"barbican-api-57bc457fc6-28p4d\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.205209 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-logs\") pod \"barbican-api-57bc457fc6-28p4d\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.215031 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-config-data\") pod \"barbican-api-57bc457fc6-28p4d\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.216136 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-config-data-custom\") pod \"barbican-api-57bc457fc6-28p4d\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.241110 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-combined-ca-bundle\") pod \"barbican-api-57bc457fc6-28p4d\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.249182 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.260629 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnghc\" (UniqueName: \"kubernetes.io/projected/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-kube-api-access-bnghc\") pod \"barbican-api-57bc457fc6-28p4d\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.357463 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.357578 4997 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.359602 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.376720 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.453550 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.785911 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="337e2c92-f622-4781-8a26-81d358d7baea" path="/var/lib/kubelet/pods/337e2c92-f622-4781-8a26-81d358d7baea/volumes" Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.825149 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-555c8b7b77-44qw4"] Dec 05 07:19:55 crc kubenswrapper[4997]: I1205 07:19:55.892997 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-db8868644-mchk9"] Dec 05 07:19:56 crc kubenswrapper[4997]: I1205 07:19:56.159983 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-jz4qt"] Dec 05 07:19:56 crc kubenswrapper[4997]: I1205 07:19:56.169753 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5477474c5d-zs25t"] Dec 05 07:19:56 crc kubenswrapper[4997]: I1205 07:19:56.243726 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57bc457fc6-28p4d"] Dec 05 07:19:56 crc kubenswrapper[4997]: I1205 07:19:56.835053 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-k8cjm" event={"ID":"78c3f4c1-158a-4d14-96c2-f9b6a663fd45","Type":"ContainerStarted","Data":"8a63d40925c22c61026b0de2d63c2a59394ad5056824d09c3f51cc7b991b8313"} Dec 05 07:19:56 crc kubenswrapper[4997]: I1205 07:19:56.862349 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57bc457fc6-28p4d" event={"ID":"6aecdebd-cbaf-4af2-acf4-06f9321b9a37","Type":"ContainerStarted","Data":"cd2440f9aa5e210d229a0413c490c0d93af695ff6e68a1e329a3902622ad4d5f"} Dec 05 07:19:56 crc kubenswrapper[4997]: I1205 07:19:56.862440 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57bc457fc6-28p4d" event={"ID":"6aecdebd-cbaf-4af2-acf4-06f9321b9a37","Type":"ContainerStarted","Data":"22f08e4078c712ccb3dc0aff6dc4f9645b5d77e7283268b483fd1a74fe7c74e1"} Dec 05 07:19:56 crc kubenswrapper[4997]: I1205 07:19:56.875244 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5477474c5d-zs25t" event={"ID":"6783d79d-4874-43c1-92bc-e09133e45989","Type":"ContainerStarted","Data":"99025ec523a5c575412ef4a88b541702a67a5c68296982dfdc6d2fccabd2bb56"} Dec 05 07:19:56 crc kubenswrapper[4997]: I1205 07:19:56.875334 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5477474c5d-zs25t" event={"ID":"6783d79d-4874-43c1-92bc-e09133e45989","Type":"ContainerStarted","Data":"61c17f937766b0c229e3792ae938a47d02305aa8b1b0a03c6224aad70b14169a"} Dec 05 07:19:56 crc kubenswrapper[4997]: I1205 07:19:56.876050 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:19:56 crc kubenswrapper[4997]: I1205 07:19:56.880017 4997 generic.go:334] "Generic (PLEG): container finished" podID="81bdd956-a535-4f79-9188-528dc91b0a1d" containerID="51ec5e9c7d9204cde67cc653b0c3bb7be47f74896a9ed1f9d466229c0a281a37" exitCode=0 Dec 05 07:19:56 crc kubenswrapper[4997]: I1205 07:19:56.880142 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" event={"ID":"81bdd956-a535-4f79-9188-528dc91b0a1d","Type":"ContainerDied","Data":"51ec5e9c7d9204cde67cc653b0c3bb7be47f74896a9ed1f9d466229c0a281a37"} Dec 05 07:19:56 crc kubenswrapper[4997]: I1205 07:19:56.880170 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" event={"ID":"81bdd956-a535-4f79-9188-528dc91b0a1d","Type":"ContainerStarted","Data":"bb0ae538937ce6a38824254600a433da51ec4c16b3f0b070d618973838f88cb0"} Dec 05 07:19:56 crc kubenswrapper[4997]: I1205 07:19:56.888560 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-555c8b7b77-44qw4" event={"ID":"a6bfb8b2-df28-4992-8831-2a5060323ddf","Type":"ContainerStarted","Data":"a602e9160cb6a58142ca3e52d7ec8d6dd98eafd893d7a66da580770ef4ba8ac2"} Dec 05 07:19:56 crc kubenswrapper[4997]: I1205 07:19:56.911088 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-k8cjm" podStartSLOduration=4.55793802 podStartE2EDuration="41.911060386s" podCreationTimestamp="2025-12-05 07:19:15 +0000 UTC" firstStartedPulling="2025-12-05 07:19:16.802753935 +0000 UTC m=+1457.331661196" lastFinishedPulling="2025-12-05 07:19:54.155876311 +0000 UTC m=+1494.684783562" observedRunningTime="2025-12-05 07:19:56.862066502 +0000 UTC m=+1497.390973763" watchObservedRunningTime="2025-12-05 07:19:56.911060386 +0000 UTC m=+1497.439967647" Dec 05 07:19:56 crc kubenswrapper[4997]: I1205 07:19:56.911190 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-db8868644-mchk9" event={"ID":"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1","Type":"ContainerStarted","Data":"be77616b5c8937572e9e713ab0982f57a84b605ed475f48883932d374be43cfa"} Dec 05 07:19:56 crc kubenswrapper[4997]: I1205 07:19:56.934102 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5477474c5d-zs25t" podStartSLOduration=2.934063857 podStartE2EDuration="2.934063857s" podCreationTimestamp="2025-12-05 07:19:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:19:56.906245035 +0000 UTC m=+1497.435152306" watchObservedRunningTime="2025-12-05 07:19:56.934063857 +0000 UTC m=+1497.462971118" Dec 05 07:19:57 crc kubenswrapper[4997]: I1205 07:19:57.954160 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57bc457fc6-28p4d" event={"ID":"6aecdebd-cbaf-4af2-acf4-06f9321b9a37","Type":"ContainerStarted","Data":"807fb7f45b92c088f92ec791ea2cb20347da760ff331b4691df3d6cd5e5e7ce6"} Dec 05 07:19:57 crc kubenswrapper[4997]: I1205 07:19:57.955980 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:57 crc kubenswrapper[4997]: I1205 07:19:57.956008 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:19:57 crc kubenswrapper[4997]: I1205 07:19:57.971157 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" event={"ID":"81bdd956-a535-4f79-9188-528dc91b0a1d","Type":"ContainerStarted","Data":"d8d919c1f1f3e3dcc58ccb6121268482f2e3f8f5f2283c067b9e44b58bff4773"} Dec 05 07:19:57 crc kubenswrapper[4997]: I1205 07:19:57.971625 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:19:57 crc kubenswrapper[4997]: I1205 07:19:57.987127 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-57bc457fc6-28p4d" podStartSLOduration=3.987099648 podStartE2EDuration="3.987099648s" podCreationTimestamp="2025-12-05 07:19:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:19:57.977603811 +0000 UTC m=+1498.506511092" watchObservedRunningTime="2025-12-05 07:19:57.987099648 +0000 UTC m=+1498.516006909" Dec 05 07:19:58 crc kubenswrapper[4997]: I1205 07:19:58.014014 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" podStartSLOduration=4.013665876 podStartE2EDuration="4.013665876s" podCreationTimestamp="2025-12-05 07:19:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:19:58.010993854 +0000 UTC m=+1498.539901125" watchObservedRunningTime="2025-12-05 07:19:58.013665876 +0000 UTC m=+1498.542573137" Dec 05 07:19:58 crc kubenswrapper[4997]: I1205 07:19:58.964710 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5cb79f8cdd-zkk89"] Dec 05 07:19:58 crc kubenswrapper[4997]: I1205 07:19:58.971577 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:58 crc kubenswrapper[4997]: I1205 07:19:58.979050 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 05 07:19:58 crc kubenswrapper[4997]: I1205 07:19:58.979317 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 05 07:19:58 crc kubenswrapper[4997]: I1205 07:19:58.991827 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5cb79f8cdd-zkk89"] Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.047278 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-combined-ca-bundle\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.047334 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2179d060-8f4c-413d-a202-e31ebc242dfc-logs\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.047385 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-config-data-custom\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.047446 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-internal-tls-certs\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.047493 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-public-tls-certs\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.047779 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ddxf\" (UniqueName: \"kubernetes.io/projected/2179d060-8f4c-413d-a202-e31ebc242dfc-kube-api-access-9ddxf\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.047848 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-config-data\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.149965 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-internal-tls-certs\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.151661 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-public-tls-certs\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.151727 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ddxf\" (UniqueName: \"kubernetes.io/projected/2179d060-8f4c-413d-a202-e31ebc242dfc-kube-api-access-9ddxf\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.151880 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-config-data\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.151993 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-combined-ca-bundle\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.152039 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2179d060-8f4c-413d-a202-e31ebc242dfc-logs\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.152136 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-config-data-custom\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.153634 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2179d060-8f4c-413d-a202-e31ebc242dfc-logs\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.161480 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-public-tls-certs\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.161688 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-internal-tls-certs\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.162139 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-combined-ca-bundle\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.162916 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-config-data-custom\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.163860 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-config-data\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.172382 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ddxf\" (UniqueName: \"kubernetes.io/projected/2179d060-8f4c-413d-a202-e31ebc242dfc-kube-api-access-9ddxf\") pod \"barbican-api-5cb79f8cdd-zkk89\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.368913 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:19:59 crc kubenswrapper[4997]: W1205 07:19:59.874474 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2179d060_8f4c_413d_a202_e31ebc242dfc.slice/crio-c89a73a98da50d8a158529471b379a982fbb6c19bed2b42f34c604c7ff17f3b3 WatchSource:0}: Error finding container c89a73a98da50d8a158529471b379a982fbb6c19bed2b42f34c604c7ff17f3b3: Status 404 returned error can't find the container with id c89a73a98da50d8a158529471b379a982fbb6c19bed2b42f34c604c7ff17f3b3 Dec 05 07:19:59 crc kubenswrapper[4997]: I1205 07:19:59.875167 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5cb79f8cdd-zkk89"] Dec 05 07:20:00 crc kubenswrapper[4997]: I1205 07:20:00.032027 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-db8868644-mchk9" event={"ID":"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1","Type":"ContainerStarted","Data":"63372a7e68a9d93e249283016625478d1f7e92640812a3890997060ed1267282"} Dec 05 07:20:00 crc kubenswrapper[4997]: I1205 07:20:00.032091 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-db8868644-mchk9" event={"ID":"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1","Type":"ContainerStarted","Data":"32e855f8095dea6a97466b7d110e63c5065aa57883dd146dc7d5cd8bf980f1a0"} Dec 05 07:20:00 crc kubenswrapper[4997]: I1205 07:20:00.035407 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cb79f8cdd-zkk89" event={"ID":"2179d060-8f4c-413d-a202-e31ebc242dfc","Type":"ContainerStarted","Data":"c89a73a98da50d8a158529471b379a982fbb6c19bed2b42f34c604c7ff17f3b3"} Dec 05 07:20:00 crc kubenswrapper[4997]: I1205 07:20:00.039333 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-555c8b7b77-44qw4" event={"ID":"a6bfb8b2-df28-4992-8831-2a5060323ddf","Type":"ContainerStarted","Data":"b77018d13ada4b2350299017976f4a2b5e3d76cc39ecbc054f294f67cbfa4f8f"} Dec 05 07:20:00 crc kubenswrapper[4997]: I1205 07:20:00.039387 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-555c8b7b77-44qw4" event={"ID":"a6bfb8b2-df28-4992-8831-2a5060323ddf","Type":"ContainerStarted","Data":"af3bd6a46ce774d4927e32dd1f7727318d24638c24860778360780d82ef22b47"} Dec 05 07:20:00 crc kubenswrapper[4997]: I1205 07:20:00.055079 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-db8868644-mchk9" podStartSLOduration=3.121926506 podStartE2EDuration="6.05505572s" podCreationTimestamp="2025-12-05 07:19:54 +0000 UTC" firstStartedPulling="2025-12-05 07:19:55.957195026 +0000 UTC m=+1496.486102287" lastFinishedPulling="2025-12-05 07:19:58.89032424 +0000 UTC m=+1499.419231501" observedRunningTime="2025-12-05 07:20:00.051414791 +0000 UTC m=+1500.580322072" watchObservedRunningTime="2025-12-05 07:20:00.05505572 +0000 UTC m=+1500.583962981" Dec 05 07:20:00 crc kubenswrapper[4997]: I1205 07:20:00.081797 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-555c8b7b77-44qw4" podStartSLOduration=3.048208314 podStartE2EDuration="6.081775672s" podCreationTimestamp="2025-12-05 07:19:54 +0000 UTC" firstStartedPulling="2025-12-05 07:19:55.854882831 +0000 UTC m=+1496.383790092" lastFinishedPulling="2025-12-05 07:19:58.888450189 +0000 UTC m=+1499.417357450" observedRunningTime="2025-12-05 07:20:00.073465927 +0000 UTC m=+1500.602373198" watchObservedRunningTime="2025-12-05 07:20:00.081775672 +0000 UTC m=+1500.610682933" Dec 05 07:20:01 crc kubenswrapper[4997]: I1205 07:20:01.058213 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cb79f8cdd-zkk89" event={"ID":"2179d060-8f4c-413d-a202-e31ebc242dfc","Type":"ContainerStarted","Data":"d14f08fb852081fdc4815d2ce697e3418dfd98d7534ab2afb3c35a5bfea2457a"} Dec 05 07:20:01 crc kubenswrapper[4997]: I1205 07:20:01.058627 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cb79f8cdd-zkk89" event={"ID":"2179d060-8f4c-413d-a202-e31ebc242dfc","Type":"ContainerStarted","Data":"5c7c977eca43b417984db669a02fedc758e9ee1c5a3cc1077d207c4c29b0e85d"} Dec 05 07:20:01 crc kubenswrapper[4997]: I1205 07:20:01.059205 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:20:01 crc kubenswrapper[4997]: I1205 07:20:01.059279 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:20:01 crc kubenswrapper[4997]: I1205 07:20:01.099813 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5cb79f8cdd-zkk89" podStartSLOduration=3.099790666 podStartE2EDuration="3.099790666s" podCreationTimestamp="2025-12-05 07:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:20:01.092104039 +0000 UTC m=+1501.621011300" watchObservedRunningTime="2025-12-05 07:20:01.099790666 +0000 UTC m=+1501.628697927" Dec 05 07:20:02 crc kubenswrapper[4997]: I1205 07:20:02.069676 4997 generic.go:334] "Generic (PLEG): container finished" podID="78c3f4c1-158a-4d14-96c2-f9b6a663fd45" containerID="8a63d40925c22c61026b0de2d63c2a59394ad5056824d09c3f51cc7b991b8313" exitCode=0 Dec 05 07:20:02 crc kubenswrapper[4997]: I1205 07:20:02.069802 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-k8cjm" event={"ID":"78c3f4c1-158a-4d14-96c2-f9b6a663fd45","Type":"ContainerDied","Data":"8a63d40925c22c61026b0de2d63c2a59394ad5056824d09c3f51cc7b991b8313"} Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.813733 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.865833 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-etc-machine-id\") pod \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.865951 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-config-data\") pod \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.866026 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "78c3f4c1-158a-4d14-96c2-f9b6a663fd45" (UID: "78c3f4c1-158a-4d14-96c2-f9b6a663fd45"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.866061 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m75ld\" (UniqueName: \"kubernetes.io/projected/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-kube-api-access-m75ld\") pod \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.866185 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-scripts\") pod \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.866261 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-db-sync-config-data\") pod \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.866387 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-combined-ca-bundle\") pod \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\" (UID: \"78c3f4c1-158a-4d14-96c2-f9b6a663fd45\") " Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.867475 4997 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.875442 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-kube-api-access-m75ld" (OuterVolumeSpecName: "kube-api-access-m75ld") pod "78c3f4c1-158a-4d14-96c2-f9b6a663fd45" (UID: "78c3f4c1-158a-4d14-96c2-f9b6a663fd45"). InnerVolumeSpecName "kube-api-access-m75ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.876506 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-scripts" (OuterVolumeSpecName: "scripts") pod "78c3f4c1-158a-4d14-96c2-f9b6a663fd45" (UID: "78c3f4c1-158a-4d14-96c2-f9b6a663fd45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.876984 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "78c3f4c1-158a-4d14-96c2-f9b6a663fd45" (UID: "78c3f4c1-158a-4d14-96c2-f9b6a663fd45"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.899906 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78c3f4c1-158a-4d14-96c2-f9b6a663fd45" (UID: "78c3f4c1-158a-4d14-96c2-f9b6a663fd45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.938269 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-config-data" (OuterVolumeSpecName: "config-data") pod "78c3f4c1-158a-4d14-96c2-f9b6a663fd45" (UID: "78c3f4c1-158a-4d14-96c2-f9b6a663fd45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.969191 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.969243 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m75ld\" (UniqueName: \"kubernetes.io/projected/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-kube-api-access-m75ld\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.969262 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.969277 4997 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:03 crc kubenswrapper[4997]: I1205 07:20:03.969293 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c3f4c1-158a-4d14-96c2-f9b6a663fd45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.097240 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-k8cjm" event={"ID":"78c3f4c1-158a-4d14-96c2-f9b6a663fd45","Type":"ContainerDied","Data":"dcbe5a58304eddda0bd0d21df8f7798731d65a144530c50d8acddfc444a5795f"} Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.097498 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcbe5a58304eddda0bd0d21df8f7798731d65a144530c50d8acddfc444a5795f" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.097576 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-k8cjm" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.590811 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 07:20:04 crc kubenswrapper[4997]: E1205 07:20:04.591313 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c3f4c1-158a-4d14-96c2-f9b6a663fd45" containerName="cinder-db-sync" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.591332 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c3f4c1-158a-4d14-96c2-f9b6a663fd45" containerName="cinder-db-sync" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.591508 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="78c3f4c1-158a-4d14-96c2-f9b6a663fd45" containerName="cinder-db-sync" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.592581 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.596738 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.596918 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.597077 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-4jgwm" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.598983 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.608178 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.687264 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-jz4qt"] Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.692198 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.692398 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc6v8\" (UniqueName: \"kubernetes.io/projected/611d40e1-86aa-40d8-b140-7b68e2b68da4-kube-api-access-fc6v8\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.692458 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.692519 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.692559 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-config-data\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.692590 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-scripts\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.692655 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/611d40e1-86aa-40d8-b140-7b68e2b68da4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.687603 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" podUID="81bdd956-a535-4f79-9188-528dc91b0a1d" containerName="dnsmasq-dns" containerID="cri-o://d8d919c1f1f3e3dcc58ccb6121268482f2e3f8f5f2283c067b9e44b58bff4773" gracePeriod=10 Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.747931 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-62wl6"] Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.749923 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.785706 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-62wl6"] Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.795045 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgfqj\" (UniqueName: \"kubernetes.io/projected/e4b49169-e268-4a11-81fc-fe32226a01fd-kube-api-access-bgfqj\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.795115 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/611d40e1-86aa-40d8-b140-7b68e2b68da4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.795189 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-dns-swift-storage-0\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.795216 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-dns-svc\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.795255 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-ovsdbserver-nb\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.795308 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc6v8\" (UniqueName: \"kubernetes.io/projected/611d40e1-86aa-40d8-b140-7b68e2b68da4-kube-api-access-fc6v8\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.795333 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.795364 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-config\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.795390 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-ovsdbserver-sb\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.795814 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/611d40e1-86aa-40d8-b140-7b68e2b68da4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.797411 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.797470 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-config-data\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.797514 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-scripts\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.813824 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-scripts\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.814267 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-config-data\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.813931 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.821766 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.830950 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc6v8\" (UniqueName: \"kubernetes.io/projected/611d40e1-86aa-40d8-b140-7b68e2b68da4-kube-api-access-fc6v8\") pod \"cinder-scheduler-0\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.900058 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgfqj\" (UniqueName: \"kubernetes.io/projected/e4b49169-e268-4a11-81fc-fe32226a01fd-kube-api-access-bgfqj\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.900203 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-dns-swift-storage-0\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.900240 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-dns-svc\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.900276 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-ovsdbserver-nb\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.900347 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-config\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.900387 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-ovsdbserver-sb\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.901843 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-ovsdbserver-sb\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.901852 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-dns-svc\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.903194 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-dns-swift-storage-0\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.905850 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-config\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.906690 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-ovsdbserver-nb\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.918270 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.930533 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.934462 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.946259 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgfqj\" (UniqueName: \"kubernetes.io/projected/e4b49169-e268-4a11-81fc-fe32226a01fd-kube-api-access-bgfqj\") pod \"dnsmasq-dns-77d8c9c7-62wl6\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.953349 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 07:20:04 crc kubenswrapper[4997]: I1205 07:20:04.983495 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.010725 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-scripts\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.011033 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-config-data\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.011062 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.011156 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/85f5ed73-561d-4876-a44c-a393e820f71c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.011210 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6zs6\" (UniqueName: \"kubernetes.io/projected/85f5ed73-561d-4876-a44c-a393e820f71c-kube-api-access-v6zs6\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.011387 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-config-data-custom\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.011501 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85f5ed73-561d-4876-a44c-a393e820f71c-logs\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.094359 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.113262 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-config-data\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.113308 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.113355 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/85f5ed73-561d-4876-a44c-a393e820f71c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.113386 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6zs6\" (UniqueName: \"kubernetes.io/projected/85f5ed73-561d-4876-a44c-a393e820f71c-kube-api-access-v6zs6\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.113438 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-config-data-custom\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.113484 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85f5ed73-561d-4876-a44c-a393e820f71c-logs\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.113538 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-scripts\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.116289 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/85f5ed73-561d-4876-a44c-a393e820f71c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.121911 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-scripts\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.122235 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85f5ed73-561d-4876-a44c-a393e820f71c-logs\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.126803 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-config-data\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.129775 4997 generic.go:334] "Generic (PLEG): container finished" podID="81bdd956-a535-4f79-9188-528dc91b0a1d" containerID="d8d919c1f1f3e3dcc58ccb6121268482f2e3f8f5f2283c067b9e44b58bff4773" exitCode=0 Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.129847 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" event={"ID":"81bdd956-a535-4f79-9188-528dc91b0a1d","Type":"ContainerDied","Data":"d8d919c1f1f3e3dcc58ccb6121268482f2e3f8f5f2283c067b9e44b58bff4773"} Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.137486 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.139519 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-config-data-custom\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.172011 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6zs6\" (UniqueName: \"kubernetes.io/projected/85f5ed73-561d-4876-a44c-a393e820f71c-kube-api-access-v6zs6\") pod \"cinder-api-0\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " pod="openstack/cinder-api-0" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.251475 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" podUID="81bdd956-a535-4f79-9188-528dc91b0a1d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.157:5353: connect: connection refused" Dec 05 07:20:05 crc kubenswrapper[4997]: I1205 07:20:05.337864 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 07:20:07 crc kubenswrapper[4997]: I1205 07:20:07.573245 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:20:07 crc kubenswrapper[4997]: I1205 07:20:07.663484 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 05 07:20:07 crc kubenswrapper[4997]: I1205 07:20:07.752063 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:20:07 crc kubenswrapper[4997]: I1205 07:20:07.808001 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-config\") pod \"81bdd956-a535-4f79-9188-528dc91b0a1d\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " Dec 05 07:20:07 crc kubenswrapper[4997]: I1205 07:20:07.808173 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-dns-swift-storage-0\") pod \"81bdd956-a535-4f79-9188-528dc91b0a1d\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " Dec 05 07:20:07 crc kubenswrapper[4997]: I1205 07:20:07.808303 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-ovsdbserver-nb\") pod \"81bdd956-a535-4f79-9188-528dc91b0a1d\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " Dec 05 07:20:07 crc kubenswrapper[4997]: I1205 07:20:07.808367 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5lm9\" (UniqueName: \"kubernetes.io/projected/81bdd956-a535-4f79-9188-528dc91b0a1d-kube-api-access-z5lm9\") pod \"81bdd956-a535-4f79-9188-528dc91b0a1d\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " Dec 05 07:20:07 crc kubenswrapper[4997]: I1205 07:20:07.808463 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-ovsdbserver-sb\") pod \"81bdd956-a535-4f79-9188-528dc91b0a1d\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " Dec 05 07:20:07 crc kubenswrapper[4997]: I1205 07:20:07.808513 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-dns-svc\") pod \"81bdd956-a535-4f79-9188-528dc91b0a1d\" (UID: \"81bdd956-a535-4f79-9188-528dc91b0a1d\") " Dec 05 07:20:07 crc kubenswrapper[4997]: I1205 07:20:07.829409 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:20:07 crc kubenswrapper[4997]: I1205 07:20:07.849553 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81bdd956-a535-4f79-9188-528dc91b0a1d-kube-api-access-z5lm9" (OuterVolumeSpecName: "kube-api-access-z5lm9") pod "81bdd956-a535-4f79-9188-528dc91b0a1d" (UID: "81bdd956-a535-4f79-9188-528dc91b0a1d"). InnerVolumeSpecName "kube-api-access-z5lm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:07 crc kubenswrapper[4997]: I1205 07:20:07.934450 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5lm9\" (UniqueName: \"kubernetes.io/projected/81bdd956-a535-4f79-9188-528dc91b0a1d-kube-api-access-z5lm9\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:07 crc kubenswrapper[4997]: I1205 07:20:07.976953 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "81bdd956-a535-4f79-9188-528dc91b0a1d" (UID: "81bdd956-a535-4f79-9188-528dc91b0a1d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:20:07 crc kubenswrapper[4997]: I1205 07:20:07.980221 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "81bdd956-a535-4f79-9188-528dc91b0a1d" (UID: "81bdd956-a535-4f79-9188-528dc91b0a1d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.036521 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.036554 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.042541 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "81bdd956-a535-4f79-9188-528dc91b0a1d" (UID: "81bdd956-a535-4f79-9188-528dc91b0a1d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.047171 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-config" (OuterVolumeSpecName: "config") pod "81bdd956-a535-4f79-9188-528dc91b0a1d" (UID: "81bdd956-a535-4f79-9188-528dc91b0a1d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.049846 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "81bdd956-a535-4f79-9188-528dc91b0a1d" (UID: "81bdd956-a535-4f79-9188-528dc91b0a1d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.138253 4997 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.138291 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.138303 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81bdd956-a535-4f79-9188-528dc91b0a1d-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.169859 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe5443e-726b-4976-a7a3-c226bd3ec481","Type":"ContainerStarted","Data":"7005e478ec1d5ab8b440588b9eca063930b2e1ddb1c02e3c5577f50224c89335"} Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.169986 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerName="ceilometer-central-agent" containerID="cri-o://3f9c3f41c902d84f5da3f081f73987894ad168e0e10bf5cdaeb82c27e1353e2a" gracePeriod=30 Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.170356 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerName="proxy-httpd" containerID="cri-o://7005e478ec1d5ab8b440588b9eca063930b2e1ddb1c02e3c5577f50224c89335" gracePeriod=30 Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.170448 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerName="sg-core" containerID="cri-o://5a0093fba43766bdbd2468af37a378fd8ab58701871d81188044c42c82c86057" gracePeriod=30 Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.170520 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerName="ceilometer-notification-agent" containerID="cri-o://6b46eaec9f50401eed6987b13e75bc96a2424eaf209614937ac23a0c2cc3c2a0" gracePeriod=30 Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.170525 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.176797 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" event={"ID":"81bdd956-a535-4f79-9188-528dc91b0a1d","Type":"ContainerDied","Data":"bb0ae538937ce6a38824254600a433da51ec4c16b3f0b070d618973838f88cb0"} Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.177544 4997 scope.go:117] "RemoveContainer" containerID="d8d919c1f1f3e3dcc58ccb6121268482f2e3f8f5f2283c067b9e44b58bff4773" Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.179889 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687dbb56f-jz4qt" Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.227025 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.236687 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.762226546 podStartE2EDuration="53.236652608s" podCreationTimestamp="2025-12-05 07:19:15 +0000 UTC" firstStartedPulling="2025-12-05 07:19:17.121257963 +0000 UTC m=+1457.650165224" lastFinishedPulling="2025-12-05 07:20:07.595684035 +0000 UTC m=+1508.124591286" observedRunningTime="2025-12-05 07:20:08.199951486 +0000 UTC m=+1508.728858747" watchObservedRunningTime="2025-12-05 07:20:08.236652608 +0000 UTC m=+1508.765559869" Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.257211 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-jz4qt"] Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.265111 4997 scope.go:117] "RemoveContainer" containerID="51ec5e9c7d9204cde67cc653b0c3bb7be47f74896a9ed1f9d466229c0a281a37" Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.272871 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-jz4qt"] Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.320946 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-62wl6"] Dec 05 07:20:08 crc kubenswrapper[4997]: W1205 07:20:08.354227 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4b49169_e268_4a11_81fc_fe32226a01fd.slice/crio-6e425e4aeb21e62c46892cff8bf494c237226acd32bfd5d16f8667eb894e5924 WatchSource:0}: Error finding container 6e425e4aeb21e62c46892cff8bf494c237226acd32bfd5d16f8667eb894e5924: Status 404 returned error can't find the container with id 6e425e4aeb21e62c46892cff8bf494c237226acd32bfd5d16f8667eb894e5924 Dec 05 07:20:08 crc kubenswrapper[4997]: I1205 07:20:08.420165 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 05 07:20:09 crc kubenswrapper[4997]: I1205 07:20:09.197491 4997 generic.go:334] "Generic (PLEG): container finished" podID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerID="5a0093fba43766bdbd2468af37a378fd8ab58701871d81188044c42c82c86057" exitCode=2 Dec 05 07:20:09 crc kubenswrapper[4997]: I1205 07:20:09.197888 4997 generic.go:334] "Generic (PLEG): container finished" podID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerID="3f9c3f41c902d84f5da3f081f73987894ad168e0e10bf5cdaeb82c27e1353e2a" exitCode=0 Dec 05 07:20:09 crc kubenswrapper[4997]: I1205 07:20:09.197939 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe5443e-726b-4976-a7a3-c226bd3ec481","Type":"ContainerDied","Data":"5a0093fba43766bdbd2468af37a378fd8ab58701871d81188044c42c82c86057"} Dec 05 07:20:09 crc kubenswrapper[4997]: I1205 07:20:09.197973 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe5443e-726b-4976-a7a3-c226bd3ec481","Type":"ContainerDied","Data":"3f9c3f41c902d84f5da3f081f73987894ad168e0e10bf5cdaeb82c27e1353e2a"} Dec 05 07:20:09 crc kubenswrapper[4997]: I1205 07:20:09.206096 4997 generic.go:334] "Generic (PLEG): container finished" podID="e4b49169-e268-4a11-81fc-fe32226a01fd" containerID="eb4dacc8abb90bf9759f70da561b93049b2c7561f701c584106c80b19aabf364" exitCode=0 Dec 05 07:20:09 crc kubenswrapper[4997]: I1205 07:20:09.206212 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" event={"ID":"e4b49169-e268-4a11-81fc-fe32226a01fd","Type":"ContainerDied","Data":"eb4dacc8abb90bf9759f70da561b93049b2c7561f701c584106c80b19aabf364"} Dec 05 07:20:09 crc kubenswrapper[4997]: I1205 07:20:09.206260 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" event={"ID":"e4b49169-e268-4a11-81fc-fe32226a01fd","Type":"ContainerStarted","Data":"6e425e4aeb21e62c46892cff8bf494c237226acd32bfd5d16f8667eb894e5924"} Dec 05 07:20:09 crc kubenswrapper[4997]: I1205 07:20:09.222264 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"611d40e1-86aa-40d8-b140-7b68e2b68da4","Type":"ContainerStarted","Data":"e0edf401c8a131ee5a66820bc685ba9229dec29f0913ae1f19ba65073988f5bf"} Dec 05 07:20:09 crc kubenswrapper[4997]: I1205 07:20:09.223404 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"85f5ed73-561d-4876-a44c-a393e820f71c","Type":"ContainerStarted","Data":"7345a8dacac46a5bf0f92440d2e59e3b62f96f451dec6dda324a613d4ab4c6fc"} Dec 05 07:20:09 crc kubenswrapper[4997]: I1205 07:20:09.770652 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81bdd956-a535-4f79-9188-528dc91b0a1d" path="/var/lib/kubelet/pods/81bdd956-a535-4f79-9188-528dc91b0a1d/volumes" Dec 05 07:20:10 crc kubenswrapper[4997]: I1205 07:20:10.252892 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" event={"ID":"e4b49169-e268-4a11-81fc-fe32226a01fd","Type":"ContainerStarted","Data":"bafe6e887ce62394ebc58d484ef2d5496cf300ba94f50eae3ce35d5603d9ea1c"} Dec 05 07:20:10 crc kubenswrapper[4997]: I1205 07:20:10.253384 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:10 crc kubenswrapper[4997]: I1205 07:20:10.256075 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"611d40e1-86aa-40d8-b140-7b68e2b68da4","Type":"ContainerStarted","Data":"502a1a0aed4cb9b9ffbc88d0121b256a67322b6cccc683e819dadfd7597beef9"} Dec 05 07:20:10 crc kubenswrapper[4997]: I1205 07:20:10.258955 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"85f5ed73-561d-4876-a44c-a393e820f71c","Type":"ContainerStarted","Data":"c1feee6a9ec30b94a61c86ebbb440bcaaaef2f31275039429a3176d29f79d1dd"} Dec 05 07:20:10 crc kubenswrapper[4997]: I1205 07:20:10.259133 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="85f5ed73-561d-4876-a44c-a393e820f71c" containerName="cinder-api-log" containerID="cri-o://c1feee6a9ec30b94a61c86ebbb440bcaaaef2f31275039429a3176d29f79d1dd" gracePeriod=30 Dec 05 07:20:10 crc kubenswrapper[4997]: I1205 07:20:10.259420 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 05 07:20:10 crc kubenswrapper[4997]: I1205 07:20:10.259466 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="85f5ed73-561d-4876-a44c-a393e820f71c" containerName="cinder-api" containerID="cri-o://9c1e37f5166ff86419dd377730b522eee072560d2d117f47e2d55de9448b1218" gracePeriod=30 Dec 05 07:20:10 crc kubenswrapper[4997]: I1205 07:20:10.279873 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" podStartSLOduration=6.27981087 podStartE2EDuration="6.27981087s" podCreationTimestamp="2025-12-05 07:20:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:20:10.268872384 +0000 UTC m=+1510.797779645" watchObservedRunningTime="2025-12-05 07:20:10.27981087 +0000 UTC m=+1510.808718131" Dec 05 07:20:10 crc kubenswrapper[4997]: I1205 07:20:10.303315 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.303288424 podStartE2EDuration="6.303288424s" podCreationTimestamp="2025-12-05 07:20:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:20:10.28980159 +0000 UTC m=+1510.818708881" watchObservedRunningTime="2025-12-05 07:20:10.303288424 +0000 UTC m=+1510.832195685" Dec 05 07:20:11 crc kubenswrapper[4997]: I1205 07:20:11.254588 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:20:11 crc kubenswrapper[4997]: I1205 07:20:11.259520 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:20:11 crc kubenswrapper[4997]: I1205 07:20:11.285030 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"611d40e1-86aa-40d8-b140-7b68e2b68da4","Type":"ContainerStarted","Data":"47ea7d4c403ab319dbb86e0b8a219d69130f2cd84976829f635588673504c525"} Dec 05 07:20:11 crc kubenswrapper[4997]: I1205 07:20:11.292710 4997 generic.go:334] "Generic (PLEG): container finished" podID="85f5ed73-561d-4876-a44c-a393e820f71c" containerID="c1feee6a9ec30b94a61c86ebbb440bcaaaef2f31275039429a3176d29f79d1dd" exitCode=143 Dec 05 07:20:11 crc kubenswrapper[4997]: I1205 07:20:11.292851 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"85f5ed73-561d-4876-a44c-a393e820f71c","Type":"ContainerStarted","Data":"9c1e37f5166ff86419dd377730b522eee072560d2d117f47e2d55de9448b1218"} Dec 05 07:20:11 crc kubenswrapper[4997]: I1205 07:20:11.292929 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"85f5ed73-561d-4876-a44c-a393e820f71c","Type":"ContainerDied","Data":"c1feee6a9ec30b94a61c86ebbb440bcaaaef2f31275039429a3176d29f79d1dd"} Dec 05 07:20:11 crc kubenswrapper[4997]: I1205 07:20:11.358452 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.5256794639999995 podStartE2EDuration="7.358426472s" podCreationTimestamp="2025-12-05 07:20:04 +0000 UTC" firstStartedPulling="2025-12-05 07:20:08.226859643 +0000 UTC m=+1508.755766904" lastFinishedPulling="2025-12-05 07:20:09.059606651 +0000 UTC m=+1509.588513912" observedRunningTime="2025-12-05 07:20:11.358416342 +0000 UTC m=+1511.887323613" watchObservedRunningTime="2025-12-05 07:20:11.358426472 +0000 UTC m=+1511.887333723" Dec 05 07:20:11 crc kubenswrapper[4997]: I1205 07:20:11.383207 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-57bc457fc6-28p4d"] Dec 05 07:20:11 crc kubenswrapper[4997]: I1205 07:20:11.383519 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-57bc457fc6-28p4d" podUID="6aecdebd-cbaf-4af2-acf4-06f9321b9a37" containerName="barbican-api-log" containerID="cri-o://cd2440f9aa5e210d229a0413c490c0d93af695ff6e68a1e329a3902622ad4d5f" gracePeriod=30 Dec 05 07:20:11 crc kubenswrapper[4997]: I1205 07:20:11.383694 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-57bc457fc6-28p4d" podUID="6aecdebd-cbaf-4af2-acf4-06f9321b9a37" containerName="barbican-api" containerID="cri-o://807fb7f45b92c088f92ec791ea2cb20347da760ff331b4691df3d6cd5e5e7ce6" gracePeriod=30 Dec 05 07:20:12 crc kubenswrapper[4997]: I1205 07:20:12.306008 4997 generic.go:334] "Generic (PLEG): container finished" podID="6aecdebd-cbaf-4af2-acf4-06f9321b9a37" containerID="cd2440f9aa5e210d229a0413c490c0d93af695ff6e68a1e329a3902622ad4d5f" exitCode=143 Dec 05 07:20:12 crc kubenswrapper[4997]: I1205 07:20:12.306072 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57bc457fc6-28p4d" event={"ID":"6aecdebd-cbaf-4af2-acf4-06f9321b9a37","Type":"ContainerDied","Data":"cd2440f9aa5e210d229a0413c490c0d93af695ff6e68a1e329a3902622ad4d5f"} Dec 05 07:20:12 crc kubenswrapper[4997]: I1205 07:20:12.415350 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:20:13 crc kubenswrapper[4997]: I1205 07:20:13.325228 4997 generic.go:334] "Generic (PLEG): container finished" podID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerID="6b46eaec9f50401eed6987b13e75bc96a2424eaf209614937ac23a0c2cc3c2a0" exitCode=0 Dec 05 07:20:13 crc kubenswrapper[4997]: I1205 07:20:13.325312 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe5443e-726b-4976-a7a3-c226bd3ec481","Type":"ContainerDied","Data":"6b46eaec9f50401eed6987b13e75bc96a2424eaf209614937ac23a0c2cc3c2a0"} Dec 05 07:20:14 crc kubenswrapper[4997]: I1205 07:20:14.987461 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.005886 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.101813 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.113449 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-logs\") pod \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.113514 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-config-data\") pod \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.113593 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-config-data-custom\") pod \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.113874 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-combined-ca-bundle\") pod \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.113927 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnghc\" (UniqueName: \"kubernetes.io/projected/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-kube-api-access-bnghc\") pod \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\" (UID: \"6aecdebd-cbaf-4af2-acf4-06f9321b9a37\") " Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.114025 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-logs" (OuterVolumeSpecName: "logs") pod "6aecdebd-cbaf-4af2-acf4-06f9321b9a37" (UID: "6aecdebd-cbaf-4af2-acf4-06f9321b9a37"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.116789 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.120516 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6aecdebd-cbaf-4af2-acf4-06f9321b9a37" (UID: "6aecdebd-cbaf-4af2-acf4-06f9321b9a37"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.122185 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-kube-api-access-bnghc" (OuterVolumeSpecName: "kube-api-access-bnghc") pod "6aecdebd-cbaf-4af2-acf4-06f9321b9a37" (UID: "6aecdebd-cbaf-4af2-acf4-06f9321b9a37"). InnerVolumeSpecName "kube-api-access-bnghc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.204769 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-vbzgc"] Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.205103 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" podUID="ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e" containerName="dnsmasq-dns" containerID="cri-o://8a9193193aa72784b1caede2e5cb4ba050873269ad92e24cf4a5a1853ef29f74" gracePeriod=10 Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.220902 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnghc\" (UniqueName: \"kubernetes.io/projected/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-kube-api-access-bnghc\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.220936 4997 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.221672 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6aecdebd-cbaf-4af2-acf4-06f9321b9a37" (UID: "6aecdebd-cbaf-4af2-acf4-06f9321b9a37"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.228274 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-config-data" (OuterVolumeSpecName: "config-data") pod "6aecdebd-cbaf-4af2-acf4-06f9321b9a37" (UID: "6aecdebd-cbaf-4af2-acf4-06f9321b9a37"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.324257 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.324299 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aecdebd-cbaf-4af2-acf4-06f9321b9a37-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.351537 4997 generic.go:334] "Generic (PLEG): container finished" podID="ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e" containerID="8a9193193aa72784b1caede2e5cb4ba050873269ad92e24cf4a5a1853ef29f74" exitCode=0 Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.351577 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" event={"ID":"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e","Type":"ContainerDied","Data":"8a9193193aa72784b1caede2e5cb4ba050873269ad92e24cf4a5a1853ef29f74"} Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.355038 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.356089 4997 generic.go:334] "Generic (PLEG): container finished" podID="6aecdebd-cbaf-4af2-acf4-06f9321b9a37" containerID="807fb7f45b92c088f92ec791ea2cb20347da760ff331b4691df3d6cd5e5e7ce6" exitCode=0 Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.356260 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57bc457fc6-28p4d" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.357308 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57bc457fc6-28p4d" event={"ID":"6aecdebd-cbaf-4af2-acf4-06f9321b9a37","Type":"ContainerDied","Data":"807fb7f45b92c088f92ec791ea2cb20347da760ff331b4691df3d6cd5e5e7ce6"} Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.357371 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57bc457fc6-28p4d" event={"ID":"6aecdebd-cbaf-4af2-acf4-06f9321b9a37","Type":"ContainerDied","Data":"22f08e4078c712ccb3dc0aff6dc4f9645b5d77e7283268b483fd1a74fe7c74e1"} Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.357416 4997 scope.go:117] "RemoveContainer" containerID="807fb7f45b92c088f92ec791ea2cb20347da760ff331b4691df3d6cd5e5e7ce6" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.410901 4997 scope.go:117] "RemoveContainer" containerID="cd2440f9aa5e210d229a0413c490c0d93af695ff6e68a1e329a3902622ad4d5f" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.452782 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.470779 4997 scope.go:117] "RemoveContainer" containerID="807fb7f45b92c088f92ec791ea2cb20347da760ff331b4691df3d6cd5e5e7ce6" Dec 05 07:20:15 crc kubenswrapper[4997]: E1205 07:20:15.481423 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"807fb7f45b92c088f92ec791ea2cb20347da760ff331b4691df3d6cd5e5e7ce6\": container with ID starting with 807fb7f45b92c088f92ec791ea2cb20347da760ff331b4691df3d6cd5e5e7ce6 not found: ID does not exist" containerID="807fb7f45b92c088f92ec791ea2cb20347da760ff331b4691df3d6cd5e5e7ce6" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.481486 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"807fb7f45b92c088f92ec791ea2cb20347da760ff331b4691df3d6cd5e5e7ce6"} err="failed to get container status \"807fb7f45b92c088f92ec791ea2cb20347da760ff331b4691df3d6cd5e5e7ce6\": rpc error: code = NotFound desc = could not find container \"807fb7f45b92c088f92ec791ea2cb20347da760ff331b4691df3d6cd5e5e7ce6\": container with ID starting with 807fb7f45b92c088f92ec791ea2cb20347da760ff331b4691df3d6cd5e5e7ce6 not found: ID does not exist" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.481525 4997 scope.go:117] "RemoveContainer" containerID="cd2440f9aa5e210d229a0413c490c0d93af695ff6e68a1e329a3902622ad4d5f" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.481669 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-57bc457fc6-28p4d"] Dec 05 07:20:15 crc kubenswrapper[4997]: E1205 07:20:15.487833 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd2440f9aa5e210d229a0413c490c0d93af695ff6e68a1e329a3902622ad4d5f\": container with ID starting with cd2440f9aa5e210d229a0413c490c0d93af695ff6e68a1e329a3902622ad4d5f not found: ID does not exist" containerID="cd2440f9aa5e210d229a0413c490c0d93af695ff6e68a1e329a3902622ad4d5f" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.487909 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd2440f9aa5e210d229a0413c490c0d93af695ff6e68a1e329a3902622ad4d5f"} err="failed to get container status \"cd2440f9aa5e210d229a0413c490c0d93af695ff6e68a1e329a3902622ad4d5f\": rpc error: code = NotFound desc = could not find container \"cd2440f9aa5e210d229a0413c490c0d93af695ff6e68a1e329a3902622ad4d5f\": container with ID starting with cd2440f9aa5e210d229a0413c490c0d93af695ff6e68a1e329a3902622ad4d5f not found: ID does not exist" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.497718 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-57bc457fc6-28p4d"] Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.762135 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aecdebd-cbaf-4af2-acf4-06f9321b9a37" path="/var/lib/kubelet/pods/6aecdebd-cbaf-4af2-acf4-06f9321b9a37/volumes" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.782407 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.858025 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-dns-swift-storage-0\") pod \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.858095 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-dns-svc\") pod \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.858190 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xwql\" (UniqueName: \"kubernetes.io/projected/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-kube-api-access-6xwql\") pod \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.858209 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-config\") pod \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.858317 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-ovsdbserver-sb\") pod \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.858368 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-ovsdbserver-nb\") pod \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\" (UID: \"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e\") " Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.863975 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-kube-api-access-6xwql" (OuterVolumeSpecName: "kube-api-access-6xwql") pod "ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e" (UID: "ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e"). InnerVolumeSpecName "kube-api-access-6xwql". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.912782 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e" (UID: "ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.913338 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e" (UID: "ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.914254 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e" (UID: "ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.917718 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-config" (OuterVolumeSpecName: "config") pod "ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e" (UID: "ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.929188 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e" (UID: "ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.956738 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.961206 4997 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.961460 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.961583 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xwql\" (UniqueName: \"kubernetes.io/projected/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-kube-api-access-6xwql\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.961753 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.961878 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:15 crc kubenswrapper[4997]: I1205 07:20:15.962029 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:16 crc kubenswrapper[4997]: I1205 07:20:16.027668 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-64fc5c4f5b-mnq4n"] Dec 05 07:20:16 crc kubenswrapper[4997]: I1205 07:20:16.028393 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-64fc5c4f5b-mnq4n" podUID="edf49641-b44d-4dba-af43-9a9da06eb55d" containerName="neutron-api" containerID="cri-o://7a4dbab571ceebe87a020f8c86c1c678db709ecca7bb391c41b9909b0ab0648f" gracePeriod=30 Dec 05 07:20:16 crc kubenswrapper[4997]: I1205 07:20:16.028958 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-64fc5c4f5b-mnq4n" podUID="edf49641-b44d-4dba-af43-9a9da06eb55d" containerName="neutron-httpd" containerID="cri-o://d179214c82ee33c9b885acffd78f36623afd0e09d1efcc0df4a6be02aeb8e158" gracePeriod=30 Dec 05 07:20:16 crc kubenswrapper[4997]: I1205 07:20:16.368229 4997 generic.go:334] "Generic (PLEG): container finished" podID="edf49641-b44d-4dba-af43-9a9da06eb55d" containerID="d179214c82ee33c9b885acffd78f36623afd0e09d1efcc0df4a6be02aeb8e158" exitCode=0 Dec 05 07:20:16 crc kubenswrapper[4997]: I1205 07:20:16.368353 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64fc5c4f5b-mnq4n" event={"ID":"edf49641-b44d-4dba-af43-9a9da06eb55d","Type":"ContainerDied","Data":"d179214c82ee33c9b885acffd78f36623afd0e09d1efcc0df4a6be02aeb8e158"} Dec 05 07:20:16 crc kubenswrapper[4997]: I1205 07:20:16.372756 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" event={"ID":"ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e","Type":"ContainerDied","Data":"f9aa9b19a4539ead367b8e31efe2226124a659adbe7e0fe5dccde26730fcda2a"} Dec 05 07:20:16 crc kubenswrapper[4997]: I1205 07:20:16.372830 4997 scope.go:117] "RemoveContainer" containerID="8a9193193aa72784b1caede2e5cb4ba050873269ad92e24cf4a5a1853ef29f74" Dec 05 07:20:16 crc kubenswrapper[4997]: I1205 07:20:16.372834 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77f55878d5-vbzgc" Dec 05 07:20:16 crc kubenswrapper[4997]: I1205 07:20:16.374968 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="611d40e1-86aa-40d8-b140-7b68e2b68da4" containerName="cinder-scheduler" containerID="cri-o://502a1a0aed4cb9b9ffbc88d0121b256a67322b6cccc683e819dadfd7597beef9" gracePeriod=30 Dec 05 07:20:16 crc kubenswrapper[4997]: I1205 07:20:16.375022 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="611d40e1-86aa-40d8-b140-7b68e2b68da4" containerName="probe" containerID="cri-o://47ea7d4c403ab319dbb86e0b8a219d69130f2cd84976829f635588673504c525" gracePeriod=30 Dec 05 07:20:16 crc kubenswrapper[4997]: I1205 07:20:16.419147 4997 scope.go:117] "RemoveContainer" containerID="8738a1e9499085a2414139ca014a0ee0f5ad8ed806324ba328c6a1fb7006b6b6" Dec 05 07:20:16 crc kubenswrapper[4997]: I1205 07:20:16.426842 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-vbzgc"] Dec 05 07:20:16 crc kubenswrapper[4997]: I1205 07:20:16.435493 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-vbzgc"] Dec 05 07:20:17 crc kubenswrapper[4997]: I1205 07:20:17.762445 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e" path="/var/lib/kubelet/pods/ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e/volumes" Dec 05 07:20:17 crc kubenswrapper[4997]: I1205 07:20:17.798214 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 05 07:20:18 crc kubenswrapper[4997]: I1205 07:20:18.401045 4997 generic.go:334] "Generic (PLEG): container finished" podID="611d40e1-86aa-40d8-b140-7b68e2b68da4" containerID="47ea7d4c403ab319dbb86e0b8a219d69130f2cd84976829f635588673504c525" exitCode=0 Dec 05 07:20:18 crc kubenswrapper[4997]: I1205 07:20:18.401648 4997 generic.go:334] "Generic (PLEG): container finished" podID="611d40e1-86aa-40d8-b140-7b68e2b68da4" containerID="502a1a0aed4cb9b9ffbc88d0121b256a67322b6cccc683e819dadfd7597beef9" exitCode=0 Dec 05 07:20:18 crc kubenswrapper[4997]: I1205 07:20:18.401702 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"611d40e1-86aa-40d8-b140-7b68e2b68da4","Type":"ContainerDied","Data":"47ea7d4c403ab319dbb86e0b8a219d69130f2cd84976829f635588673504c525"} Dec 05 07:20:18 crc kubenswrapper[4997]: I1205 07:20:18.401743 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"611d40e1-86aa-40d8-b140-7b68e2b68da4","Type":"ContainerDied","Data":"502a1a0aed4cb9b9ffbc88d0121b256a67322b6cccc683e819dadfd7597beef9"} Dec 05 07:20:18 crc kubenswrapper[4997]: I1205 07:20:18.993213 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.030992 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/611d40e1-86aa-40d8-b140-7b68e2b68da4-etc-machine-id\") pod \"611d40e1-86aa-40d8-b140-7b68e2b68da4\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.031435 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc6v8\" (UniqueName: \"kubernetes.io/projected/611d40e1-86aa-40d8-b140-7b68e2b68da4-kube-api-access-fc6v8\") pod \"611d40e1-86aa-40d8-b140-7b68e2b68da4\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.031479 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-config-data-custom\") pod \"611d40e1-86aa-40d8-b140-7b68e2b68da4\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.031545 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-scripts\") pod \"611d40e1-86aa-40d8-b140-7b68e2b68da4\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.031121 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/611d40e1-86aa-40d8-b140-7b68e2b68da4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "611d40e1-86aa-40d8-b140-7b68e2b68da4" (UID: "611d40e1-86aa-40d8-b140-7b68e2b68da4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.031686 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-combined-ca-bundle\") pod \"611d40e1-86aa-40d8-b140-7b68e2b68da4\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.031740 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-config-data\") pod \"611d40e1-86aa-40d8-b140-7b68e2b68da4\" (UID: \"611d40e1-86aa-40d8-b140-7b68e2b68da4\") " Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.032245 4997 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/611d40e1-86aa-40d8-b140-7b68e2b68da4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.040404 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "611d40e1-86aa-40d8-b140-7b68e2b68da4" (UID: "611d40e1-86aa-40d8-b140-7b68e2b68da4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.053341 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-scripts" (OuterVolumeSpecName: "scripts") pod "611d40e1-86aa-40d8-b140-7b68e2b68da4" (UID: "611d40e1-86aa-40d8-b140-7b68e2b68da4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.058110 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/611d40e1-86aa-40d8-b140-7b68e2b68da4-kube-api-access-fc6v8" (OuterVolumeSpecName: "kube-api-access-fc6v8") pod "611d40e1-86aa-40d8-b140-7b68e2b68da4" (UID: "611d40e1-86aa-40d8-b140-7b68e2b68da4"). InnerVolumeSpecName "kube-api-access-fc6v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.109432 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "611d40e1-86aa-40d8-b140-7b68e2b68da4" (UID: "611d40e1-86aa-40d8-b140-7b68e2b68da4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.135296 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc6v8\" (UniqueName: \"kubernetes.io/projected/611d40e1-86aa-40d8-b140-7b68e2b68da4-kube-api-access-fc6v8\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.135345 4997 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.135360 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.135376 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.173891 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-config-data" (OuterVolumeSpecName: "config-data") pod "611d40e1-86aa-40d8-b140-7b68e2b68da4" (UID: "611d40e1-86aa-40d8-b140-7b68e2b68da4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.220737 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.226971 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.237661 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/611d40e1-86aa-40d8-b140-7b68e2b68da4-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.422221 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.424217 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"611d40e1-86aa-40d8-b140-7b68e2b68da4","Type":"ContainerDied","Data":"e0edf401c8a131ee5a66820bc685ba9229dec29f0913ae1f19ba65073988f5bf"} Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.426983 4997 scope.go:117] "RemoveContainer" containerID="47ea7d4c403ab319dbb86e0b8a219d69130f2cd84976829f635588673504c525" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.462832 4997 scope.go:117] "RemoveContainer" containerID="502a1a0aed4cb9b9ffbc88d0121b256a67322b6cccc683e819dadfd7597beef9" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.492647 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.509314 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.529270 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 07:20:19 crc kubenswrapper[4997]: E1205 07:20:19.530032 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="611d40e1-86aa-40d8-b140-7b68e2b68da4" containerName="probe" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.530063 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="611d40e1-86aa-40d8-b140-7b68e2b68da4" containerName="probe" Dec 05 07:20:19 crc kubenswrapper[4997]: E1205 07:20:19.530080 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aecdebd-cbaf-4af2-acf4-06f9321b9a37" containerName="barbican-api-log" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.530093 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aecdebd-cbaf-4af2-acf4-06f9321b9a37" containerName="barbican-api-log" Dec 05 07:20:19 crc kubenswrapper[4997]: E1205 07:20:19.530121 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e" containerName="init" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.530129 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e" containerName="init" Dec 05 07:20:19 crc kubenswrapper[4997]: E1205 07:20:19.530145 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81bdd956-a535-4f79-9188-528dc91b0a1d" containerName="dnsmasq-dns" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.530153 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="81bdd956-a535-4f79-9188-528dc91b0a1d" containerName="dnsmasq-dns" Dec 05 07:20:19 crc kubenswrapper[4997]: E1205 07:20:19.530178 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aecdebd-cbaf-4af2-acf4-06f9321b9a37" containerName="barbican-api" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.530185 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aecdebd-cbaf-4af2-acf4-06f9321b9a37" containerName="barbican-api" Dec 05 07:20:19 crc kubenswrapper[4997]: E1205 07:20:19.530196 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e" containerName="dnsmasq-dns" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.530205 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e" containerName="dnsmasq-dns" Dec 05 07:20:19 crc kubenswrapper[4997]: E1205 07:20:19.530222 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="611d40e1-86aa-40d8-b140-7b68e2b68da4" containerName="cinder-scheduler" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.530232 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="611d40e1-86aa-40d8-b140-7b68e2b68da4" containerName="cinder-scheduler" Dec 05 07:20:19 crc kubenswrapper[4997]: E1205 07:20:19.530253 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81bdd956-a535-4f79-9188-528dc91b0a1d" containerName="init" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.530262 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="81bdd956-a535-4f79-9188-528dc91b0a1d" containerName="init" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.530510 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aecdebd-cbaf-4af2-acf4-06f9321b9a37" containerName="barbican-api-log" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.530531 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="81bdd956-a535-4f79-9188-528dc91b0a1d" containerName="dnsmasq-dns" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.530552 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aecdebd-cbaf-4af2-acf4-06f9321b9a37" containerName="barbican-api" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.530570 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccb6efaf-8a4c-47c5-8e0e-9b779fae9c2e" containerName="dnsmasq-dns" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.530582 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="611d40e1-86aa-40d8-b140-7b68e2b68da4" containerName="cinder-scheduler" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.530595 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="611d40e1-86aa-40d8-b140-7b68e2b68da4" containerName="probe" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.532192 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.536313 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.543572 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.645265 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-config-data\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.645551 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.645713 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-scripts\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.645794 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.645947 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe32fb42-61e6-4cd4-a75b-b684a4171224-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.646201 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vblvp\" (UniqueName: \"kubernetes.io/projected/fe32fb42-61e6-4cd4-a75b-b684a4171224-kube-api-access-vblvp\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.748065 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe32fb42-61e6-4cd4-a75b-b684a4171224-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.748150 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vblvp\" (UniqueName: \"kubernetes.io/projected/fe32fb42-61e6-4cd4-a75b-b684a4171224-kube-api-access-vblvp\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.748221 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-config-data\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.748266 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.748346 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-scripts\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.748369 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.748980 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe32fb42-61e6-4cd4-a75b-b684a4171224-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.753211 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-config-data\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.753341 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-scripts\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.754106 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.762680 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="611d40e1-86aa-40d8-b140-7b68e2b68da4" path="/var/lib/kubelet/pods/611d40e1-86aa-40d8-b140-7b68e2b68da4/volumes" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.764317 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.768764 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vblvp\" (UniqueName: \"kubernetes.io/projected/fe32fb42-61e6-4cd4-a75b-b684a4171224-kube-api-access-vblvp\") pod \"cinder-scheduler-0\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " pod="openstack/cinder-scheduler-0" Dec 05 07:20:19 crc kubenswrapper[4997]: I1205 07:20:19.854892 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 07:20:20 crc kubenswrapper[4997]: I1205 07:20:20.351261 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 07:20:20 crc kubenswrapper[4997]: I1205 07:20:20.442556 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fe32fb42-61e6-4cd4-a75b-b684a4171224","Type":"ContainerStarted","Data":"b35c9b6435ee745827f49e345e658de1e9600ff778a331dca340af2f70164c49"} Dec 05 07:20:22 crc kubenswrapper[4997]: I1205 07:20:22.476343 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fe32fb42-61e6-4cd4-a75b-b684a4171224","Type":"ContainerStarted","Data":"b762656e8963968dd62e28b778f1374e72ed4bfb67fbcee7edeacd3920071001"} Dec 05 07:20:23 crc kubenswrapper[4997]: I1205 07:20:23.505093 4997 generic.go:334] "Generic (PLEG): container finished" podID="edf49641-b44d-4dba-af43-9a9da06eb55d" containerID="7a4dbab571ceebe87a020f8c86c1c678db709ecca7bb391c41b9909b0ab0648f" exitCode=0 Dec 05 07:20:23 crc kubenswrapper[4997]: I1205 07:20:23.505172 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64fc5c4f5b-mnq4n" event={"ID":"edf49641-b44d-4dba-af43-9a9da06eb55d","Type":"ContainerDied","Data":"7a4dbab571ceebe87a020f8c86c1c678db709ecca7bb391c41b9909b0ab0648f"} Dec 05 07:20:23 crc kubenswrapper[4997]: I1205 07:20:23.509496 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fe32fb42-61e6-4cd4-a75b-b684a4171224","Type":"ContainerStarted","Data":"921875cb2f26d740baa583b8c4418587dfe774e4ea61c9ad78bb1723d653c892"} Dec 05 07:20:23 crc kubenswrapper[4997]: I1205 07:20:23.539750 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.5397142729999995 podStartE2EDuration="4.539714273s" podCreationTimestamp="2025-12-05 07:20:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:20:23.529836986 +0000 UTC m=+1524.058744267" watchObservedRunningTime="2025-12-05 07:20:23.539714273 +0000 UTC m=+1524.068621534" Dec 05 07:20:23 crc kubenswrapper[4997]: I1205 07:20:23.936459 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.015844 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-combined-ca-bundle\") pod \"edf49641-b44d-4dba-af43-9a9da06eb55d\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.016578 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-ovndb-tls-certs\") pod \"edf49641-b44d-4dba-af43-9a9da06eb55d\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.016678 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wk8jl\" (UniqueName: \"kubernetes.io/projected/edf49641-b44d-4dba-af43-9a9da06eb55d-kube-api-access-wk8jl\") pod \"edf49641-b44d-4dba-af43-9a9da06eb55d\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.016773 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-httpd-config\") pod \"edf49641-b44d-4dba-af43-9a9da06eb55d\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.016990 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-config\") pod \"edf49641-b44d-4dba-af43-9a9da06eb55d\" (UID: \"edf49641-b44d-4dba-af43-9a9da06eb55d\") " Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.022984 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "edf49641-b44d-4dba-af43-9a9da06eb55d" (UID: "edf49641-b44d-4dba-af43-9a9da06eb55d"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.023539 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edf49641-b44d-4dba-af43-9a9da06eb55d-kube-api-access-wk8jl" (OuterVolumeSpecName: "kube-api-access-wk8jl") pod "edf49641-b44d-4dba-af43-9a9da06eb55d" (UID: "edf49641-b44d-4dba-af43-9a9da06eb55d"). InnerVolumeSpecName "kube-api-access-wk8jl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.073133 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-config" (OuterVolumeSpecName: "config") pod "edf49641-b44d-4dba-af43-9a9da06eb55d" (UID: "edf49641-b44d-4dba-af43-9a9da06eb55d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.074348 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "edf49641-b44d-4dba-af43-9a9da06eb55d" (UID: "edf49641-b44d-4dba-af43-9a9da06eb55d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.108641 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "edf49641-b44d-4dba-af43-9a9da06eb55d" (UID: "edf49641-b44d-4dba-af43-9a9da06eb55d"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.119760 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.119811 4997 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.119823 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wk8jl\" (UniqueName: \"kubernetes.io/projected/edf49641-b44d-4dba-af43-9a9da06eb55d-kube-api-access-wk8jl\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.119839 4997 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.119849 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/edf49641-b44d-4dba-af43-9a9da06eb55d-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.523763 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64fc5c4f5b-mnq4n" event={"ID":"edf49641-b44d-4dba-af43-9a9da06eb55d","Type":"ContainerDied","Data":"5baf1bf21c042acc4112da5f4d2072fb8d5c38aa4ba01691dc519354750c00ba"} Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.523822 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64fc5c4f5b-mnq4n" Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.523867 4997 scope.go:117] "RemoveContainer" containerID="d179214c82ee33c9b885acffd78f36623afd0e09d1efcc0df4a6be02aeb8e158" Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.576676 4997 scope.go:117] "RemoveContainer" containerID="7a4dbab571ceebe87a020f8c86c1c678db709ecca7bb391c41b9909b0ab0648f" Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.586639 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-64fc5c4f5b-mnq4n"] Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.594478 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-64fc5c4f5b-mnq4n"] Dec 05 07:20:24 crc kubenswrapper[4997]: I1205 07:20:24.855178 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 05 07:20:25 crc kubenswrapper[4997]: I1205 07:20:25.767359 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edf49641-b44d-4dba-af43-9a9da06eb55d" path="/var/lib/kubelet/pods/edf49641-b44d-4dba-af43-9a9da06eb55d/volumes" Dec 05 07:20:26 crc kubenswrapper[4997]: I1205 07:20:26.780475 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.086802 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.395148 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 05 07:20:30 crc kubenswrapper[4997]: E1205 07:20:30.396468 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf49641-b44d-4dba-af43-9a9da06eb55d" containerName="neutron-api" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.396495 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf49641-b44d-4dba-af43-9a9da06eb55d" containerName="neutron-api" Dec 05 07:20:30 crc kubenswrapper[4997]: E1205 07:20:30.396537 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf49641-b44d-4dba-af43-9a9da06eb55d" containerName="neutron-httpd" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.396545 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf49641-b44d-4dba-af43-9a9da06eb55d" containerName="neutron-httpd" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.396976 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="edf49641-b44d-4dba-af43-9a9da06eb55d" containerName="neutron-httpd" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.397012 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="edf49641-b44d-4dba-af43-9a9da06eb55d" containerName="neutron-api" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.398142 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.404428 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.405453 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.405857 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-lws9v" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.448202 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.449865 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1868d5ba-1b02-41e2-b74c-58579afa6567-openstack-config\") pod \"openstackclient\" (UID: \"1868d5ba-1b02-41e2-b74c-58579afa6567\") " pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.449961 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1868d5ba-1b02-41e2-b74c-58579afa6567-openstack-config-secret\") pod \"openstackclient\" (UID: \"1868d5ba-1b02-41e2-b74c-58579afa6567\") " pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.449992 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjsfz\" (UniqueName: \"kubernetes.io/projected/1868d5ba-1b02-41e2-b74c-58579afa6567-kube-api-access-wjsfz\") pod \"openstackclient\" (UID: \"1868d5ba-1b02-41e2-b74c-58579afa6567\") " pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.450022 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1868d5ba-1b02-41e2-b74c-58579afa6567-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1868d5ba-1b02-41e2-b74c-58579afa6567\") " pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.551699 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1868d5ba-1b02-41e2-b74c-58579afa6567-openstack-config-secret\") pod \"openstackclient\" (UID: \"1868d5ba-1b02-41e2-b74c-58579afa6567\") " pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.551757 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjsfz\" (UniqueName: \"kubernetes.io/projected/1868d5ba-1b02-41e2-b74c-58579afa6567-kube-api-access-wjsfz\") pod \"openstackclient\" (UID: \"1868d5ba-1b02-41e2-b74c-58579afa6567\") " pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.551786 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1868d5ba-1b02-41e2-b74c-58579afa6567-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1868d5ba-1b02-41e2-b74c-58579afa6567\") " pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.551893 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1868d5ba-1b02-41e2-b74c-58579afa6567-openstack-config\") pod \"openstackclient\" (UID: \"1868d5ba-1b02-41e2-b74c-58579afa6567\") " pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.552988 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1868d5ba-1b02-41e2-b74c-58579afa6567-openstack-config\") pod \"openstackclient\" (UID: \"1868d5ba-1b02-41e2-b74c-58579afa6567\") " pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.560235 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1868d5ba-1b02-41e2-b74c-58579afa6567-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1868d5ba-1b02-41e2-b74c-58579afa6567\") " pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.567118 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1868d5ba-1b02-41e2-b74c-58579afa6567-openstack-config-secret\") pod \"openstackclient\" (UID: \"1868d5ba-1b02-41e2-b74c-58579afa6567\") " pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.574225 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjsfz\" (UniqueName: \"kubernetes.io/projected/1868d5ba-1b02-41e2-b74c-58579afa6567-kube-api-access-wjsfz\") pod \"openstackclient\" (UID: \"1868d5ba-1b02-41e2-b74c-58579afa6567\") " pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.731733 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.745868 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.784905 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.797827 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.799588 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.821506 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 07:20:30 crc kubenswrapper[4997]: E1205 07:20:30.908717 4997 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 07:20:30 crc kubenswrapper[4997]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_1868d5ba-1b02-41e2-b74c-58579afa6567_0(7614ffa73b5763010387e984568934f2e6076ac279163ed56dd2946d87393207): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"7614ffa73b5763010387e984568934f2e6076ac279163ed56dd2946d87393207" Netns:"/var/run/netns/85c68c18-1ee7-44a4-a06c-4289d735d6f8" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=7614ffa73b5763010387e984568934f2e6076ac279163ed56dd2946d87393207;K8S_POD_UID=1868d5ba-1b02-41e2-b74c-58579afa6567" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/1868d5ba-1b02-41e2-b74c-58579afa6567]: expected pod UID "1868d5ba-1b02-41e2-b74c-58579afa6567" but got "126f5f4c-a3f5-4744-99cd-691bdd33444b" from Kube API Dec 05 07:20:30 crc kubenswrapper[4997]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 07:20:30 crc kubenswrapper[4997]: > Dec 05 07:20:30 crc kubenswrapper[4997]: E1205 07:20:30.908816 4997 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 07:20:30 crc kubenswrapper[4997]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_1868d5ba-1b02-41e2-b74c-58579afa6567_0(7614ffa73b5763010387e984568934f2e6076ac279163ed56dd2946d87393207): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"7614ffa73b5763010387e984568934f2e6076ac279163ed56dd2946d87393207" Netns:"/var/run/netns/85c68c18-1ee7-44a4-a06c-4289d735d6f8" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=7614ffa73b5763010387e984568934f2e6076ac279163ed56dd2946d87393207;K8S_POD_UID=1868d5ba-1b02-41e2-b74c-58579afa6567" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/1868d5ba-1b02-41e2-b74c-58579afa6567]: expected pod UID "1868d5ba-1b02-41e2-b74c-58579afa6567" but got "126f5f4c-a3f5-4744-99cd-691bdd33444b" from Kube API Dec 05 07:20:30 crc kubenswrapper[4997]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 07:20:30 crc kubenswrapper[4997]: > pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.971110 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvzz5\" (UniqueName: \"kubernetes.io/projected/126f5f4c-a3f5-4744-99cd-691bdd33444b-kube-api-access-bvzz5\") pod \"openstackclient\" (UID: \"126f5f4c-a3f5-4744-99cd-691bdd33444b\") " pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.971192 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/126f5f4c-a3f5-4744-99cd-691bdd33444b-openstack-config-secret\") pod \"openstackclient\" (UID: \"126f5f4c-a3f5-4744-99cd-691bdd33444b\") " pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.971344 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/126f5f4c-a3f5-4744-99cd-691bdd33444b-openstack-config\") pod \"openstackclient\" (UID: \"126f5f4c-a3f5-4744-99cd-691bdd33444b\") " pod="openstack/openstackclient" Dec 05 07:20:30 crc kubenswrapper[4997]: I1205 07:20:30.971384 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/126f5f4c-a3f5-4744-99cd-691bdd33444b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"126f5f4c-a3f5-4744-99cd-691bdd33444b\") " pod="openstack/openstackclient" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.073580 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/126f5f4c-a3f5-4744-99cd-691bdd33444b-openstack-config\") pod \"openstackclient\" (UID: \"126f5f4c-a3f5-4744-99cd-691bdd33444b\") " pod="openstack/openstackclient" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.073686 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/126f5f4c-a3f5-4744-99cd-691bdd33444b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"126f5f4c-a3f5-4744-99cd-691bdd33444b\") " pod="openstack/openstackclient" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.073736 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvzz5\" (UniqueName: \"kubernetes.io/projected/126f5f4c-a3f5-4744-99cd-691bdd33444b-kube-api-access-bvzz5\") pod \"openstackclient\" (UID: \"126f5f4c-a3f5-4744-99cd-691bdd33444b\") " pod="openstack/openstackclient" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.073801 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/126f5f4c-a3f5-4744-99cd-691bdd33444b-openstack-config-secret\") pod \"openstackclient\" (UID: \"126f5f4c-a3f5-4744-99cd-691bdd33444b\") " pod="openstack/openstackclient" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.074741 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/126f5f4c-a3f5-4744-99cd-691bdd33444b-openstack-config\") pod \"openstackclient\" (UID: \"126f5f4c-a3f5-4744-99cd-691bdd33444b\") " pod="openstack/openstackclient" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.081535 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/126f5f4c-a3f5-4744-99cd-691bdd33444b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"126f5f4c-a3f5-4744-99cd-691bdd33444b\") " pod="openstack/openstackclient" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.082182 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/126f5f4c-a3f5-4744-99cd-691bdd33444b-openstack-config-secret\") pod \"openstackclient\" (UID: \"126f5f4c-a3f5-4744-99cd-691bdd33444b\") " pod="openstack/openstackclient" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.092451 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvzz5\" (UniqueName: \"kubernetes.io/projected/126f5f4c-a3f5-4744-99cd-691bdd33444b-kube-api-access-bvzz5\") pod \"openstackclient\" (UID: \"126f5f4c-a3f5-4744-99cd-691bdd33444b\") " pod="openstack/openstackclient" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.248596 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.606606 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.614112 4997 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="1868d5ba-1b02-41e2-b74c-58579afa6567" podUID="126f5f4c-a3f5-4744-99cd-691bdd33444b" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.647857 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.725321 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.727585 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.788174 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjsfz\" (UniqueName: \"kubernetes.io/projected/1868d5ba-1b02-41e2-b74c-58579afa6567-kube-api-access-wjsfz\") pod \"1868d5ba-1b02-41e2-b74c-58579afa6567\" (UID: \"1868d5ba-1b02-41e2-b74c-58579afa6567\") " Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.788726 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1868d5ba-1b02-41e2-b74c-58579afa6567-openstack-config-secret\") pod \"1868d5ba-1b02-41e2-b74c-58579afa6567\" (UID: \"1868d5ba-1b02-41e2-b74c-58579afa6567\") " Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.788807 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1868d5ba-1b02-41e2-b74c-58579afa6567-openstack-config\") pod \"1868d5ba-1b02-41e2-b74c-58579afa6567\" (UID: \"1868d5ba-1b02-41e2-b74c-58579afa6567\") " Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.788966 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1868d5ba-1b02-41e2-b74c-58579afa6567-combined-ca-bundle\") pod \"1868d5ba-1b02-41e2-b74c-58579afa6567\" (UID: \"1868d5ba-1b02-41e2-b74c-58579afa6567\") " Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.791263 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1868d5ba-1b02-41e2-b74c-58579afa6567-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "1868d5ba-1b02-41e2-b74c-58579afa6567" (UID: "1868d5ba-1b02-41e2-b74c-58579afa6567"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.791752 4997 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1868d5ba-1b02-41e2-b74c-58579afa6567-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.800771 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1868d5ba-1b02-41e2-b74c-58579afa6567-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "1868d5ba-1b02-41e2-b74c-58579afa6567" (UID: "1868d5ba-1b02-41e2-b74c-58579afa6567"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.800807 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1868d5ba-1b02-41e2-b74c-58579afa6567-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1868d5ba-1b02-41e2-b74c-58579afa6567" (UID: "1868d5ba-1b02-41e2-b74c-58579afa6567"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.805859 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1868d5ba-1b02-41e2-b74c-58579afa6567-kube-api-access-wjsfz" (OuterVolumeSpecName: "kube-api-access-wjsfz") pod "1868d5ba-1b02-41e2-b74c-58579afa6567" (UID: "1868d5ba-1b02-41e2-b74c-58579afa6567"). InnerVolumeSpecName "kube-api-access-wjsfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.894009 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1868d5ba-1b02-41e2-b74c-58579afa6567-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.894055 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjsfz\" (UniqueName: \"kubernetes.io/projected/1868d5ba-1b02-41e2-b74c-58579afa6567-kube-api-access-wjsfz\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:31 crc kubenswrapper[4997]: I1205 07:20:31.894064 4997 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1868d5ba-1b02-41e2-b74c-58579afa6567-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:32 crc kubenswrapper[4997]: I1205 07:20:32.624663 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"126f5f4c-a3f5-4744-99cd-691bdd33444b","Type":"ContainerStarted","Data":"ecdd9f3a74f95882b1e7871d732e4b4ee89d4d435634c6dd4d33de2c021b455b"} Dec 05 07:20:32 crc kubenswrapper[4997]: I1205 07:20:32.624724 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 07:20:32 crc kubenswrapper[4997]: I1205 07:20:32.643401 4997 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="1868d5ba-1b02-41e2-b74c-58579afa6567" podUID="126f5f4c-a3f5-4744-99cd-691bdd33444b" Dec 05 07:20:33 crc kubenswrapper[4997]: I1205 07:20:33.762797 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1868d5ba-1b02-41e2-b74c-58579afa6567" path="/var/lib/kubelet/pods/1868d5ba-1b02-41e2-b74c-58579afa6567/volumes" Dec 05 07:20:35 crc kubenswrapper[4997]: I1205 07:20:35.890147 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-66698d9c4f-5rmpv"] Dec 05 07:20:35 crc kubenswrapper[4997]: I1205 07:20:35.893177 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:35 crc kubenswrapper[4997]: I1205 07:20:35.896955 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 05 07:20:35 crc kubenswrapper[4997]: I1205 07:20:35.897184 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 05 07:20:35 crc kubenswrapper[4997]: I1205 07:20:35.901099 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 05 07:20:35 crc kubenswrapper[4997]: I1205 07:20:35.914016 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-66698d9c4f-5rmpv"] Dec 05 07:20:35 crc kubenswrapper[4997]: I1205 07:20:35.999036 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/27ff3957-ee08-40be-a41d-02979f192fda-etc-swift\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:35 crc kubenswrapper[4997]: I1205 07:20:35.999121 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-combined-ca-bundle\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:35 crc kubenswrapper[4997]: I1205 07:20:35.999149 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-config-data\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:35 crc kubenswrapper[4997]: I1205 07:20:35.999200 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nlwv\" (UniqueName: \"kubernetes.io/projected/27ff3957-ee08-40be-a41d-02979f192fda-kube-api-access-2nlwv\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:35 crc kubenswrapper[4997]: I1205 07:20:35.999261 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-internal-tls-certs\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:35 crc kubenswrapper[4997]: I1205 07:20:35.999315 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-public-tls-certs\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:35 crc kubenswrapper[4997]: I1205 07:20:35.999346 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27ff3957-ee08-40be-a41d-02979f192fda-log-httpd\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:35 crc kubenswrapper[4997]: I1205 07:20:35.999365 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27ff3957-ee08-40be-a41d-02979f192fda-run-httpd\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:36 crc kubenswrapper[4997]: I1205 07:20:36.100898 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/27ff3957-ee08-40be-a41d-02979f192fda-etc-swift\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:36 crc kubenswrapper[4997]: I1205 07:20:36.100989 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-combined-ca-bundle\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:36 crc kubenswrapper[4997]: I1205 07:20:36.101017 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-config-data\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:36 crc kubenswrapper[4997]: I1205 07:20:36.101072 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nlwv\" (UniqueName: \"kubernetes.io/projected/27ff3957-ee08-40be-a41d-02979f192fda-kube-api-access-2nlwv\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:36 crc kubenswrapper[4997]: I1205 07:20:36.101171 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-internal-tls-certs\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:36 crc kubenswrapper[4997]: I1205 07:20:36.101197 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-public-tls-certs\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:36 crc kubenswrapper[4997]: I1205 07:20:36.101322 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27ff3957-ee08-40be-a41d-02979f192fda-log-httpd\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:36 crc kubenswrapper[4997]: I1205 07:20:36.101350 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27ff3957-ee08-40be-a41d-02979f192fda-run-httpd\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:36 crc kubenswrapper[4997]: I1205 07:20:36.101933 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27ff3957-ee08-40be-a41d-02979f192fda-run-httpd\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:36 crc kubenswrapper[4997]: I1205 07:20:36.110403 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27ff3957-ee08-40be-a41d-02979f192fda-log-httpd\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:36 crc kubenswrapper[4997]: I1205 07:20:36.111883 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-config-data\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:36 crc kubenswrapper[4997]: I1205 07:20:36.112415 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-combined-ca-bundle\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:36 crc kubenswrapper[4997]: I1205 07:20:36.116044 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-internal-tls-certs\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:36 crc kubenswrapper[4997]: I1205 07:20:36.116141 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-public-tls-certs\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:36 crc kubenswrapper[4997]: I1205 07:20:36.121439 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/27ff3957-ee08-40be-a41d-02979f192fda-etc-swift\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:36 crc kubenswrapper[4997]: I1205 07:20:36.136392 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nlwv\" (UniqueName: \"kubernetes.io/projected/27ff3957-ee08-40be-a41d-02979f192fda-kube-api-access-2nlwv\") pod \"swift-proxy-66698d9c4f-5rmpv\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:36 crc kubenswrapper[4997]: I1205 07:20:36.226439 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:38 crc kubenswrapper[4997]: I1205 07:20:38.704413 4997 generic.go:334] "Generic (PLEG): container finished" podID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerID="7005e478ec1d5ab8b440588b9eca063930b2e1ddb1c02e3c5577f50224c89335" exitCode=137 Dec 05 07:20:38 crc kubenswrapper[4997]: I1205 07:20:38.704496 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe5443e-726b-4976-a7a3-c226bd3ec481","Type":"ContainerDied","Data":"7005e478ec1d5ab8b440588b9eca063930b2e1ddb1c02e3c5577f50224c89335"} Dec 05 07:20:40 crc kubenswrapper[4997]: I1205 07:20:40.340273 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="85f5ed73-561d-4876-a44c-a393e820f71c" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.162:8776/healthcheck\": dial tcp 10.217.0.162:8776: connect: connection refused" Dec 05 07:20:40 crc kubenswrapper[4997]: I1205 07:20:40.731485 4997 generic.go:334] "Generic (PLEG): container finished" podID="85f5ed73-561d-4876-a44c-a393e820f71c" containerID="9c1e37f5166ff86419dd377730b522eee072560d2d117f47e2d55de9448b1218" exitCode=137 Dec 05 07:20:40 crc kubenswrapper[4997]: I1205 07:20:40.731555 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"85f5ed73-561d-4876-a44c-a393e820f71c","Type":"ContainerDied","Data":"9c1e37f5166ff86419dd377730b522eee072560d2d117f47e2d55de9448b1218"} Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.104734 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-gf4w4"] Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.106580 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-gf4w4" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.116666 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-gf4w4"] Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.258763 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-pspnv"] Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.260492 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-pspnv" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.266367 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-pspnv"] Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.271179 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc6qr\" (UniqueName: \"kubernetes.io/projected/bdbf2efb-5ab9-4018-86f0-59b9e37db599-kube-api-access-fc6qr\") pod \"nova-api-db-create-gf4w4\" (UID: \"bdbf2efb-5ab9-4018-86f0-59b9e37db599\") " pod="openstack/nova-api-db-create-gf4w4" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.271286 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdbf2efb-5ab9-4018-86f0-59b9e37db599-operator-scripts\") pod \"nova-api-db-create-gf4w4\" (UID: \"bdbf2efb-5ab9-4018-86f0-59b9e37db599\") " pod="openstack/nova-api-db-create-gf4w4" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.314515 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.374062 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-scripts\") pod \"dbe5443e-726b-4976-a7a3-c226bd3ec481\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.374226 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-combined-ca-bundle\") pod \"dbe5443e-726b-4976-a7a3-c226bd3ec481\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.374312 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe5443e-726b-4976-a7a3-c226bd3ec481-log-httpd\") pod \"dbe5443e-726b-4976-a7a3-c226bd3ec481\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.374453 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxr7q\" (UniqueName: \"kubernetes.io/projected/dbe5443e-726b-4976-a7a3-c226bd3ec481-kube-api-access-hxr7q\") pod \"dbe5443e-726b-4976-a7a3-c226bd3ec481\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.374500 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-config-data\") pod \"dbe5443e-726b-4976-a7a3-c226bd3ec481\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.374560 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-sg-core-conf-yaml\") pod \"dbe5443e-726b-4976-a7a3-c226bd3ec481\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.374641 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe5443e-726b-4976-a7a3-c226bd3ec481-run-httpd\") pod \"dbe5443e-726b-4976-a7a3-c226bd3ec481\" (UID: \"dbe5443e-726b-4976-a7a3-c226bd3ec481\") " Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.375073 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc6qr\" (UniqueName: \"kubernetes.io/projected/bdbf2efb-5ab9-4018-86f0-59b9e37db599-kube-api-access-fc6qr\") pod \"nova-api-db-create-gf4w4\" (UID: \"bdbf2efb-5ab9-4018-86f0-59b9e37db599\") " pod="openstack/nova-api-db-create-gf4w4" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.375129 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a88492d3-fed2-456f-9f18-5ddfe089bd4b-operator-scripts\") pod \"nova-cell0-db-create-pspnv\" (UID: \"a88492d3-fed2-456f-9f18-5ddfe089bd4b\") " pod="openstack/nova-cell0-db-create-pspnv" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.375150 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdbf2efb-5ab9-4018-86f0-59b9e37db599-operator-scripts\") pod \"nova-api-db-create-gf4w4\" (UID: \"bdbf2efb-5ab9-4018-86f0-59b9e37db599\") " pod="openstack/nova-api-db-create-gf4w4" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.375183 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrfl2\" (UniqueName: \"kubernetes.io/projected/a88492d3-fed2-456f-9f18-5ddfe089bd4b-kube-api-access-qrfl2\") pod \"nova-cell0-db-create-pspnv\" (UID: \"a88492d3-fed2-456f-9f18-5ddfe089bd4b\") " pod="openstack/nova-cell0-db-create-pspnv" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.376284 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbe5443e-726b-4976-a7a3-c226bd3ec481-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dbe5443e-726b-4976-a7a3-c226bd3ec481" (UID: "dbe5443e-726b-4976-a7a3-c226bd3ec481"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.381108 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdbf2efb-5ab9-4018-86f0-59b9e37db599-operator-scripts\") pod \"nova-api-db-create-gf4w4\" (UID: \"bdbf2efb-5ab9-4018-86f0-59b9e37db599\") " pod="openstack/nova-api-db-create-gf4w4" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.397843 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-8ba6-account-create-update-26l28"] Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.397947 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbe5443e-726b-4976-a7a3-c226bd3ec481-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dbe5443e-726b-4976-a7a3-c226bd3ec481" (UID: "dbe5443e-726b-4976-a7a3-c226bd3ec481"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.398053 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbe5443e-726b-4976-a7a3-c226bd3ec481-kube-api-access-hxr7q" (OuterVolumeSpecName: "kube-api-access-hxr7q") pod "dbe5443e-726b-4976-a7a3-c226bd3ec481" (UID: "dbe5443e-726b-4976-a7a3-c226bd3ec481"). InnerVolumeSpecName "kube-api-access-hxr7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:42 crc kubenswrapper[4997]: E1205 07:20:42.398393 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerName="ceilometer-central-agent" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.398414 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerName="ceilometer-central-agent" Dec 05 07:20:42 crc kubenswrapper[4997]: E1205 07:20:42.398422 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerName="ceilometer-notification-agent" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.398430 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerName="ceilometer-notification-agent" Dec 05 07:20:42 crc kubenswrapper[4997]: E1205 07:20:42.398442 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerName="sg-core" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.398448 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerName="sg-core" Dec 05 07:20:42 crc kubenswrapper[4997]: E1205 07:20:42.398471 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerName="proxy-httpd" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.398477 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerName="proxy-httpd" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.398462 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-scripts" (OuterVolumeSpecName: "scripts") pod "dbe5443e-726b-4976-a7a3-c226bd3ec481" (UID: "dbe5443e-726b-4976-a7a3-c226bd3ec481"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.398731 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerName="proxy-httpd" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.398755 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerName="sg-core" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.398769 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerName="ceilometer-notification-agent" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.398800 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbe5443e-726b-4976-a7a3-c226bd3ec481" containerName="ceilometer-central-agent" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.399681 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8ba6-account-create-update-26l28" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.403488 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.407394 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc6qr\" (UniqueName: \"kubernetes.io/projected/bdbf2efb-5ab9-4018-86f0-59b9e37db599-kube-api-access-fc6qr\") pod \"nova-api-db-create-gf4w4\" (UID: \"bdbf2efb-5ab9-4018-86f0-59b9e37db599\") " pod="openstack/nova-api-db-create-gf4w4" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.411060 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-bmcxb"] Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.413314 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bmcxb" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.432588 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8ba6-account-create-update-26l28"] Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.444277 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-gf4w4" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.478337 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7fae453-348d-44cf-97dc-5e45252e0024-operator-scripts\") pod \"nova-api-8ba6-account-create-update-26l28\" (UID: \"f7fae453-348d-44cf-97dc-5e45252e0024\") " pod="openstack/nova-api-8ba6-account-create-update-26l28" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.478453 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnbdk\" (UniqueName: \"kubernetes.io/projected/dfe50d2f-e9dc-4222-a00d-f879885600bd-kube-api-access-fnbdk\") pod \"nova-cell1-db-create-bmcxb\" (UID: \"dfe50d2f-e9dc-4222-a00d-f879885600bd\") " pod="openstack/nova-cell1-db-create-bmcxb" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.478496 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a88492d3-fed2-456f-9f18-5ddfe089bd4b-operator-scripts\") pod \"nova-cell0-db-create-pspnv\" (UID: \"a88492d3-fed2-456f-9f18-5ddfe089bd4b\") " pod="openstack/nova-cell0-db-create-pspnv" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.478517 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrfl2\" (UniqueName: \"kubernetes.io/projected/a88492d3-fed2-456f-9f18-5ddfe089bd4b-kube-api-access-qrfl2\") pod \"nova-cell0-db-create-pspnv\" (UID: \"a88492d3-fed2-456f-9f18-5ddfe089bd4b\") " pod="openstack/nova-cell0-db-create-pspnv" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.478550 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfe50d2f-e9dc-4222-a00d-f879885600bd-operator-scripts\") pod \"nova-cell1-db-create-bmcxb\" (UID: \"dfe50d2f-e9dc-4222-a00d-f879885600bd\") " pod="openstack/nova-cell1-db-create-bmcxb" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.478589 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kcnw\" (UniqueName: \"kubernetes.io/projected/f7fae453-348d-44cf-97dc-5e45252e0024-kube-api-access-7kcnw\") pod \"nova-api-8ba6-account-create-update-26l28\" (UID: \"f7fae453-348d-44cf-97dc-5e45252e0024\") " pod="openstack/nova-api-8ba6-account-create-update-26l28" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.478866 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxr7q\" (UniqueName: \"kubernetes.io/projected/dbe5443e-726b-4976-a7a3-c226bd3ec481-kube-api-access-hxr7q\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.478887 4997 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe5443e-726b-4976-a7a3-c226bd3ec481-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.478896 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.478906 4997 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe5443e-726b-4976-a7a3-c226bd3ec481-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.485604 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a88492d3-fed2-456f-9f18-5ddfe089bd4b-operator-scripts\") pod \"nova-cell0-db-create-pspnv\" (UID: \"a88492d3-fed2-456f-9f18-5ddfe089bd4b\") " pod="openstack/nova-cell0-db-create-pspnv" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.506192 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrfl2\" (UniqueName: \"kubernetes.io/projected/a88492d3-fed2-456f-9f18-5ddfe089bd4b-kube-api-access-qrfl2\") pod \"nova-cell0-db-create-pspnv\" (UID: \"a88492d3-fed2-456f-9f18-5ddfe089bd4b\") " pod="openstack/nova-cell0-db-create-pspnv" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.507159 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bmcxb"] Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.560942 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "dbe5443e-726b-4976-a7a3-c226bd3ec481" (UID: "dbe5443e-726b-4976-a7a3-c226bd3ec481"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.581215 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7fae453-348d-44cf-97dc-5e45252e0024-operator-scripts\") pod \"nova-api-8ba6-account-create-update-26l28\" (UID: \"f7fae453-348d-44cf-97dc-5e45252e0024\") " pod="openstack/nova-api-8ba6-account-create-update-26l28" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.581380 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnbdk\" (UniqueName: \"kubernetes.io/projected/dfe50d2f-e9dc-4222-a00d-f879885600bd-kube-api-access-fnbdk\") pod \"nova-cell1-db-create-bmcxb\" (UID: \"dfe50d2f-e9dc-4222-a00d-f879885600bd\") " pod="openstack/nova-cell1-db-create-bmcxb" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.581463 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfe50d2f-e9dc-4222-a00d-f879885600bd-operator-scripts\") pod \"nova-cell1-db-create-bmcxb\" (UID: \"dfe50d2f-e9dc-4222-a00d-f879885600bd\") " pod="openstack/nova-cell1-db-create-bmcxb" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.581510 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kcnw\" (UniqueName: \"kubernetes.io/projected/f7fae453-348d-44cf-97dc-5e45252e0024-kube-api-access-7kcnw\") pod \"nova-api-8ba6-account-create-update-26l28\" (UID: \"f7fae453-348d-44cf-97dc-5e45252e0024\") " pod="openstack/nova-api-8ba6-account-create-update-26l28" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.587115 4997 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.588235 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfe50d2f-e9dc-4222-a00d-f879885600bd-operator-scripts\") pod \"nova-cell1-db-create-bmcxb\" (UID: \"dfe50d2f-e9dc-4222-a00d-f879885600bd\") " pod="openstack/nova-cell1-db-create-bmcxb" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.588536 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7fae453-348d-44cf-97dc-5e45252e0024-operator-scripts\") pod \"nova-api-8ba6-account-create-update-26l28\" (UID: \"f7fae453-348d-44cf-97dc-5e45252e0024\") " pod="openstack/nova-api-8ba6-account-create-update-26l28" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.610089 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnbdk\" (UniqueName: \"kubernetes.io/projected/dfe50d2f-e9dc-4222-a00d-f879885600bd-kube-api-access-fnbdk\") pod \"nova-cell1-db-create-bmcxb\" (UID: \"dfe50d2f-e9dc-4222-a00d-f879885600bd\") " pod="openstack/nova-cell1-db-create-bmcxb" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.611303 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kcnw\" (UniqueName: \"kubernetes.io/projected/f7fae453-348d-44cf-97dc-5e45252e0024-kube-api-access-7kcnw\") pod \"nova-api-8ba6-account-create-update-26l28\" (UID: \"f7fae453-348d-44cf-97dc-5e45252e0024\") " pod="openstack/nova-api-8ba6-account-create-update-26l28" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.658120 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-pspnv" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.686167 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-66698d9c4f-5rmpv"] Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.717873 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-2ec3-account-create-update-w5xrv"] Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.719434 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2ec3-account-create-update-w5xrv" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.728480 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.734901 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-2ec3-account-create-update-w5xrv"] Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.739782 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8ba6-account-create-update-26l28" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.794600 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9wjk\" (UniqueName: \"kubernetes.io/projected/a3705104-b487-4dd7-b283-db1edcf99ec5-kube-api-access-v9wjk\") pod \"nova-cell0-2ec3-account-create-update-w5xrv\" (UID: \"a3705104-b487-4dd7-b283-db1edcf99ec5\") " pod="openstack/nova-cell0-2ec3-account-create-update-w5xrv" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.795546 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3705104-b487-4dd7-b283-db1edcf99ec5-operator-scripts\") pod \"nova-cell0-2ec3-account-create-update-w5xrv\" (UID: \"a3705104-b487-4dd7-b283-db1edcf99ec5\") " pod="openstack/nova-cell0-2ec3-account-create-update-w5xrv" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.794918 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-418a-account-create-update-mbkgn"] Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.797517 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-418a-account-create-update-mbkgn" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.802811 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.810504 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bmcxb" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.813657 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe5443e-726b-4976-a7a3-c226bd3ec481","Type":"ContainerDied","Data":"b07e02d6dbb72e64ea73e58b4669858e343f72aae7b1d6391e5a98ab1c793e71"} Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.813713 4997 scope.go:117] "RemoveContainer" containerID="7005e478ec1d5ab8b440588b9eca063930b2e1ddb1c02e3c5577f50224c89335" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.813744 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.822423 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-66698d9c4f-5rmpv" event={"ID":"27ff3957-ee08-40be-a41d-02979f192fda","Type":"ContainerStarted","Data":"8dc02dafe00b2c60dc0fe31d05475bf78494b2f5bdf3443d1d12c8e10e0bcb87"} Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.844084 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-418a-account-create-update-mbkgn"] Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.904148 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3705104-b487-4dd7-b283-db1edcf99ec5-operator-scripts\") pod \"nova-cell0-2ec3-account-create-update-w5xrv\" (UID: \"a3705104-b487-4dd7-b283-db1edcf99ec5\") " pod="openstack/nova-cell0-2ec3-account-create-update-w5xrv" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.904385 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9wjk\" (UniqueName: \"kubernetes.io/projected/a3705104-b487-4dd7-b283-db1edcf99ec5-kube-api-access-v9wjk\") pod \"nova-cell0-2ec3-account-create-update-w5xrv\" (UID: \"a3705104-b487-4dd7-b283-db1edcf99ec5\") " pod="openstack/nova-cell0-2ec3-account-create-update-w5xrv" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.904470 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phk6p\" (UniqueName: \"kubernetes.io/projected/d0d26c9f-3fc4-4626-9212-abac9cd65264-kube-api-access-phk6p\") pod \"nova-cell1-418a-account-create-update-mbkgn\" (UID: \"d0d26c9f-3fc4-4626-9212-abac9cd65264\") " pod="openstack/nova-cell1-418a-account-create-update-mbkgn" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.904575 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0d26c9f-3fc4-4626-9212-abac9cd65264-operator-scripts\") pod \"nova-cell1-418a-account-create-update-mbkgn\" (UID: \"d0d26c9f-3fc4-4626-9212-abac9cd65264\") " pod="openstack/nova-cell1-418a-account-create-update-mbkgn" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.907059 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3705104-b487-4dd7-b283-db1edcf99ec5-operator-scripts\") pod \"nova-cell0-2ec3-account-create-update-w5xrv\" (UID: \"a3705104-b487-4dd7-b283-db1edcf99ec5\") " pod="openstack/nova-cell0-2ec3-account-create-update-w5xrv" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.921522 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dbe5443e-726b-4976-a7a3-c226bd3ec481" (UID: "dbe5443e-726b-4976-a7a3-c226bd3ec481"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.927012 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9wjk\" (UniqueName: \"kubernetes.io/projected/a3705104-b487-4dd7-b283-db1edcf99ec5-kube-api-access-v9wjk\") pod \"nova-cell0-2ec3-account-create-update-w5xrv\" (UID: \"a3705104-b487-4dd7-b283-db1edcf99ec5\") " pod="openstack/nova-cell0-2ec3-account-create-update-w5xrv" Dec 05 07:20:42 crc kubenswrapper[4997]: I1205 07:20:42.976663 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-config-data" (OuterVolumeSpecName: "config-data") pod "dbe5443e-726b-4976-a7a3-c226bd3ec481" (UID: "dbe5443e-726b-4976-a7a3-c226bd3ec481"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.017340 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0d26c9f-3fc4-4626-9212-abac9cd65264-operator-scripts\") pod \"nova-cell1-418a-account-create-update-mbkgn\" (UID: \"d0d26c9f-3fc4-4626-9212-abac9cd65264\") " pod="openstack/nova-cell1-418a-account-create-update-mbkgn" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.017596 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phk6p\" (UniqueName: \"kubernetes.io/projected/d0d26c9f-3fc4-4626-9212-abac9cd65264-kube-api-access-phk6p\") pod \"nova-cell1-418a-account-create-update-mbkgn\" (UID: \"d0d26c9f-3fc4-4626-9212-abac9cd65264\") " pod="openstack/nova-cell1-418a-account-create-update-mbkgn" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.017689 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.017717 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe5443e-726b-4976-a7a3-c226bd3ec481-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.026704 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0d26c9f-3fc4-4626-9212-abac9cd65264-operator-scripts\") pod \"nova-cell1-418a-account-create-update-mbkgn\" (UID: \"d0d26c9f-3fc4-4626-9212-abac9cd65264\") " pod="openstack/nova-cell1-418a-account-create-update-mbkgn" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.043883 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phk6p\" (UniqueName: \"kubernetes.io/projected/d0d26c9f-3fc4-4626-9212-abac9cd65264-kube-api-access-phk6p\") pod \"nova-cell1-418a-account-create-update-mbkgn\" (UID: \"d0d26c9f-3fc4-4626-9212-abac9cd65264\") " pod="openstack/nova-cell1-418a-account-create-update-mbkgn" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.081751 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2ec3-account-create-update-w5xrv" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.082411 4997 scope.go:117] "RemoveContainer" containerID="5a0093fba43766bdbd2468af37a378fd8ab58701871d81188044c42c82c86057" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.099124 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.131536 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-418a-account-create-update-mbkgn" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.214962 4997 scope.go:117] "RemoveContainer" containerID="6b46eaec9f50401eed6987b13e75bc96a2424eaf209614937ac23a0c2cc3c2a0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.223135 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/85f5ed73-561d-4876-a44c-a393e820f71c-etc-machine-id\") pod \"85f5ed73-561d-4876-a44c-a393e820f71c\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.223268 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6zs6\" (UniqueName: \"kubernetes.io/projected/85f5ed73-561d-4876-a44c-a393e820f71c-kube-api-access-v6zs6\") pod \"85f5ed73-561d-4876-a44c-a393e820f71c\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.223343 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-scripts\") pod \"85f5ed73-561d-4876-a44c-a393e820f71c\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.223443 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85f5ed73-561d-4876-a44c-a393e820f71c-logs\") pod \"85f5ed73-561d-4876-a44c-a393e820f71c\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.223516 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-combined-ca-bundle\") pod \"85f5ed73-561d-4876-a44c-a393e820f71c\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.223594 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-config-data\") pod \"85f5ed73-561d-4876-a44c-a393e820f71c\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.223657 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-config-data-custom\") pod \"85f5ed73-561d-4876-a44c-a393e820f71c\" (UID: \"85f5ed73-561d-4876-a44c-a393e820f71c\") " Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.225442 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/85f5ed73-561d-4876-a44c-a393e820f71c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "85f5ed73-561d-4876-a44c-a393e820f71c" (UID: "85f5ed73-561d-4876-a44c-a393e820f71c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.225864 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85f5ed73-561d-4876-a44c-a393e820f71c-logs" (OuterVolumeSpecName: "logs") pod "85f5ed73-561d-4876-a44c-a393e820f71c" (UID: "85f5ed73-561d-4876-a44c-a393e820f71c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.233213 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-scripts" (OuterVolumeSpecName: "scripts") pod "85f5ed73-561d-4876-a44c-a393e820f71c" (UID: "85f5ed73-561d-4876-a44c-a393e820f71c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.233302 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.233989 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85f5ed73-561d-4876-a44c-a393e820f71c-kube-api-access-v6zs6" (OuterVolumeSpecName: "kube-api-access-v6zs6") pod "85f5ed73-561d-4876-a44c-a393e820f71c" (UID: "85f5ed73-561d-4876-a44c-a393e820f71c"). InnerVolumeSpecName "kube-api-access-v6zs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.249908 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "85f5ed73-561d-4876-a44c-a393e820f71c" (UID: "85f5ed73-561d-4876-a44c-a393e820f71c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.273207 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.290953 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-gf4w4"] Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.331510 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85f5ed73-561d-4876-a44c-a393e820f71c-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.331553 4997 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.331570 4997 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/85f5ed73-561d-4876-a44c-a393e820f71c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.331583 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6zs6\" (UniqueName: \"kubernetes.io/projected/85f5ed73-561d-4876-a44c-a393e820f71c-kube-api-access-v6zs6\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.331657 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.352361 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-config-data" (OuterVolumeSpecName: "config-data") pod "85f5ed73-561d-4876-a44c-a393e820f71c" (UID: "85f5ed73-561d-4876-a44c-a393e820f71c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.359212 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:20:43 crc kubenswrapper[4997]: E1205 07:20:43.359861 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85f5ed73-561d-4876-a44c-a393e820f71c" containerName="cinder-api" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.359887 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="85f5ed73-561d-4876-a44c-a393e820f71c" containerName="cinder-api" Dec 05 07:20:43 crc kubenswrapper[4997]: E1205 07:20:43.359905 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85f5ed73-561d-4876-a44c-a393e820f71c" containerName="cinder-api-log" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.359913 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="85f5ed73-561d-4876-a44c-a393e820f71c" containerName="cinder-api-log" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.360097 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="85f5ed73-561d-4876-a44c-a393e820f71c" containerName="cinder-api" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.360117 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="85f5ed73-561d-4876-a44c-a393e820f71c" containerName="cinder-api-log" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.364303 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.367623 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.368829 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.372222 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85f5ed73-561d-4876-a44c-a393e820f71c" (UID: "85f5ed73-561d-4876-a44c-a393e820f71c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.396588 4997 scope.go:117] "RemoveContainer" containerID="3f9c3f41c902d84f5da3f081f73987894ad168e0e10bf5cdaeb82c27e1353e2a" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.396856 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.434581 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.434631 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85f5ed73-561d-4876-a44c-a393e820f71c-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.540126 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.540918 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blhqf\" (UniqueName: \"kubernetes.io/projected/212bb0f0-5d52-4e83-962d-2a14ff302c2f-kube-api-access-blhqf\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.541010 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/212bb0f0-5d52-4e83-962d-2a14ff302c2f-log-httpd\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.541091 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-config-data\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.541171 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/212bb0f0-5d52-4e83-962d-2a14ff302c2f-run-httpd\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.541206 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-scripts\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.541235 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.565808 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-pspnv"] Dec 05 07:20:43 crc kubenswrapper[4997]: W1205 07:20:43.601883 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda88492d3_fed2_456f_9f18_5ddfe089bd4b.slice/crio-125e08a3faaa2a9a8444ce19d214421271f60c8ba054e0b64147cd6f7f2fdc6a WatchSource:0}: Error finding container 125e08a3faaa2a9a8444ce19d214421271f60c8ba054e0b64147cd6f7f2fdc6a: Status 404 returned error can't find the container with id 125e08a3faaa2a9a8444ce19d214421271f60c8ba054e0b64147cd6f7f2fdc6a Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.622962 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8ba6-account-create-update-26l28"] Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.643854 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/212bb0f0-5d52-4e83-962d-2a14ff302c2f-run-httpd\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.643924 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-scripts\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.643957 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.643993 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.644070 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blhqf\" (UniqueName: \"kubernetes.io/projected/212bb0f0-5d52-4e83-962d-2a14ff302c2f-kube-api-access-blhqf\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.644134 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/212bb0f0-5d52-4e83-962d-2a14ff302c2f-log-httpd\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.644182 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-config-data\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.649425 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/212bb0f0-5d52-4e83-962d-2a14ff302c2f-run-httpd\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.652818 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/212bb0f0-5d52-4e83-962d-2a14ff302c2f-log-httpd\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.658372 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.661129 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:20:43 crc kubenswrapper[4997]: E1205 07:20:43.662031 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-blhqf scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="212bb0f0-5d52-4e83-962d-2a14ff302c2f" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.664960 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-config-data\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.668883 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.678148 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-scripts\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.686274 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blhqf\" (UniqueName: \"kubernetes.io/projected/212bb0f0-5d52-4e83-962d-2a14ff302c2f-kube-api-access-blhqf\") pod \"ceilometer-0\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.782381 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbe5443e-726b-4976-a7a3-c226bd3ec481" path="/var/lib/kubelet/pods/dbe5443e-726b-4976-a7a3-c226bd3ec481/volumes" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.847749 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-2ec3-account-create-update-w5xrv"] Dec 05 07:20:43 crc kubenswrapper[4997]: W1205 07:20:43.873727 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3705104_b487_4dd7_b283_db1edcf99ec5.slice/crio-6e53abc955c7fb28c84133c95a478dfa289bad3caddf0fffb08e48700a7e4823 WatchSource:0}: Error finding container 6e53abc955c7fb28c84133c95a478dfa289bad3caddf0fffb08e48700a7e4823: Status 404 returned error can't find the container with id 6e53abc955c7fb28c84133c95a478dfa289bad3caddf0fffb08e48700a7e4823 Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.874647 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"126f5f4c-a3f5-4744-99cd-691bdd33444b","Type":"ContainerStarted","Data":"6ea597f614ba1df9ed5420084187fc9eec112551e2d7b84b30b72cdf64053515"} Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.879431 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-pspnv" event={"ID":"a88492d3-fed2-456f-9f18-5ddfe089bd4b","Type":"ContainerStarted","Data":"125e08a3faaa2a9a8444ce19d214421271f60c8ba054e0b64147cd6f7f2fdc6a"} Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.885119 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bmcxb"] Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.888086 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8ba6-account-create-update-26l28" event={"ID":"f7fae453-348d-44cf-97dc-5e45252e0024","Type":"ContainerStarted","Data":"fe5fa46ed5ebf38b4baab2059dbae0cb5ca954377818a8e3871ab234d2fe41d5"} Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.909167 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"85f5ed73-561d-4876-a44c-a393e820f71c","Type":"ContainerDied","Data":"7345a8dacac46a5bf0f92440d2e59e3b62f96f451dec6dda324a613d4ab4c6fc"} Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.909247 4997 scope.go:117] "RemoveContainer" containerID="9c1e37f5166ff86419dd377730b522eee072560d2d117f47e2d55de9448b1218" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.909432 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.921093 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.361051181 podStartE2EDuration="13.92106386s" podCreationTimestamp="2025-12-05 07:20:30 +0000 UTC" firstStartedPulling="2025-12-05 07:20:31.727366166 +0000 UTC m=+1532.256273427" lastFinishedPulling="2025-12-05 07:20:42.287378845 +0000 UTC m=+1542.816286106" observedRunningTime="2025-12-05 07:20:43.898132599 +0000 UTC m=+1544.427039880" watchObservedRunningTime="2025-12-05 07:20:43.92106386 +0000 UTC m=+1544.449971121" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.949992 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-gf4w4" event={"ID":"bdbf2efb-5ab9-4018-86f0-59b9e37db599","Type":"ContainerStarted","Data":"50a7e2c25e5cd83e508a065267024218c87cd9a607ba3f50b881843c73167053"} Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.975709 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.976111 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.976145 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-66698d9c4f-5rmpv" event={"ID":"27ff3957-ee08-40be-a41d-02979f192fda","Type":"ContainerStarted","Data":"6f749a47e57e47c7006263ba4386dbf5b34d5aa77a620b464309e9c5d820056b"} Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.992001 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 05 07:20:43 crc kubenswrapper[4997]: I1205 07:20:43.992800 4997 scope.go:117] "RemoveContainer" containerID="c1feee6a9ec30b94a61c86ebbb440bcaaaef2f31275039429a3176d29f79d1dd" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.009368 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.044810 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.048201 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.053300 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.053871 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.059519 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.060858 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.069028 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-sg-core-conf-yaml\") pod \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.069200 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-scripts\") pod \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.069314 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-config-data\") pod \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.069407 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/212bb0f0-5d52-4e83-962d-2a14ff302c2f-log-httpd\") pod \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.069445 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-combined-ca-bundle\") pod \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.069469 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blhqf\" (UniqueName: \"kubernetes.io/projected/212bb0f0-5d52-4e83-962d-2a14ff302c2f-kube-api-access-blhqf\") pod \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.069510 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/212bb0f0-5d52-4e83-962d-2a14ff302c2f-run-httpd\") pod \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\" (UID: \"212bb0f0-5d52-4e83-962d-2a14ff302c2f\") " Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.073001 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/212bb0f0-5d52-4e83-962d-2a14ff302c2f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "212bb0f0-5d52-4e83-962d-2a14ff302c2f" (UID: "212bb0f0-5d52-4e83-962d-2a14ff302c2f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.073071 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-418a-account-create-update-mbkgn"] Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.076300 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/212bb0f0-5d52-4e83-962d-2a14ff302c2f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "212bb0f0-5d52-4e83-962d-2a14ff302c2f" (UID: "212bb0f0-5d52-4e83-962d-2a14ff302c2f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.079916 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "212bb0f0-5d52-4e83-962d-2a14ff302c2f" (UID: "212bb0f0-5d52-4e83-962d-2a14ff302c2f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.087398 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/212bb0f0-5d52-4e83-962d-2a14ff302c2f-kube-api-access-blhqf" (OuterVolumeSpecName: "kube-api-access-blhqf") pod "212bb0f0-5d52-4e83-962d-2a14ff302c2f" (UID: "212bb0f0-5d52-4e83-962d-2a14ff302c2f"). InnerVolumeSpecName "kube-api-access-blhqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.092710 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-scripts" (OuterVolumeSpecName: "scripts") pod "212bb0f0-5d52-4e83-962d-2a14ff302c2f" (UID: "212bb0f0-5d52-4e83-962d-2a14ff302c2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.109299 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-config-data" (OuterVolumeSpecName: "config-data") pod "212bb0f0-5d52-4e83-962d-2a14ff302c2f" (UID: "212bb0f0-5d52-4e83-962d-2a14ff302c2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.110878 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "212bb0f0-5d52-4e83-962d-2a14ff302c2f" (UID: "212bb0f0-5d52-4e83-962d-2a14ff302c2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:44 crc kubenswrapper[4997]: W1205 07:20:44.111231 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0d26c9f_3fc4_4626_9212_abac9cd65264.slice/crio-4e8824eef354e0666c5b28b531d2b45cec11ff741f96ae50029e3f30842be7b2 WatchSource:0}: Error finding container 4e8824eef354e0666c5b28b531d2b45cec11ff741f96ae50029e3f30842be7b2: Status 404 returned error can't find the container with id 4e8824eef354e0666c5b28b531d2b45cec11ff741f96ae50029e3f30842be7b2 Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.173137 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mn64\" (UniqueName: \"kubernetes.io/projected/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-kube-api-access-6mn64\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.173225 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-config-data\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.173247 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-scripts\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.173290 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.173316 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-logs\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.173341 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-config-data-custom\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.173393 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.173412 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.173440 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.173494 4997 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.173505 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.173513 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.173524 4997 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/212bb0f0-5d52-4e83-962d-2a14ff302c2f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.173532 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/212bb0f0-5d52-4e83-962d-2a14ff302c2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.173545 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blhqf\" (UniqueName: \"kubernetes.io/projected/212bb0f0-5d52-4e83-962d-2a14ff302c2f-kube-api-access-blhqf\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.173555 4997 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/212bb0f0-5d52-4e83-962d-2a14ff302c2f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.275264 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mn64\" (UniqueName: \"kubernetes.io/projected/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-kube-api-access-6mn64\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.275390 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-config-data\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.275417 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-scripts\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.275491 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.275525 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-logs\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.275556 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-config-data-custom\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.275692 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.275715 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.275747 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.276420 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-logs\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.276505 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.281714 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-scripts\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.283209 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.289629 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.295504 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-config-data-custom\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.300509 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-config-data\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.303241 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.310247 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mn64\" (UniqueName: \"kubernetes.io/projected/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-kube-api-access-6mn64\") pod \"cinder-api-0\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.378990 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.991370 4997 generic.go:334] "Generic (PLEG): container finished" podID="d0d26c9f-3fc4-4626-9212-abac9cd65264" containerID="0b1d64b7f2fccb29886cde4842d5d80e274474638cdc49f408b3dfb215f85fd4" exitCode=0 Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.991448 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-418a-account-create-update-mbkgn" event={"ID":"d0d26c9f-3fc4-4626-9212-abac9cd65264","Type":"ContainerDied","Data":"0b1d64b7f2fccb29886cde4842d5d80e274474638cdc49f408b3dfb215f85fd4"} Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.991900 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-418a-account-create-update-mbkgn" event={"ID":"d0d26c9f-3fc4-4626-9212-abac9cd65264","Type":"ContainerStarted","Data":"4e8824eef354e0666c5b28b531d2b45cec11ff741f96ae50029e3f30842be7b2"} Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.994695 4997 generic.go:334] "Generic (PLEG): container finished" podID="bdbf2efb-5ab9-4018-86f0-59b9e37db599" containerID="49dec3f8ec745fed264ce35f45332d233e06748c10111467fac5a3e086c7c245" exitCode=0 Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.994747 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-gf4w4" event={"ID":"bdbf2efb-5ab9-4018-86f0-59b9e37db599","Type":"ContainerDied","Data":"49dec3f8ec745fed264ce35f45332d233e06748c10111467fac5a3e086c7c245"} Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.999246 4997 generic.go:334] "Generic (PLEG): container finished" podID="a3705104-b487-4dd7-b283-db1edcf99ec5" containerID="dce22783874c237d14eefd9cc90fd2f082fc0d59392932b78b0131a8307d7d6b" exitCode=0 Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.999312 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2ec3-account-create-update-w5xrv" event={"ID":"a3705104-b487-4dd7-b283-db1edcf99ec5","Type":"ContainerDied","Data":"dce22783874c237d14eefd9cc90fd2f082fc0d59392932b78b0131a8307d7d6b"} Dec 05 07:20:44 crc kubenswrapper[4997]: I1205 07:20:44.999466 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2ec3-account-create-update-w5xrv" event={"ID":"a3705104-b487-4dd7-b283-db1edcf99ec5","Type":"ContainerStarted","Data":"6e53abc955c7fb28c84133c95a478dfa289bad3caddf0fffb08e48700a7e4823"} Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.002982 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-66698d9c4f-5rmpv" event={"ID":"27ff3957-ee08-40be-a41d-02979f192fda","Type":"ContainerStarted","Data":"025cfd66536602dc692bb16ce499c3e7d95d59c6864c476cec229bbace2c59d3"} Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.003629 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.003661 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.005260 4997 generic.go:334] "Generic (PLEG): container finished" podID="a88492d3-fed2-456f-9f18-5ddfe089bd4b" containerID="5eaf0178fac7b663a1feb1d79be38e9e15c37e90ce1a021fc1f087e1974f2fa5" exitCode=0 Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.005397 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-pspnv" event={"ID":"a88492d3-fed2-456f-9f18-5ddfe089bd4b","Type":"ContainerDied","Data":"5eaf0178fac7b663a1feb1d79be38e9e15c37e90ce1a021fc1f087e1974f2fa5"} Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.007723 4997 generic.go:334] "Generic (PLEG): container finished" podID="dfe50d2f-e9dc-4222-a00d-f879885600bd" containerID="3fac4170eed53a3f784a00877568373b3cd58e632ea290c38068fa431ff21b48" exitCode=0 Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.007786 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bmcxb" event={"ID":"dfe50d2f-e9dc-4222-a00d-f879885600bd","Type":"ContainerDied","Data":"3fac4170eed53a3f784a00877568373b3cd58e632ea290c38068fa431ff21b48"} Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.007812 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bmcxb" event={"ID":"dfe50d2f-e9dc-4222-a00d-f879885600bd","Type":"ContainerStarted","Data":"2501cc38e2cecccee97646b46b866141b83637110c383d2435d75585ab8ae424"} Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.010056 4997 generic.go:334] "Generic (PLEG): container finished" podID="f7fae453-348d-44cf-97dc-5e45252e0024" containerID="8f71ef8ac8eb6da4977035d50f2e232a8ea7cd965854f2b988cfb9cf8185b3a4" exitCode=0 Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.010218 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8ba6-account-create-update-26l28" event={"ID":"f7fae453-348d-44cf-97dc-5e45252e0024","Type":"ContainerDied","Data":"8f71ef8ac8eb6da4977035d50f2e232a8ea7cd965854f2b988cfb9cf8185b3a4"} Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.013639 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: W1205 07:20:45.053148 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d3915c5_30d2_43be_abd4_438cd9d8ebf7.slice/crio-5a78b3f3d78b1711cb5549e03a9ce76384ae3d42e9846ce209eed741535e916f WatchSource:0}: Error finding container 5a78b3f3d78b1711cb5549e03a9ce76384ae3d42e9846ce209eed741535e916f: Status 404 returned error can't find the container with id 5a78b3f3d78b1711cb5549e03a9ce76384ae3d42e9846ce209eed741535e916f Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.060921 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.099905 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-66698d9c4f-5rmpv" podStartSLOduration=10.09987502 podStartE2EDuration="10.09987502s" podCreationTimestamp="2025-12-05 07:20:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:20:45.083808305 +0000 UTC m=+1545.612715576" watchObservedRunningTime="2025-12-05 07:20:45.09987502 +0000 UTC m=+1545.628782281" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.254917 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.296606 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.335570 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.340172 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.344148 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.344168 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.357681 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.412818 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.412885 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsv6t\" (UniqueName: \"kubernetes.io/projected/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-kube-api-access-qsv6t\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.412910 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.412994 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-config-data\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.413026 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-scripts\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.413053 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-run-httpd\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.413100 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-log-httpd\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.523708 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-config-data\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.523829 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-scripts\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.523874 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-run-httpd\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.523937 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-log-httpd\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.524010 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.524056 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsv6t\" (UniqueName: \"kubernetes.io/projected/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-kube-api-access-qsv6t\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.524089 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.532822 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-log-httpd\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.535830 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-run-httpd\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.538260 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.547885 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-config-data\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.550592 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.555992 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-scripts\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.602168 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsv6t\" (UniqueName: \"kubernetes.io/projected/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-kube-api-access-qsv6t\") pod \"ceilometer-0\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.676677 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.803439 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="212bb0f0-5d52-4e83-962d-2a14ff302c2f" path="/var/lib/kubelet/pods/212bb0f0-5d52-4e83-962d-2a14ff302c2f/volumes" Dec 05 07:20:45 crc kubenswrapper[4997]: I1205 07:20:45.814060 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85f5ed73-561d-4876-a44c-a393e820f71c" path="/var/lib/kubelet/pods/85f5ed73-561d-4876-a44c-a393e820f71c/volumes" Dec 05 07:20:46 crc kubenswrapper[4997]: I1205 07:20:46.026310 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d3915c5-30d2-43be-abd4-438cd9d8ebf7","Type":"ContainerStarted","Data":"5a78b3f3d78b1711cb5549e03a9ce76384ae3d42e9846ce209eed741535e916f"} Dec 05 07:20:46 crc kubenswrapper[4997]: I1205 07:20:46.187457 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:20:46 crc kubenswrapper[4997]: I1205 07:20:46.449970 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:20:46 crc kubenswrapper[4997]: I1205 07:20:46.450728 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="494f1331-dcea-4745-add0-50794d9e38c9" containerName="glance-log" containerID="cri-o://60b0bdb40b4513ec19380aa4ea11f57950298ef219df1a43ffc7f22f094a77b2" gracePeriod=30 Dec 05 07:20:46 crc kubenswrapper[4997]: I1205 07:20:46.450944 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="494f1331-dcea-4745-add0-50794d9e38c9" containerName="glance-httpd" containerID="cri-o://6d255e415e3ec2bac137bdf03a4ca5835d173b925ab6f3c048d7f48d26a75cae" gracePeriod=30 Dec 05 07:20:46 crc kubenswrapper[4997]: I1205 07:20:46.523181 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-pspnv" Dec 05 07:20:46 crc kubenswrapper[4997]: I1205 07:20:46.554091 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrfl2\" (UniqueName: \"kubernetes.io/projected/a88492d3-fed2-456f-9f18-5ddfe089bd4b-kube-api-access-qrfl2\") pod \"a88492d3-fed2-456f-9f18-5ddfe089bd4b\" (UID: \"a88492d3-fed2-456f-9f18-5ddfe089bd4b\") " Dec 05 07:20:46 crc kubenswrapper[4997]: I1205 07:20:46.554143 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a88492d3-fed2-456f-9f18-5ddfe089bd4b-operator-scripts\") pod \"a88492d3-fed2-456f-9f18-5ddfe089bd4b\" (UID: \"a88492d3-fed2-456f-9f18-5ddfe089bd4b\") " Dec 05 07:20:46 crc kubenswrapper[4997]: I1205 07:20:46.555667 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a88492d3-fed2-456f-9f18-5ddfe089bd4b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a88492d3-fed2-456f-9f18-5ddfe089bd4b" (UID: "a88492d3-fed2-456f-9f18-5ddfe089bd4b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:20:46 crc kubenswrapper[4997]: I1205 07:20:46.566520 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a88492d3-fed2-456f-9f18-5ddfe089bd4b-kube-api-access-qrfl2" (OuterVolumeSpecName: "kube-api-access-qrfl2") pod "a88492d3-fed2-456f-9f18-5ddfe089bd4b" (UID: "a88492d3-fed2-456f-9f18-5ddfe089bd4b"). InnerVolumeSpecName "kube-api-access-qrfl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:46 crc kubenswrapper[4997]: I1205 07:20:46.656836 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrfl2\" (UniqueName: \"kubernetes.io/projected/a88492d3-fed2-456f-9f18-5ddfe089bd4b-kube-api-access-qrfl2\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:46 crc kubenswrapper[4997]: I1205 07:20:46.657258 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a88492d3-fed2-456f-9f18-5ddfe089bd4b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.022853 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8ba6-account-create-update-26l28" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.035836 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2ec3-account-create-update-w5xrv" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.065691 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3705104-b487-4dd7-b283-db1edcf99ec5-operator-scripts\") pod \"a3705104-b487-4dd7-b283-db1edcf99ec5\" (UID: \"a3705104-b487-4dd7-b283-db1edcf99ec5\") " Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.065754 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9wjk\" (UniqueName: \"kubernetes.io/projected/a3705104-b487-4dd7-b283-db1edcf99ec5-kube-api-access-v9wjk\") pod \"a3705104-b487-4dd7-b283-db1edcf99ec5\" (UID: \"a3705104-b487-4dd7-b283-db1edcf99ec5\") " Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.066018 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kcnw\" (UniqueName: \"kubernetes.io/projected/f7fae453-348d-44cf-97dc-5e45252e0024-kube-api-access-7kcnw\") pod \"f7fae453-348d-44cf-97dc-5e45252e0024\" (UID: \"f7fae453-348d-44cf-97dc-5e45252e0024\") " Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.066078 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7fae453-348d-44cf-97dc-5e45252e0024-operator-scripts\") pod \"f7fae453-348d-44cf-97dc-5e45252e0024\" (UID: \"f7fae453-348d-44cf-97dc-5e45252e0024\") " Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.066427 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3705104-b487-4dd7-b283-db1edcf99ec5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a3705104-b487-4dd7-b283-db1edcf99ec5" (UID: "a3705104-b487-4dd7-b283-db1edcf99ec5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.066691 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3705104-b487-4dd7-b283-db1edcf99ec5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.067216 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7fae453-348d-44cf-97dc-5e45252e0024-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f7fae453-348d-44cf-97dc-5e45252e0024" (UID: "f7fae453-348d-44cf-97dc-5e45252e0024"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.072923 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3705104-b487-4dd7-b283-db1edcf99ec5-kube-api-access-v9wjk" (OuterVolumeSpecName: "kube-api-access-v9wjk") pod "a3705104-b487-4dd7-b283-db1edcf99ec5" (UID: "a3705104-b487-4dd7-b283-db1edcf99ec5"). InnerVolumeSpecName "kube-api-access-v9wjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.073783 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bmcxb" event={"ID":"dfe50d2f-e9dc-4222-a00d-f879885600bd","Type":"ContainerDied","Data":"2501cc38e2cecccee97646b46b866141b83637110c383d2435d75585ab8ae424"} Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.073831 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2501cc38e2cecccee97646b46b866141b83637110c383d2435d75585ab8ae424" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.074023 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7fae453-348d-44cf-97dc-5e45252e0024-kube-api-access-7kcnw" (OuterVolumeSpecName: "kube-api-access-7kcnw") pod "f7fae453-348d-44cf-97dc-5e45252e0024" (UID: "f7fae453-348d-44cf-97dc-5e45252e0024"). InnerVolumeSpecName "kube-api-access-7kcnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.100461 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8ba6-account-create-update-26l28" event={"ID":"f7fae453-348d-44cf-97dc-5e45252e0024","Type":"ContainerDied","Data":"fe5fa46ed5ebf38b4baab2059dbae0cb5ca954377818a8e3871ab234d2fe41d5"} Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.100528 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe5fa46ed5ebf38b4baab2059dbae0cb5ca954377818a8e3871ab234d2fe41d5" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.100664 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8ba6-account-create-update-26l28" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.121629 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-418a-account-create-update-mbkgn" event={"ID":"d0d26c9f-3fc4-4626-9212-abac9cd65264","Type":"ContainerDied","Data":"4e8824eef354e0666c5b28b531d2b45cec11ff741f96ae50029e3f30842be7b2"} Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.121698 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e8824eef354e0666c5b28b531d2b45cec11ff741f96ae50029e3f30842be7b2" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.134074 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-418a-account-create-update-mbkgn" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.134114 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2ec3-account-create-update-w5xrv" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.134163 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2ec3-account-create-update-w5xrv" event={"ID":"a3705104-b487-4dd7-b283-db1edcf99ec5","Type":"ContainerDied","Data":"6e53abc955c7fb28c84133c95a478dfa289bad3caddf0fffb08e48700a7e4823"} Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.134255 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e53abc955c7fb28c84133c95a478dfa289bad3caddf0fffb08e48700a7e4823" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.138628 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-gf4w4" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.138854 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d3915c5-30d2-43be-abd4-438cd9d8ebf7","Type":"ContainerStarted","Data":"057a82bd61c9e67d2a9d3ed1d85e6973ab9ffca83e3bd81b232a535bb4ad1866"} Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.144754 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f","Type":"ContainerStarted","Data":"3ceb714740247b51dca6da7adce9cf54f20f09220e25eedd2307c1e3c036b8c0"} Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.160360 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-pspnv" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.160884 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-pspnv" event={"ID":"a88492d3-fed2-456f-9f18-5ddfe089bd4b","Type":"ContainerDied","Data":"125e08a3faaa2a9a8444ce19d214421271f60c8ba054e0b64147cd6f7f2fdc6a"} Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.160951 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="125e08a3faaa2a9a8444ce19d214421271f60c8ba054e0b64147cd6f7f2fdc6a" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.165763 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bmcxb" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.170798 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc6qr\" (UniqueName: \"kubernetes.io/projected/bdbf2efb-5ab9-4018-86f0-59b9e37db599-kube-api-access-fc6qr\") pod \"bdbf2efb-5ab9-4018-86f0-59b9e37db599\" (UID: \"bdbf2efb-5ab9-4018-86f0-59b9e37db599\") " Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.171131 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdbf2efb-5ab9-4018-86f0-59b9e37db599-operator-scripts\") pod \"bdbf2efb-5ab9-4018-86f0-59b9e37db599\" (UID: \"bdbf2efb-5ab9-4018-86f0-59b9e37db599\") " Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.171238 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phk6p\" (UniqueName: \"kubernetes.io/projected/d0d26c9f-3fc4-4626-9212-abac9cd65264-kube-api-access-phk6p\") pod \"d0d26c9f-3fc4-4626-9212-abac9cd65264\" (UID: \"d0d26c9f-3fc4-4626-9212-abac9cd65264\") " Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.171455 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0d26c9f-3fc4-4626-9212-abac9cd65264-operator-scripts\") pod \"d0d26c9f-3fc4-4626-9212-abac9cd65264\" (UID: \"d0d26c9f-3fc4-4626-9212-abac9cd65264\") " Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.172148 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kcnw\" (UniqueName: \"kubernetes.io/projected/f7fae453-348d-44cf-97dc-5e45252e0024-kube-api-access-7kcnw\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.172186 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7fae453-348d-44cf-97dc-5e45252e0024-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.172202 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9wjk\" (UniqueName: \"kubernetes.io/projected/a3705104-b487-4dd7-b283-db1edcf99ec5-kube-api-access-v9wjk\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.172937 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0d26c9f-3fc4-4626-9212-abac9cd65264-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d0d26c9f-3fc4-4626-9212-abac9cd65264" (UID: "d0d26c9f-3fc4-4626-9212-abac9cd65264"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.177099 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdbf2efb-5ab9-4018-86f0-59b9e37db599-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bdbf2efb-5ab9-4018-86f0-59b9e37db599" (UID: "bdbf2efb-5ab9-4018-86f0-59b9e37db599"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.184664 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdbf2efb-5ab9-4018-86f0-59b9e37db599-kube-api-access-fc6qr" (OuterVolumeSpecName: "kube-api-access-fc6qr") pod "bdbf2efb-5ab9-4018-86f0-59b9e37db599" (UID: "bdbf2efb-5ab9-4018-86f0-59b9e37db599"). InnerVolumeSpecName "kube-api-access-fc6qr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.198119 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0d26c9f-3fc4-4626-9212-abac9cd65264-kube-api-access-phk6p" (OuterVolumeSpecName: "kube-api-access-phk6p") pod "d0d26c9f-3fc4-4626-9212-abac9cd65264" (UID: "d0d26c9f-3fc4-4626-9212-abac9cd65264"). InnerVolumeSpecName "kube-api-access-phk6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.206479 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-gf4w4" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.206928 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-gf4w4" event={"ID":"bdbf2efb-5ab9-4018-86f0-59b9e37db599","Type":"ContainerDied","Data":"50a7e2c25e5cd83e508a065267024218c87cd9a607ba3f50b881843c73167053"} Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.207156 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50a7e2c25e5cd83e508a065267024218c87cd9a607ba3f50b881843c73167053" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.236713 4997 generic.go:334] "Generic (PLEG): container finished" podID="494f1331-dcea-4745-add0-50794d9e38c9" containerID="60b0bdb40b4513ec19380aa4ea11f57950298ef219df1a43ffc7f22f094a77b2" exitCode=143 Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.236777 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"494f1331-dcea-4745-add0-50794d9e38c9","Type":"ContainerDied","Data":"60b0bdb40b4513ec19380aa4ea11f57950298ef219df1a43ffc7f22f094a77b2"} Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.277537 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnbdk\" (UniqueName: \"kubernetes.io/projected/dfe50d2f-e9dc-4222-a00d-f879885600bd-kube-api-access-fnbdk\") pod \"dfe50d2f-e9dc-4222-a00d-f879885600bd\" (UID: \"dfe50d2f-e9dc-4222-a00d-f879885600bd\") " Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.277735 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfe50d2f-e9dc-4222-a00d-f879885600bd-operator-scripts\") pod \"dfe50d2f-e9dc-4222-a00d-f879885600bd\" (UID: \"dfe50d2f-e9dc-4222-a00d-f879885600bd\") " Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.278577 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdbf2efb-5ab9-4018-86f0-59b9e37db599-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.278600 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phk6p\" (UniqueName: \"kubernetes.io/projected/d0d26c9f-3fc4-4626-9212-abac9cd65264-kube-api-access-phk6p\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.278626 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0d26c9f-3fc4-4626-9212-abac9cd65264-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.278679 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc6qr\" (UniqueName: \"kubernetes.io/projected/bdbf2efb-5ab9-4018-86f0-59b9e37db599-kube-api-access-fc6qr\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.292057 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfe50d2f-e9dc-4222-a00d-f879885600bd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dfe50d2f-e9dc-4222-a00d-f879885600bd" (UID: "dfe50d2f-e9dc-4222-a00d-f879885600bd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.297858 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfe50d2f-e9dc-4222-a00d-f879885600bd-kube-api-access-fnbdk" (OuterVolumeSpecName: "kube-api-access-fnbdk") pod "dfe50d2f-e9dc-4222-a00d-f879885600bd" (UID: "dfe50d2f-e9dc-4222-a00d-f879885600bd"). InnerVolumeSpecName "kube-api-access-fnbdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.380442 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnbdk\" (UniqueName: \"kubernetes.io/projected/dfe50d2f-e9dc-4222-a00d-f879885600bd-kube-api-access-fnbdk\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.380482 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfe50d2f-e9dc-4222-a00d-f879885600bd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.539760 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.540079 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" containerName="glance-log" containerID="cri-o://919328b940a3d4e6e28f76104075700a77242e6a1b78fc6e92aac2aa62da5236" gracePeriod=30 Dec 05 07:20:47 crc kubenswrapper[4997]: I1205 07:20:47.540217 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" containerName="glance-httpd" containerID="cri-o://48adac2190d73ab746eea173c0da78a53022e28b46d2c8de7580024fb3ddb353" gracePeriod=30 Dec 05 07:20:48 crc kubenswrapper[4997]: I1205 07:20:48.251663 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d3915c5-30d2-43be-abd4-438cd9d8ebf7","Type":"ContainerStarted","Data":"8f36673252ccfc4fc6b7d58b342de4054e621d8bc6722f917b4cf09a6dd9bcaf"} Dec 05 07:20:48 crc kubenswrapper[4997]: I1205 07:20:48.253343 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 05 07:20:48 crc kubenswrapper[4997]: I1205 07:20:48.255227 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f","Type":"ContainerStarted","Data":"67362fe303ad19856160a40743ba3d5c2e4efe4c0269e64ad5a7763a0de865eb"} Dec 05 07:20:48 crc kubenswrapper[4997]: I1205 07:20:48.255280 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f","Type":"ContainerStarted","Data":"52f17ac84e23b345d3bafec1cf2403e818e1a9f66defc8fbae156bb75206d250"} Dec 05 07:20:48 crc kubenswrapper[4997]: I1205 07:20:48.258136 4997 generic.go:334] "Generic (PLEG): container finished" podID="2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" containerID="919328b940a3d4e6e28f76104075700a77242e6a1b78fc6e92aac2aa62da5236" exitCode=143 Dec 05 07:20:48 crc kubenswrapper[4997]: I1205 07:20:48.258282 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bmcxb" Dec 05 07:20:48 crc kubenswrapper[4997]: I1205 07:20:48.258755 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4","Type":"ContainerDied","Data":"919328b940a3d4e6e28f76104075700a77242e6a1b78fc6e92aac2aa62da5236"} Dec 05 07:20:48 crc kubenswrapper[4997]: I1205 07:20:48.258987 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-418a-account-create-update-mbkgn" Dec 05 07:20:48 crc kubenswrapper[4997]: I1205 07:20:48.278161 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.27813747 podStartE2EDuration="5.27813747s" podCreationTimestamp="2025-12-05 07:20:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:20:48.276164527 +0000 UTC m=+1548.805071788" watchObservedRunningTime="2025-12-05 07:20:48.27813747 +0000 UTC m=+1548.807044751" Dec 05 07:20:48 crc kubenswrapper[4997]: I1205 07:20:48.427455 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:20:49 crc kubenswrapper[4997]: I1205 07:20:49.276469 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f","Type":"ContainerStarted","Data":"fdff4564bab55db2af38192a28710bd2c95c8abfb451c8db33b363c73fa52270"} Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.238017 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.256832 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-scripts\") pod \"494f1331-dcea-4745-add0-50794d9e38c9\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.256990 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/494f1331-dcea-4745-add0-50794d9e38c9-httpd-run\") pod \"494f1331-dcea-4745-add0-50794d9e38c9\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.257756 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/494f1331-dcea-4745-add0-50794d9e38c9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "494f1331-dcea-4745-add0-50794d9e38c9" (UID: "494f1331-dcea-4745-add0-50794d9e38c9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.257911 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttgzh\" (UniqueName: \"kubernetes.io/projected/494f1331-dcea-4745-add0-50794d9e38c9-kube-api-access-ttgzh\") pod \"494f1331-dcea-4745-add0-50794d9e38c9\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.257961 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-public-tls-certs\") pod \"494f1331-dcea-4745-add0-50794d9e38c9\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.258496 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/494f1331-dcea-4745-add0-50794d9e38c9-logs\") pod \"494f1331-dcea-4745-add0-50794d9e38c9\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.258540 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-combined-ca-bundle\") pod \"494f1331-dcea-4745-add0-50794d9e38c9\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.258725 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-config-data\") pod \"494f1331-dcea-4745-add0-50794d9e38c9\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.258870 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"494f1331-dcea-4745-add0-50794d9e38c9\" (UID: \"494f1331-dcea-4745-add0-50794d9e38c9\") " Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.259942 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/494f1331-dcea-4745-add0-50794d9e38c9-logs" (OuterVolumeSpecName: "logs") pod "494f1331-dcea-4745-add0-50794d9e38c9" (UID: "494f1331-dcea-4745-add0-50794d9e38c9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.260047 4997 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/494f1331-dcea-4745-add0-50794d9e38c9-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.270724 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/494f1331-dcea-4745-add0-50794d9e38c9-kube-api-access-ttgzh" (OuterVolumeSpecName: "kube-api-access-ttgzh") pod "494f1331-dcea-4745-add0-50794d9e38c9" (UID: "494f1331-dcea-4745-add0-50794d9e38c9"). InnerVolumeSpecName "kube-api-access-ttgzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.270602 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-scripts" (OuterVolumeSpecName: "scripts") pod "494f1331-dcea-4745-add0-50794d9e38c9" (UID: "494f1331-dcea-4745-add0-50794d9e38c9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.287830 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "494f1331-dcea-4745-add0-50794d9e38c9" (UID: "494f1331-dcea-4745-add0-50794d9e38c9"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.326046 4997 generic.go:334] "Generic (PLEG): container finished" podID="494f1331-dcea-4745-add0-50794d9e38c9" containerID="6d255e415e3ec2bac137bdf03a4ca5835d173b925ab6f3c048d7f48d26a75cae" exitCode=0 Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.326284 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.326685 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"494f1331-dcea-4745-add0-50794d9e38c9","Type":"ContainerDied","Data":"6d255e415e3ec2bac137bdf03a4ca5835d173b925ab6f3c048d7f48d26a75cae"} Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.326906 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"494f1331-dcea-4745-add0-50794d9e38c9","Type":"ContainerDied","Data":"a5d9bd50afec7230b300e10c9fb8e9dccf422550391720d17984c22fb562aebe"} Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.327404 4997 scope.go:117] "RemoveContainer" containerID="6d255e415e3ec2bac137bdf03a4ca5835d173b925ab6f3c048d7f48d26a75cae" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.328257 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "494f1331-dcea-4745-add0-50794d9e38c9" (UID: "494f1331-dcea-4745-add0-50794d9e38c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.344993 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-config-data" (OuterVolumeSpecName: "config-data") pod "494f1331-dcea-4745-add0-50794d9e38c9" (UID: "494f1331-dcea-4745-add0-50794d9e38c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.355773 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f","Type":"ContainerStarted","Data":"593eadf1941f2907ac94627f9a4ca8fb140aa60eb770ecdc965c4ebe69e45c1b"} Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.356264 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerName="ceilometer-central-agent" containerID="cri-o://52f17ac84e23b345d3bafec1cf2403e818e1a9f66defc8fbae156bb75206d250" gracePeriod=30 Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.356423 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.356487 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerName="proxy-httpd" containerID="cri-o://593eadf1941f2907ac94627f9a4ca8fb140aa60eb770ecdc965c4ebe69e45c1b" gracePeriod=30 Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.356568 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerName="sg-core" containerID="cri-o://fdff4564bab55db2af38192a28710bd2c95c8abfb451c8db33b363c73fa52270" gracePeriod=30 Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.356652 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerName="ceilometer-notification-agent" containerID="cri-o://67362fe303ad19856160a40743ba3d5c2e4efe4c0269e64ad5a7763a0de865eb" gracePeriod=30 Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.382742 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.392114 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttgzh\" (UniqueName: \"kubernetes.io/projected/494f1331-dcea-4745-add0-50794d9e38c9-kube-api-access-ttgzh\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.392180 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/494f1331-dcea-4745-add0-50794d9e38c9-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.392199 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.392216 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.392280 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.394264 4997 scope.go:117] "RemoveContainer" containerID="60b0bdb40b4513ec19380aa4ea11f57950298ef219df1a43ffc7f22f094a77b2" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.399799 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.845749456 podStartE2EDuration="5.399766883s" podCreationTimestamp="2025-12-05 07:20:45 +0000 UTC" firstStartedPulling="2025-12-05 07:20:46.221804022 +0000 UTC m=+1546.750711283" lastFinishedPulling="2025-12-05 07:20:49.775821459 +0000 UTC m=+1550.304728710" observedRunningTime="2025-12-05 07:20:50.394952132 +0000 UTC m=+1550.923859413" watchObservedRunningTime="2025-12-05 07:20:50.399766883 +0000 UTC m=+1550.928674144" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.414929 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "494f1331-dcea-4745-add0-50794d9e38c9" (UID: "494f1331-dcea-4745-add0-50794d9e38c9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.432829 4997 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.458636 4997 scope.go:117] "RemoveContainer" containerID="6d255e415e3ec2bac137bdf03a4ca5835d173b925ab6f3c048d7f48d26a75cae" Dec 05 07:20:50 crc kubenswrapper[4997]: E1205 07:20:50.459721 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d255e415e3ec2bac137bdf03a4ca5835d173b925ab6f3c048d7f48d26a75cae\": container with ID starting with 6d255e415e3ec2bac137bdf03a4ca5835d173b925ab6f3c048d7f48d26a75cae not found: ID does not exist" containerID="6d255e415e3ec2bac137bdf03a4ca5835d173b925ab6f3c048d7f48d26a75cae" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.459764 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d255e415e3ec2bac137bdf03a4ca5835d173b925ab6f3c048d7f48d26a75cae"} err="failed to get container status \"6d255e415e3ec2bac137bdf03a4ca5835d173b925ab6f3c048d7f48d26a75cae\": rpc error: code = NotFound desc = could not find container \"6d255e415e3ec2bac137bdf03a4ca5835d173b925ab6f3c048d7f48d26a75cae\": container with ID starting with 6d255e415e3ec2bac137bdf03a4ca5835d173b925ab6f3c048d7f48d26a75cae not found: ID does not exist" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.459793 4997 scope.go:117] "RemoveContainer" containerID="60b0bdb40b4513ec19380aa4ea11f57950298ef219df1a43ffc7f22f094a77b2" Dec 05 07:20:50 crc kubenswrapper[4997]: E1205 07:20:50.460989 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60b0bdb40b4513ec19380aa4ea11f57950298ef219df1a43ffc7f22f094a77b2\": container with ID starting with 60b0bdb40b4513ec19380aa4ea11f57950298ef219df1a43ffc7f22f094a77b2 not found: ID does not exist" containerID="60b0bdb40b4513ec19380aa4ea11f57950298ef219df1a43ffc7f22f094a77b2" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.461027 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60b0bdb40b4513ec19380aa4ea11f57950298ef219df1a43ffc7f22f094a77b2"} err="failed to get container status \"60b0bdb40b4513ec19380aa4ea11f57950298ef219df1a43ffc7f22f094a77b2\": rpc error: code = NotFound desc = could not find container \"60b0bdb40b4513ec19380aa4ea11f57950298ef219df1a43ffc7f22f094a77b2\": container with ID starting with 60b0bdb40b4513ec19380aa4ea11f57950298ef219df1a43ffc7f22f094a77b2 not found: ID does not exist" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.496693 4997 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.496743 4997 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/494f1331-dcea-4745-add0-50794d9e38c9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.664233 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.672931 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.702767 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:20:50 crc kubenswrapper[4997]: E1205 07:20:50.703215 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7fae453-348d-44cf-97dc-5e45252e0024" containerName="mariadb-account-create-update" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.703239 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7fae453-348d-44cf-97dc-5e45252e0024" containerName="mariadb-account-create-update" Dec 05 07:20:50 crc kubenswrapper[4997]: E1205 07:20:50.703249 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdbf2efb-5ab9-4018-86f0-59b9e37db599" containerName="mariadb-database-create" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.703256 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdbf2efb-5ab9-4018-86f0-59b9e37db599" containerName="mariadb-database-create" Dec 05 07:20:50 crc kubenswrapper[4997]: E1205 07:20:50.703288 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="494f1331-dcea-4745-add0-50794d9e38c9" containerName="glance-log" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.703295 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="494f1331-dcea-4745-add0-50794d9e38c9" containerName="glance-log" Dec 05 07:20:50 crc kubenswrapper[4997]: E1205 07:20:50.703310 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="494f1331-dcea-4745-add0-50794d9e38c9" containerName="glance-httpd" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.703316 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="494f1331-dcea-4745-add0-50794d9e38c9" containerName="glance-httpd" Dec 05 07:20:50 crc kubenswrapper[4997]: E1205 07:20:50.703327 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0d26c9f-3fc4-4626-9212-abac9cd65264" containerName="mariadb-account-create-update" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.703333 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d26c9f-3fc4-4626-9212-abac9cd65264" containerName="mariadb-account-create-update" Dec 05 07:20:50 crc kubenswrapper[4997]: E1205 07:20:50.703342 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a88492d3-fed2-456f-9f18-5ddfe089bd4b" containerName="mariadb-database-create" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.703348 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a88492d3-fed2-456f-9f18-5ddfe089bd4b" containerName="mariadb-database-create" Dec 05 07:20:50 crc kubenswrapper[4997]: E1205 07:20:50.703360 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3705104-b487-4dd7-b283-db1edcf99ec5" containerName="mariadb-account-create-update" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.703365 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3705104-b487-4dd7-b283-db1edcf99ec5" containerName="mariadb-account-create-update" Dec 05 07:20:50 crc kubenswrapper[4997]: E1205 07:20:50.703373 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfe50d2f-e9dc-4222-a00d-f879885600bd" containerName="mariadb-database-create" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.703380 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfe50d2f-e9dc-4222-a00d-f879885600bd" containerName="mariadb-database-create" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.703580 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7fae453-348d-44cf-97dc-5e45252e0024" containerName="mariadb-account-create-update" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.703594 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0d26c9f-3fc4-4626-9212-abac9cd65264" containerName="mariadb-account-create-update" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.703605 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3705104-b487-4dd7-b283-db1edcf99ec5" containerName="mariadb-account-create-update" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.703636 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a88492d3-fed2-456f-9f18-5ddfe089bd4b" containerName="mariadb-database-create" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.703646 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="494f1331-dcea-4745-add0-50794d9e38c9" containerName="glance-httpd" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.703655 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="494f1331-dcea-4745-add0-50794d9e38c9" containerName="glance-log" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.703668 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfe50d2f-e9dc-4222-a00d-f879885600bd" containerName="mariadb-database-create" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.703679 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdbf2efb-5ab9-4018-86f0-59b9e37db599" containerName="mariadb-database-create" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.705156 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.708210 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.708768 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.728737 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.804816 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-logs\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.804888 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.804933 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.804971 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-scripts\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.805282 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.805406 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-config-data\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.805458 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m4pw\" (UniqueName: \"kubernetes.io/projected/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-kube-api-access-6m4pw\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.805605 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.907862 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-scripts\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.908380 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.908420 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-config-data\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.908447 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m4pw\" (UniqueName: \"kubernetes.io/projected/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-kube-api-access-6m4pw\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.908487 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.908539 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-logs\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.908576 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.908600 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.909099 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.909180 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.912816 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-logs\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.914934 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-scripts\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.917682 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.919340 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.926119 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-config-data\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.938978 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m4pw\" (UniqueName: \"kubernetes.io/projected/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-kube-api-access-6m4pw\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:50 crc kubenswrapper[4997]: I1205 07:20:50.946152 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " pod="openstack/glance-default-external-api-0" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.141668 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.440456 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.445539 4997 generic.go:334] "Generic (PLEG): container finished" podID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerID="593eadf1941f2907ac94627f9a4ca8fb140aa60eb770ecdc965c4ebe69e45c1b" exitCode=0 Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.446004 4997 generic.go:334] "Generic (PLEG): container finished" podID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerID="fdff4564bab55db2af38192a28710bd2c95c8abfb451c8db33b363c73fa52270" exitCode=2 Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.446100 4997 generic.go:334] "Generic (PLEG): container finished" podID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerID="67362fe303ad19856160a40743ba3d5c2e4efe4c0269e64ad5a7763a0de865eb" exitCode=0 Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.446245 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f","Type":"ContainerDied","Data":"593eadf1941f2907ac94627f9a4ca8fb140aa60eb770ecdc965c4ebe69e45c1b"} Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.446369 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f","Type":"ContainerDied","Data":"fdff4564bab55db2af38192a28710bd2c95c8abfb451c8db33b363c73fa52270"} Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.446458 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f","Type":"ContainerDied","Data":"67362fe303ad19856160a40743ba3d5c2e4efe4c0269e64ad5a7763a0de865eb"} Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.462706 4997 generic.go:334] "Generic (PLEG): container finished" podID="2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" containerID="48adac2190d73ab746eea173c0da78a53022e28b46d2c8de7580024fb3ddb353" exitCode=0 Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.462758 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4","Type":"ContainerDied","Data":"48adac2190d73ab746eea173c0da78a53022e28b46d2c8de7580024fb3ddb353"} Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.483278 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.557568 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.643579 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-scripts\") pod \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.644007 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-combined-ca-bundle\") pod \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.644698 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z54hg\" (UniqueName: \"kubernetes.io/projected/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-kube-api-access-z54hg\") pod \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.644745 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-logs\") pod \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.644793 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.644938 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-httpd-run\") pod \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.645106 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-config-data\") pod \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.645193 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-internal-tls-certs\") pod \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\" (UID: \"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4\") " Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.646056 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" (UID: "2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.646078 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-logs" (OuterVolumeSpecName: "logs") pod "2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" (UID: "2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.662976 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" (UID: "2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.668628 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-kube-api-access-z54hg" (OuterVolumeSpecName: "kube-api-access-z54hg") pod "2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" (UID: "2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4"). InnerVolumeSpecName "kube-api-access-z54hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.673477 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-scripts" (OuterVolumeSpecName: "scripts") pod "2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" (UID: "2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.715850 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" (UID: "2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.723990 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-config-data" (OuterVolumeSpecName: "config-data") pod "2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" (UID: "2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.748434 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.748471 4997 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.748483 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.748492 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.748501 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.748510 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z54hg\" (UniqueName: \"kubernetes.io/projected/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-kube-api-access-z54hg\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.748521 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.778368 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="494f1331-dcea-4745-add0-50794d9e38c9" path="/var/lib/kubelet/pods/494f1331-dcea-4745-add0-50794d9e38c9/volumes" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.779752 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" (UID: "2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.780591 4997 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.851742 4997 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:51 crc kubenswrapper[4997]: I1205 07:20:51.851783 4997 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:52 crc kubenswrapper[4997]: W1205 07:20:52.132020 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67d9e01d_5189_4aac_8f1a_b1f09fe138b8.slice/crio-784b13fb1238ccd16ae49390528e4338365e3eaa7c310f19fc69bb52dc47a6b7 WatchSource:0}: Error finding container 784b13fb1238ccd16ae49390528e4338365e3eaa7c310f19fc69bb52dc47a6b7: Status 404 returned error can't find the container with id 784b13fb1238ccd16ae49390528e4338365e3eaa7c310f19fc69bb52dc47a6b7 Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.132133 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.475833 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67d9e01d-5189-4aac-8f1a-b1f09fe138b8","Type":"ContainerStarted","Data":"784b13fb1238ccd16ae49390528e4338365e3eaa7c310f19fc69bb52dc47a6b7"} Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.490006 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4","Type":"ContainerDied","Data":"bf48eed91f67604c5ca7387bd16157d7eb043065e6137abaa35f1e5fd434d20f"} Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.490282 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.490326 4997 scope.go:117] "RemoveContainer" containerID="48adac2190d73ab746eea173c0da78a53022e28b46d2c8de7580024fb3ddb353" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.533169 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.624974 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.639001 4997 scope.go:117] "RemoveContainer" containerID="919328b940a3d4e6e28f76104075700a77242e6a1b78fc6e92aac2aa62da5236" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.644606 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:20:52 crc kubenswrapper[4997]: E1205 07:20:52.645199 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" containerName="glance-httpd" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.645229 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" containerName="glance-httpd" Dec 05 07:20:52 crc kubenswrapper[4997]: E1205 07:20:52.645262 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" containerName="glance-log" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.645272 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" containerName="glance-log" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.645506 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" containerName="glance-log" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.645536 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" containerName="glance-httpd" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.647043 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.652075 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.652213 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.652349 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.679999 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.680138 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e08a39e3-f267-496d-80b3-b12a9eef14c1-logs\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.680189 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.680216 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.680271 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e08a39e3-f267-496d-80b3-b12a9eef14c1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.680367 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kpvt\" (UniqueName: \"kubernetes.io/projected/e08a39e3-f267-496d-80b3-b12a9eef14c1-kube-api-access-5kpvt\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.680502 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.680552 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.782224 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.782295 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.782338 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.782400 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e08a39e3-f267-496d-80b3-b12a9eef14c1-logs\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.782434 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.782456 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.782486 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e08a39e3-f267-496d-80b3-b12a9eef14c1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.782535 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kpvt\" (UniqueName: \"kubernetes.io/projected/e08a39e3-f267-496d-80b3-b12a9eef14c1-kube-api-access-5kpvt\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.783240 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e08a39e3-f267-496d-80b3-b12a9eef14c1-logs\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.783389 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e08a39e3-f267-496d-80b3-b12a9eef14c1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.783994 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.806331 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.806820 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.807501 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.809388 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.822003 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kpvt\" (UniqueName: \"kubernetes.io/projected/e08a39e3-f267-496d-80b3-b12a9eef14c1-kube-api-access-5kpvt\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.835853 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " pod="openstack/glance-default-internal-api-0" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.937869 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-qzrvc"] Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.939294 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-qzrvc" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.942000 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.942098 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-qwgjd" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.944786 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.959969 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-qzrvc"] Dec 05 07:20:52 crc kubenswrapper[4997]: I1205 07:20:52.992370 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 07:20:53 crc kubenswrapper[4997]: I1205 07:20:53.094973 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-scripts\") pod \"nova-cell0-conductor-db-sync-qzrvc\" (UID: \"778a1f81-32bc-4c34-b77d-3675d24d5467\") " pod="openstack/nova-cell0-conductor-db-sync-qzrvc" Dec 05 07:20:53 crc kubenswrapper[4997]: I1205 07:20:53.095478 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vjb2\" (UniqueName: \"kubernetes.io/projected/778a1f81-32bc-4c34-b77d-3675d24d5467-kube-api-access-5vjb2\") pod \"nova-cell0-conductor-db-sync-qzrvc\" (UID: \"778a1f81-32bc-4c34-b77d-3675d24d5467\") " pod="openstack/nova-cell0-conductor-db-sync-qzrvc" Dec 05 07:20:53 crc kubenswrapper[4997]: I1205 07:20:53.095906 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-qzrvc\" (UID: \"778a1f81-32bc-4c34-b77d-3675d24d5467\") " pod="openstack/nova-cell0-conductor-db-sync-qzrvc" Dec 05 07:20:53 crc kubenswrapper[4997]: I1205 07:20:53.096090 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-config-data\") pod \"nova-cell0-conductor-db-sync-qzrvc\" (UID: \"778a1f81-32bc-4c34-b77d-3675d24d5467\") " pod="openstack/nova-cell0-conductor-db-sync-qzrvc" Dec 05 07:20:53 crc kubenswrapper[4997]: I1205 07:20:53.198280 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-scripts\") pod \"nova-cell0-conductor-db-sync-qzrvc\" (UID: \"778a1f81-32bc-4c34-b77d-3675d24d5467\") " pod="openstack/nova-cell0-conductor-db-sync-qzrvc" Dec 05 07:20:53 crc kubenswrapper[4997]: I1205 07:20:53.198343 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vjb2\" (UniqueName: \"kubernetes.io/projected/778a1f81-32bc-4c34-b77d-3675d24d5467-kube-api-access-5vjb2\") pod \"nova-cell0-conductor-db-sync-qzrvc\" (UID: \"778a1f81-32bc-4c34-b77d-3675d24d5467\") " pod="openstack/nova-cell0-conductor-db-sync-qzrvc" Dec 05 07:20:53 crc kubenswrapper[4997]: I1205 07:20:53.198447 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-qzrvc\" (UID: \"778a1f81-32bc-4c34-b77d-3675d24d5467\") " pod="openstack/nova-cell0-conductor-db-sync-qzrvc" Dec 05 07:20:53 crc kubenswrapper[4997]: I1205 07:20:53.198488 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-config-data\") pod \"nova-cell0-conductor-db-sync-qzrvc\" (UID: \"778a1f81-32bc-4c34-b77d-3675d24d5467\") " pod="openstack/nova-cell0-conductor-db-sync-qzrvc" Dec 05 07:20:53 crc kubenswrapper[4997]: I1205 07:20:53.206665 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-qzrvc\" (UID: \"778a1f81-32bc-4c34-b77d-3675d24d5467\") " pod="openstack/nova-cell0-conductor-db-sync-qzrvc" Dec 05 07:20:53 crc kubenswrapper[4997]: I1205 07:20:53.208088 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-config-data\") pod \"nova-cell0-conductor-db-sync-qzrvc\" (UID: \"778a1f81-32bc-4c34-b77d-3675d24d5467\") " pod="openstack/nova-cell0-conductor-db-sync-qzrvc" Dec 05 07:20:53 crc kubenswrapper[4997]: I1205 07:20:53.213958 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-scripts\") pod \"nova-cell0-conductor-db-sync-qzrvc\" (UID: \"778a1f81-32bc-4c34-b77d-3675d24d5467\") " pod="openstack/nova-cell0-conductor-db-sync-qzrvc" Dec 05 07:20:53 crc kubenswrapper[4997]: I1205 07:20:53.228687 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vjb2\" (UniqueName: \"kubernetes.io/projected/778a1f81-32bc-4c34-b77d-3675d24d5467-kube-api-access-5vjb2\") pod \"nova-cell0-conductor-db-sync-qzrvc\" (UID: \"778a1f81-32bc-4c34-b77d-3675d24d5467\") " pod="openstack/nova-cell0-conductor-db-sync-qzrvc" Dec 05 07:20:53 crc kubenswrapper[4997]: I1205 07:20:53.280636 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-qzrvc" Dec 05 07:20:53 crc kubenswrapper[4997]: I1205 07:20:53.529417 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67d9e01d-5189-4aac-8f1a-b1f09fe138b8","Type":"ContainerStarted","Data":"d4b6f46bb0d9eab2ada6b1349324da20be55e9bbe3ce62bb67888afe1b1a469c"} Dec 05 07:20:53 crc kubenswrapper[4997]: I1205 07:20:53.770225 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4" path="/var/lib/kubelet/pods/2ed6d9d7-6364-4dc5-b017-b9ba5e07f4e4/volumes" Dec 05 07:20:53 crc kubenswrapper[4997]: I1205 07:20:53.771328 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:20:53 crc kubenswrapper[4997]: W1205 07:20:53.783856 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode08a39e3_f267_496d_80b3_b12a9eef14c1.slice/crio-89452f4406872d0491819e7c170b3af0dc88bb38d26f4432ce97d3d75d550177 WatchSource:0}: Error finding container 89452f4406872d0491819e7c170b3af0dc88bb38d26f4432ce97d3d75d550177: Status 404 returned error can't find the container with id 89452f4406872d0491819e7c170b3af0dc88bb38d26f4432ce97d3d75d550177 Dec 05 07:20:53 crc kubenswrapper[4997]: I1205 07:20:53.888643 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-qzrvc"] Dec 05 07:20:53 crc kubenswrapper[4997]: W1205 07:20:53.901352 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod778a1f81_32bc_4c34_b77d_3675d24d5467.slice/crio-31601a6b0af246491220bd2c26af1630dbd55e6abc42e0acb34fbca7b310b7ff WatchSource:0}: Error finding container 31601a6b0af246491220bd2c26af1630dbd55e6abc42e0acb34fbca7b310b7ff: Status 404 returned error can't find the container with id 31601a6b0af246491220bd2c26af1630dbd55e6abc42e0acb34fbca7b310b7ff Dec 05 07:20:54 crc kubenswrapper[4997]: I1205 07:20:54.557479 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67d9e01d-5189-4aac-8f1a-b1f09fe138b8","Type":"ContainerStarted","Data":"793a4abf63252321ba4fb0dafb87c02e599c7ee609d534abf01e11eb3eff2e86"} Dec 05 07:20:54 crc kubenswrapper[4997]: I1205 07:20:54.566350 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e08a39e3-f267-496d-80b3-b12a9eef14c1","Type":"ContainerStarted","Data":"5624c5a4633297b9a0b3880002a7fb4866e561e1bf970698aec5d0b949bbfcc6"} Dec 05 07:20:54 crc kubenswrapper[4997]: I1205 07:20:54.566417 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e08a39e3-f267-496d-80b3-b12a9eef14c1","Type":"ContainerStarted","Data":"89452f4406872d0491819e7c170b3af0dc88bb38d26f4432ce97d3d75d550177"} Dec 05 07:20:54 crc kubenswrapper[4997]: I1205 07:20:54.586530 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-qzrvc" event={"ID":"778a1f81-32bc-4c34-b77d-3675d24d5467","Type":"ContainerStarted","Data":"31601a6b0af246491220bd2c26af1630dbd55e6abc42e0acb34fbca7b310b7ff"} Dec 05 07:20:54 crc kubenswrapper[4997]: I1205 07:20:54.588232 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.5882054960000005 podStartE2EDuration="4.588205496s" podCreationTimestamp="2025-12-05 07:20:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:20:54.582342557 +0000 UTC m=+1555.111249818" watchObservedRunningTime="2025-12-05 07:20:54.588205496 +0000 UTC m=+1555.117112757" Dec 05 07:20:55 crc kubenswrapper[4997]: I1205 07:20:55.602831 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e08a39e3-f267-496d-80b3-b12a9eef14c1","Type":"ContainerStarted","Data":"d09f8f75d78ec8de394fe2d3878fbdc0a819558c405bcad43068ea535ea0a421"} Dec 05 07:20:55 crc kubenswrapper[4997]: I1205 07:20:55.633035 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.633002954 podStartE2EDuration="3.633002954s" podCreationTimestamp="2025-12-05 07:20:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:20:55.628063121 +0000 UTC m=+1556.156970402" watchObservedRunningTime="2025-12-05 07:20:55.633002954 +0000 UTC m=+1556.161910215" Dec 05 07:20:55 crc kubenswrapper[4997]: E1205 07:20:55.985482 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bd457ad_f0e3_4de8_9bf4_8206c4cd661f.slice/crio-conmon-52f17ac84e23b345d3bafec1cf2403e818e1a9f66defc8fbae156bb75206d250.scope\": RecentStats: unable to find data in memory cache]" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.377216 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.495279 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-run-httpd\") pod \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.495863 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsv6t\" (UniqueName: \"kubernetes.io/projected/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-kube-api-access-qsv6t\") pod \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.495910 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-config-data\") pod \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.496100 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-sg-core-conf-yaml\") pod \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.496141 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-log-httpd\") pod \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.496140 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" (UID: "1bd457ad-f0e3-4de8-9bf4-8206c4cd661f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.496227 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-combined-ca-bundle\") pod \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.496264 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-scripts\") pod \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\" (UID: \"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f\") " Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.496723 4997 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.497023 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" (UID: "1bd457ad-f0e3-4de8-9bf4-8206c4cd661f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.507508 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-scripts" (OuterVolumeSpecName: "scripts") pod "1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" (UID: "1bd457ad-f0e3-4de8-9bf4-8206c4cd661f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.521010 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-kube-api-access-qsv6t" (OuterVolumeSpecName: "kube-api-access-qsv6t") pod "1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" (UID: "1bd457ad-f0e3-4de8-9bf4-8206c4cd661f"). InnerVolumeSpecName "kube-api-access-qsv6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.542442 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" (UID: "1bd457ad-f0e3-4de8-9bf4-8206c4cd661f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.605370 4997 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.605418 4997 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.605428 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.605441 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsv6t\" (UniqueName: \"kubernetes.io/projected/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-kube-api-access-qsv6t\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.618803 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" (UID: "1bd457ad-f0e3-4de8-9bf4-8206c4cd661f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.626782 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-config-data" (OuterVolumeSpecName: "config-data") pod "1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" (UID: "1bd457ad-f0e3-4de8-9bf4-8206c4cd661f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.656910 4997 generic.go:334] "Generic (PLEG): container finished" podID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerID="52f17ac84e23b345d3bafec1cf2403e818e1a9f66defc8fbae156bb75206d250" exitCode=0 Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.658103 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.659854 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f","Type":"ContainerDied","Data":"52f17ac84e23b345d3bafec1cf2403e818e1a9f66defc8fbae156bb75206d250"} Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.659949 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bd457ad-f0e3-4de8-9bf4-8206c4cd661f","Type":"ContainerDied","Data":"3ceb714740247b51dca6da7adce9cf54f20f09220e25eedd2307c1e3c036b8c0"} Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.659973 4997 scope.go:117] "RemoveContainer" containerID="593eadf1941f2907ac94627f9a4ca8fb140aa60eb770ecdc965c4ebe69e45c1b" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.699566 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.707020 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.707059 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.722894 4997 scope.go:117] "RemoveContainer" containerID="fdff4564bab55db2af38192a28710bd2c95c8abfb451c8db33b363c73fa52270" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.725702 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.750214 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:20:56 crc kubenswrapper[4997]: E1205 07:20:56.750712 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerName="ceilometer-notification-agent" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.750727 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerName="ceilometer-notification-agent" Dec 05 07:20:56 crc kubenswrapper[4997]: E1205 07:20:56.750737 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerName="ceilometer-central-agent" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.750743 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerName="ceilometer-central-agent" Dec 05 07:20:56 crc kubenswrapper[4997]: E1205 07:20:56.750764 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerName="sg-core" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.750771 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerName="sg-core" Dec 05 07:20:56 crc kubenswrapper[4997]: E1205 07:20:56.750780 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerName="proxy-httpd" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.750786 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerName="proxy-httpd" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.751016 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerName="proxy-httpd" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.751037 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerName="sg-core" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.751052 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerName="ceilometer-notification-agent" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.751060 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" containerName="ceilometer-central-agent" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.752921 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.756960 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.757328 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.760109 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.769883 4997 scope.go:117] "RemoveContainer" containerID="67362fe303ad19856160a40743ba3d5c2e4efe4c0269e64ad5a7763a0de865eb" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.810408 4997 scope.go:117] "RemoveContainer" containerID="52f17ac84e23b345d3bafec1cf2403e818e1a9f66defc8fbae156bb75206d250" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.856731 4997 scope.go:117] "RemoveContainer" containerID="593eadf1941f2907ac94627f9a4ca8fb140aa60eb770ecdc965c4ebe69e45c1b" Dec 05 07:20:56 crc kubenswrapper[4997]: E1205 07:20:56.857408 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"593eadf1941f2907ac94627f9a4ca8fb140aa60eb770ecdc965c4ebe69e45c1b\": container with ID starting with 593eadf1941f2907ac94627f9a4ca8fb140aa60eb770ecdc965c4ebe69e45c1b not found: ID does not exist" containerID="593eadf1941f2907ac94627f9a4ca8fb140aa60eb770ecdc965c4ebe69e45c1b" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.857461 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"593eadf1941f2907ac94627f9a4ca8fb140aa60eb770ecdc965c4ebe69e45c1b"} err="failed to get container status \"593eadf1941f2907ac94627f9a4ca8fb140aa60eb770ecdc965c4ebe69e45c1b\": rpc error: code = NotFound desc = could not find container \"593eadf1941f2907ac94627f9a4ca8fb140aa60eb770ecdc965c4ebe69e45c1b\": container with ID starting with 593eadf1941f2907ac94627f9a4ca8fb140aa60eb770ecdc965c4ebe69e45c1b not found: ID does not exist" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.857498 4997 scope.go:117] "RemoveContainer" containerID="fdff4564bab55db2af38192a28710bd2c95c8abfb451c8db33b363c73fa52270" Dec 05 07:20:56 crc kubenswrapper[4997]: E1205 07:20:56.857996 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdff4564bab55db2af38192a28710bd2c95c8abfb451c8db33b363c73fa52270\": container with ID starting with fdff4564bab55db2af38192a28710bd2c95c8abfb451c8db33b363c73fa52270 not found: ID does not exist" containerID="fdff4564bab55db2af38192a28710bd2c95c8abfb451c8db33b363c73fa52270" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.858017 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdff4564bab55db2af38192a28710bd2c95c8abfb451c8db33b363c73fa52270"} err="failed to get container status \"fdff4564bab55db2af38192a28710bd2c95c8abfb451c8db33b363c73fa52270\": rpc error: code = NotFound desc = could not find container \"fdff4564bab55db2af38192a28710bd2c95c8abfb451c8db33b363c73fa52270\": container with ID starting with fdff4564bab55db2af38192a28710bd2c95c8abfb451c8db33b363c73fa52270 not found: ID does not exist" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.858030 4997 scope.go:117] "RemoveContainer" containerID="67362fe303ad19856160a40743ba3d5c2e4efe4c0269e64ad5a7763a0de865eb" Dec 05 07:20:56 crc kubenswrapper[4997]: E1205 07:20:56.858272 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67362fe303ad19856160a40743ba3d5c2e4efe4c0269e64ad5a7763a0de865eb\": container with ID starting with 67362fe303ad19856160a40743ba3d5c2e4efe4c0269e64ad5a7763a0de865eb not found: ID does not exist" containerID="67362fe303ad19856160a40743ba3d5c2e4efe4c0269e64ad5a7763a0de865eb" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.858288 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67362fe303ad19856160a40743ba3d5c2e4efe4c0269e64ad5a7763a0de865eb"} err="failed to get container status \"67362fe303ad19856160a40743ba3d5c2e4efe4c0269e64ad5a7763a0de865eb\": rpc error: code = NotFound desc = could not find container \"67362fe303ad19856160a40743ba3d5c2e4efe4c0269e64ad5a7763a0de865eb\": container with ID starting with 67362fe303ad19856160a40743ba3d5c2e4efe4c0269e64ad5a7763a0de865eb not found: ID does not exist" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.858300 4997 scope.go:117] "RemoveContainer" containerID="52f17ac84e23b345d3bafec1cf2403e818e1a9f66defc8fbae156bb75206d250" Dec 05 07:20:56 crc kubenswrapper[4997]: E1205 07:20:56.858546 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52f17ac84e23b345d3bafec1cf2403e818e1a9f66defc8fbae156bb75206d250\": container with ID starting with 52f17ac84e23b345d3bafec1cf2403e818e1a9f66defc8fbae156bb75206d250 not found: ID does not exist" containerID="52f17ac84e23b345d3bafec1cf2403e818e1a9f66defc8fbae156bb75206d250" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.858562 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52f17ac84e23b345d3bafec1cf2403e818e1a9f66defc8fbae156bb75206d250"} err="failed to get container status \"52f17ac84e23b345d3bafec1cf2403e818e1a9f66defc8fbae156bb75206d250\": rpc error: code = NotFound desc = could not find container \"52f17ac84e23b345d3bafec1cf2403e818e1a9f66defc8fbae156bb75206d250\": container with ID starting with 52f17ac84e23b345d3bafec1cf2403e818e1a9f66defc8fbae156bb75206d250 not found: ID does not exist" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.917705 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2b7a1c8e-7026-4860-9ab9-0f66cc187559-log-httpd\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.918194 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.918256 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.919297 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-scripts\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.920057 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8jg8\" (UniqueName: \"kubernetes.io/projected/2b7a1c8e-7026-4860-9ab9-0f66cc187559-kube-api-access-q8jg8\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.920877 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2b7a1c8e-7026-4860-9ab9-0f66cc187559-run-httpd\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.920918 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-config-data\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:56 crc kubenswrapper[4997]: I1205 07:20:56.992819 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 05 07:20:57 crc kubenswrapper[4997]: I1205 07:20:57.026095 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:57 crc kubenswrapper[4997]: I1205 07:20:57.026195 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:57 crc kubenswrapper[4997]: I1205 07:20:57.026225 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-scripts\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:57 crc kubenswrapper[4997]: I1205 07:20:57.026271 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8jg8\" (UniqueName: \"kubernetes.io/projected/2b7a1c8e-7026-4860-9ab9-0f66cc187559-kube-api-access-q8jg8\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:57 crc kubenswrapper[4997]: I1205 07:20:57.026313 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2b7a1c8e-7026-4860-9ab9-0f66cc187559-run-httpd\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:57 crc kubenswrapper[4997]: I1205 07:20:57.026334 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-config-data\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:57 crc kubenswrapper[4997]: I1205 07:20:57.026382 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2b7a1c8e-7026-4860-9ab9-0f66cc187559-log-httpd\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:57 crc kubenswrapper[4997]: I1205 07:20:57.027044 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2b7a1c8e-7026-4860-9ab9-0f66cc187559-log-httpd\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:57 crc kubenswrapper[4997]: I1205 07:20:57.027756 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2b7a1c8e-7026-4860-9ab9-0f66cc187559-run-httpd\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:57 crc kubenswrapper[4997]: I1205 07:20:57.030772 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:57 crc kubenswrapper[4997]: I1205 07:20:57.034529 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-config-data\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:57 crc kubenswrapper[4997]: I1205 07:20:57.042460 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:57 crc kubenswrapper[4997]: I1205 07:20:57.050683 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-scripts\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:57 crc kubenswrapper[4997]: I1205 07:20:57.055043 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8jg8\" (UniqueName: \"kubernetes.io/projected/2b7a1c8e-7026-4860-9ab9-0f66cc187559-kube-api-access-q8jg8\") pod \"ceilometer-0\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " pod="openstack/ceilometer-0" Dec 05 07:20:57 crc kubenswrapper[4997]: I1205 07:20:57.089318 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:20:57 crc kubenswrapper[4997]: I1205 07:20:57.670285 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:20:57 crc kubenswrapper[4997]: W1205 07:20:57.684063 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b7a1c8e_7026_4860_9ab9_0f66cc187559.slice/crio-f769606626428f7693ced509c300ba94f2fcfa44d6724b2861c2742206928d2c WatchSource:0}: Error finding container f769606626428f7693ced509c300ba94f2fcfa44d6724b2861c2742206928d2c: Status 404 returned error can't find the container with id f769606626428f7693ced509c300ba94f2fcfa44d6724b2861c2742206928d2c Dec 05 07:20:57 crc kubenswrapper[4997]: I1205 07:20:57.764648 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bd457ad-f0e3-4de8-9bf4-8206c4cd661f" path="/var/lib/kubelet/pods/1bd457ad-f0e3-4de8-9bf4-8206c4cd661f/volumes" Dec 05 07:20:58 crc kubenswrapper[4997]: I1205 07:20:58.706114 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2b7a1c8e-7026-4860-9ab9-0f66cc187559","Type":"ContainerStarted","Data":"ac59ac472762c63070b59bce0b71ce6e17b3f41b388a25c53f2bd606a010ee10"} Dec 05 07:20:58 crc kubenswrapper[4997]: I1205 07:20:58.706646 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2b7a1c8e-7026-4860-9ab9-0f66cc187559","Type":"ContainerStarted","Data":"f769606626428f7693ced509c300ba94f2fcfa44d6724b2861c2742206928d2c"} Dec 05 07:20:58 crc kubenswrapper[4997]: I1205 07:20:58.976026 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:21:01 crc kubenswrapper[4997]: I1205 07:21:01.142272 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 07:21:01 crc kubenswrapper[4997]: I1205 07:21:01.142867 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 07:21:01 crc kubenswrapper[4997]: I1205 07:21:01.187274 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 07:21:01 crc kubenswrapper[4997]: I1205 07:21:01.206249 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 07:21:01 crc kubenswrapper[4997]: I1205 07:21:01.747072 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 07:21:01 crc kubenswrapper[4997]: I1205 07:21:01.747150 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 07:21:02 crc kubenswrapper[4997]: I1205 07:21:02.993439 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 07:21:02 crc kubenswrapper[4997]: I1205 07:21:02.993835 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 07:21:03 crc kubenswrapper[4997]: I1205 07:21:03.040595 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 07:21:03 crc kubenswrapper[4997]: I1205 07:21:03.069824 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 07:21:03 crc kubenswrapper[4997]: I1205 07:21:03.776259 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 07:21:03 crc kubenswrapper[4997]: I1205 07:21:03.776716 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 07:21:03 crc kubenswrapper[4997]: I1205 07:21:03.934809 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-txfw9"] Dec 05 07:21:03 crc kubenswrapper[4997]: I1205 07:21:03.937106 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-txfw9" Dec 05 07:21:03 crc kubenswrapper[4997]: I1205 07:21:03.972735 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-txfw9"] Dec 05 07:21:04 crc kubenswrapper[4997]: I1205 07:21:04.020093 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/517be6bb-125b-4649-a4a1-764b628c1f25-utilities\") pod \"redhat-operators-txfw9\" (UID: \"517be6bb-125b-4649-a4a1-764b628c1f25\") " pod="openshift-marketplace/redhat-operators-txfw9" Dec 05 07:21:04 crc kubenswrapper[4997]: I1205 07:21:04.020210 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9zv7\" (UniqueName: \"kubernetes.io/projected/517be6bb-125b-4649-a4a1-764b628c1f25-kube-api-access-d9zv7\") pod \"redhat-operators-txfw9\" (UID: \"517be6bb-125b-4649-a4a1-764b628c1f25\") " pod="openshift-marketplace/redhat-operators-txfw9" Dec 05 07:21:04 crc kubenswrapper[4997]: I1205 07:21:04.020307 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/517be6bb-125b-4649-a4a1-764b628c1f25-catalog-content\") pod \"redhat-operators-txfw9\" (UID: \"517be6bb-125b-4649-a4a1-764b628c1f25\") " pod="openshift-marketplace/redhat-operators-txfw9" Dec 05 07:21:04 crc kubenswrapper[4997]: I1205 07:21:04.065591 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 07:21:04 crc kubenswrapper[4997]: I1205 07:21:04.065768 4997 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 07:21:04 crc kubenswrapper[4997]: I1205 07:21:04.069466 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 07:21:04 crc kubenswrapper[4997]: I1205 07:21:04.123248 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/517be6bb-125b-4649-a4a1-764b628c1f25-catalog-content\") pod \"redhat-operators-txfw9\" (UID: \"517be6bb-125b-4649-a4a1-764b628c1f25\") " pod="openshift-marketplace/redhat-operators-txfw9" Dec 05 07:21:04 crc kubenswrapper[4997]: I1205 07:21:04.123428 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/517be6bb-125b-4649-a4a1-764b628c1f25-utilities\") pod \"redhat-operators-txfw9\" (UID: \"517be6bb-125b-4649-a4a1-764b628c1f25\") " pod="openshift-marketplace/redhat-operators-txfw9" Dec 05 07:21:04 crc kubenswrapper[4997]: I1205 07:21:04.123475 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9zv7\" (UniqueName: \"kubernetes.io/projected/517be6bb-125b-4649-a4a1-764b628c1f25-kube-api-access-d9zv7\") pod \"redhat-operators-txfw9\" (UID: \"517be6bb-125b-4649-a4a1-764b628c1f25\") " pod="openshift-marketplace/redhat-operators-txfw9" Dec 05 07:21:04 crc kubenswrapper[4997]: I1205 07:21:04.124654 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/517be6bb-125b-4649-a4a1-764b628c1f25-catalog-content\") pod \"redhat-operators-txfw9\" (UID: \"517be6bb-125b-4649-a4a1-764b628c1f25\") " pod="openshift-marketplace/redhat-operators-txfw9" Dec 05 07:21:04 crc kubenswrapper[4997]: I1205 07:21:04.125023 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/517be6bb-125b-4649-a4a1-764b628c1f25-utilities\") pod \"redhat-operators-txfw9\" (UID: \"517be6bb-125b-4649-a4a1-764b628c1f25\") " pod="openshift-marketplace/redhat-operators-txfw9" Dec 05 07:21:04 crc kubenswrapper[4997]: I1205 07:21:04.162007 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9zv7\" (UniqueName: \"kubernetes.io/projected/517be6bb-125b-4649-a4a1-764b628c1f25-kube-api-access-d9zv7\") pod \"redhat-operators-txfw9\" (UID: \"517be6bb-125b-4649-a4a1-764b628c1f25\") " pod="openshift-marketplace/redhat-operators-txfw9" Dec 05 07:21:04 crc kubenswrapper[4997]: I1205 07:21:04.265240 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-txfw9" Dec 05 07:21:05 crc kubenswrapper[4997]: I1205 07:21:05.918248 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-txfw9"] Dec 05 07:21:06 crc kubenswrapper[4997]: E1205 07:21:06.308817 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod517be6bb_125b_4649_a4a1_764b628c1f25.slice/crio-conmon-6680aacd156af62c8359f41f4b22b8cc1b010dc7eb05d3bbe0ba8985d562b989.scope\": RecentStats: unable to find data in memory cache]" Dec 05 07:21:06 crc kubenswrapper[4997]: I1205 07:21:06.428278 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 07:21:06 crc kubenswrapper[4997]: I1205 07:21:06.428425 4997 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 07:21:06 crc kubenswrapper[4997]: I1205 07:21:06.625928 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 07:21:06 crc kubenswrapper[4997]: I1205 07:21:06.820950 4997 generic.go:334] "Generic (PLEG): container finished" podID="517be6bb-125b-4649-a4a1-764b628c1f25" containerID="6680aacd156af62c8359f41f4b22b8cc1b010dc7eb05d3bbe0ba8985d562b989" exitCode=0 Dec 05 07:21:06 crc kubenswrapper[4997]: I1205 07:21:06.821486 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txfw9" event={"ID":"517be6bb-125b-4649-a4a1-764b628c1f25","Type":"ContainerDied","Data":"6680aacd156af62c8359f41f4b22b8cc1b010dc7eb05d3bbe0ba8985d562b989"} Dec 05 07:21:06 crc kubenswrapper[4997]: I1205 07:21:06.821529 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txfw9" event={"ID":"517be6bb-125b-4649-a4a1-764b628c1f25","Type":"ContainerStarted","Data":"1017cd5b4dbabdf101e8ca16cd7aa92834815c9ba38a2e91f34c549938b01fb7"} Dec 05 07:21:06 crc kubenswrapper[4997]: I1205 07:21:06.828008 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-qzrvc" event={"ID":"778a1f81-32bc-4c34-b77d-3675d24d5467","Type":"ContainerStarted","Data":"3f78e02557266ee9488a60e7e7d398025ad709b887c5b297cbc3363b33225525"} Dec 05 07:21:06 crc kubenswrapper[4997]: I1205 07:21:06.839470 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2b7a1c8e-7026-4860-9ab9-0f66cc187559","Type":"ContainerStarted","Data":"26f28ed47a68cff357f468ae8e1db8ac2a479eb104daa85ba40c09b316f06904"} Dec 05 07:21:06 crc kubenswrapper[4997]: I1205 07:21:06.886677 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-qzrvc" podStartSLOduration=3.305929973 podStartE2EDuration="14.886646783s" podCreationTimestamp="2025-12-05 07:20:52 +0000 UTC" firstStartedPulling="2025-12-05 07:20:53.905200325 +0000 UTC m=+1554.434107586" lastFinishedPulling="2025-12-05 07:21:05.485917135 +0000 UTC m=+1566.014824396" observedRunningTime="2025-12-05 07:21:06.88060249 +0000 UTC m=+1567.409509751" watchObservedRunningTime="2025-12-05 07:21:06.886646783 +0000 UTC m=+1567.415554044" Dec 05 07:21:07 crc kubenswrapper[4997]: I1205 07:21:07.853488 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txfw9" event={"ID":"517be6bb-125b-4649-a4a1-764b628c1f25","Type":"ContainerStarted","Data":"233f9f67946a8d324702ffea0d5561d6cb11f83092f9441f3eef275bb113d6bd"} Dec 05 07:21:07 crc kubenswrapper[4997]: I1205 07:21:07.860159 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2b7a1c8e-7026-4860-9ab9-0f66cc187559","Type":"ContainerStarted","Data":"c81fec47b7d1f800bb7239bb0f25c51ec857bc80d34743f312b791b92369ad13"} Dec 05 07:21:08 crc kubenswrapper[4997]: I1205 07:21:08.874841 4997 generic.go:334] "Generic (PLEG): container finished" podID="517be6bb-125b-4649-a4a1-764b628c1f25" containerID="233f9f67946a8d324702ffea0d5561d6cb11f83092f9441f3eef275bb113d6bd" exitCode=0 Dec 05 07:21:08 crc kubenswrapper[4997]: I1205 07:21:08.874962 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txfw9" event={"ID":"517be6bb-125b-4649-a4a1-764b628c1f25","Type":"ContainerDied","Data":"233f9f67946a8d324702ffea0d5561d6cb11f83092f9441f3eef275bb113d6bd"} Dec 05 07:21:09 crc kubenswrapper[4997]: I1205 07:21:09.890029 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2b7a1c8e-7026-4860-9ab9-0f66cc187559","Type":"ContainerStarted","Data":"04962c1512e9695617fdd16a154e03c3004c5cdbbec285021d959c5476bf76fc"} Dec 05 07:21:09 crc kubenswrapper[4997]: I1205 07:21:09.890559 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 07:21:09 crc kubenswrapper[4997]: I1205 07:21:09.890256 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerName="ceilometer-notification-agent" containerID="cri-o://26f28ed47a68cff357f468ae8e1db8ac2a479eb104daa85ba40c09b316f06904" gracePeriod=30 Dec 05 07:21:09 crc kubenswrapper[4997]: I1205 07:21:09.890186 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerName="ceilometer-central-agent" containerID="cri-o://ac59ac472762c63070b59bce0b71ce6e17b3f41b388a25c53f2bd606a010ee10" gracePeriod=30 Dec 05 07:21:09 crc kubenswrapper[4997]: I1205 07:21:09.890257 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerName="sg-core" containerID="cri-o://c81fec47b7d1f800bb7239bb0f25c51ec857bc80d34743f312b791b92369ad13" gracePeriod=30 Dec 05 07:21:09 crc kubenswrapper[4997]: I1205 07:21:09.890323 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerName="proxy-httpd" containerID="cri-o://04962c1512e9695617fdd16a154e03c3004c5cdbbec285021d959c5476bf76fc" gracePeriod=30 Dec 05 07:21:09 crc kubenswrapper[4997]: I1205 07:21:09.916136 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.08362458 podStartE2EDuration="13.916116903s" podCreationTimestamp="2025-12-05 07:20:56 +0000 UTC" firstStartedPulling="2025-12-05 07:20:57.686742502 +0000 UTC m=+1558.215649763" lastFinishedPulling="2025-12-05 07:21:09.519234825 +0000 UTC m=+1570.048142086" observedRunningTime="2025-12-05 07:21:09.914242061 +0000 UTC m=+1570.443149322" watchObservedRunningTime="2025-12-05 07:21:09.916116903 +0000 UTC m=+1570.445024164" Dec 05 07:21:10 crc kubenswrapper[4997]: I1205 07:21:10.913099 4997 generic.go:334] "Generic (PLEG): container finished" podID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerID="c81fec47b7d1f800bb7239bb0f25c51ec857bc80d34743f312b791b92369ad13" exitCode=2 Dec 05 07:21:10 crc kubenswrapper[4997]: I1205 07:21:10.913550 4997 generic.go:334] "Generic (PLEG): container finished" podID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerID="26f28ed47a68cff357f468ae8e1db8ac2a479eb104daa85ba40c09b316f06904" exitCode=0 Dec 05 07:21:10 crc kubenswrapper[4997]: I1205 07:21:10.913563 4997 generic.go:334] "Generic (PLEG): container finished" podID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerID="ac59ac472762c63070b59bce0b71ce6e17b3f41b388a25c53f2bd606a010ee10" exitCode=0 Dec 05 07:21:10 crc kubenswrapper[4997]: I1205 07:21:10.913194 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2b7a1c8e-7026-4860-9ab9-0f66cc187559","Type":"ContainerDied","Data":"c81fec47b7d1f800bb7239bb0f25c51ec857bc80d34743f312b791b92369ad13"} Dec 05 07:21:10 crc kubenswrapper[4997]: I1205 07:21:10.913661 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2b7a1c8e-7026-4860-9ab9-0f66cc187559","Type":"ContainerDied","Data":"26f28ed47a68cff357f468ae8e1db8ac2a479eb104daa85ba40c09b316f06904"} Dec 05 07:21:10 crc kubenswrapper[4997]: I1205 07:21:10.913680 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2b7a1c8e-7026-4860-9ab9-0f66cc187559","Type":"ContainerDied","Data":"ac59ac472762c63070b59bce0b71ce6e17b3f41b388a25c53f2bd606a010ee10"} Dec 05 07:21:10 crc kubenswrapper[4997]: I1205 07:21:10.916263 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txfw9" event={"ID":"517be6bb-125b-4649-a4a1-764b628c1f25","Type":"ContainerStarted","Data":"3739f7c6336aa11acde4ea7624bd022a7df58589f6491a955cd7be2ce2cf3ebc"} Dec 05 07:21:10 crc kubenswrapper[4997]: I1205 07:21:10.942186 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-txfw9" podStartSLOduration=5.134073377 podStartE2EDuration="7.942157073s" podCreationTimestamp="2025-12-05 07:21:03 +0000 UTC" firstStartedPulling="2025-12-05 07:21:06.823712492 +0000 UTC m=+1567.352619743" lastFinishedPulling="2025-12-05 07:21:09.631796178 +0000 UTC m=+1570.160703439" observedRunningTime="2025-12-05 07:21:10.933180991 +0000 UTC m=+1571.462088252" watchObservedRunningTime="2025-12-05 07:21:10.942157073 +0000 UTC m=+1571.471064334" Dec 05 07:21:14 crc kubenswrapper[4997]: I1205 07:21:14.266470 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-txfw9" Dec 05 07:21:14 crc kubenswrapper[4997]: I1205 07:21:14.266957 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-txfw9" Dec 05 07:21:15 crc kubenswrapper[4997]: I1205 07:21:15.328228 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-txfw9" podUID="517be6bb-125b-4649-a4a1-764b628c1f25" containerName="registry-server" probeResult="failure" output=< Dec 05 07:21:15 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 07:21:15 crc kubenswrapper[4997]: > Dec 05 07:21:20 crc kubenswrapper[4997]: I1205 07:21:20.044555 4997 generic.go:334] "Generic (PLEG): container finished" podID="778a1f81-32bc-4c34-b77d-3675d24d5467" containerID="3f78e02557266ee9488a60e7e7d398025ad709b887c5b297cbc3363b33225525" exitCode=0 Dec 05 07:21:20 crc kubenswrapper[4997]: I1205 07:21:20.044662 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-qzrvc" event={"ID":"778a1f81-32bc-4c34-b77d-3675d24d5467","Type":"ContainerDied","Data":"3f78e02557266ee9488a60e7e7d398025ad709b887c5b297cbc3363b33225525"} Dec 05 07:21:21 crc kubenswrapper[4997]: I1205 07:21:21.483042 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-qzrvc" Dec 05 07:21:21 crc kubenswrapper[4997]: I1205 07:21:21.565257 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-config-data\") pod \"778a1f81-32bc-4c34-b77d-3675d24d5467\" (UID: \"778a1f81-32bc-4c34-b77d-3675d24d5467\") " Dec 05 07:21:21 crc kubenswrapper[4997]: I1205 07:21:21.565325 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-combined-ca-bundle\") pod \"778a1f81-32bc-4c34-b77d-3675d24d5467\" (UID: \"778a1f81-32bc-4c34-b77d-3675d24d5467\") " Dec 05 07:21:21 crc kubenswrapper[4997]: I1205 07:21:21.565371 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vjb2\" (UniqueName: \"kubernetes.io/projected/778a1f81-32bc-4c34-b77d-3675d24d5467-kube-api-access-5vjb2\") pod \"778a1f81-32bc-4c34-b77d-3675d24d5467\" (UID: \"778a1f81-32bc-4c34-b77d-3675d24d5467\") " Dec 05 07:21:21 crc kubenswrapper[4997]: I1205 07:21:21.565527 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-scripts\") pod \"778a1f81-32bc-4c34-b77d-3675d24d5467\" (UID: \"778a1f81-32bc-4c34-b77d-3675d24d5467\") " Dec 05 07:21:21 crc kubenswrapper[4997]: I1205 07:21:21.573920 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/778a1f81-32bc-4c34-b77d-3675d24d5467-kube-api-access-5vjb2" (OuterVolumeSpecName: "kube-api-access-5vjb2") pod "778a1f81-32bc-4c34-b77d-3675d24d5467" (UID: "778a1f81-32bc-4c34-b77d-3675d24d5467"). InnerVolumeSpecName "kube-api-access-5vjb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:21:21 crc kubenswrapper[4997]: I1205 07:21:21.578850 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-scripts" (OuterVolumeSpecName: "scripts") pod "778a1f81-32bc-4c34-b77d-3675d24d5467" (UID: "778a1f81-32bc-4c34-b77d-3675d24d5467"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:21 crc kubenswrapper[4997]: I1205 07:21:21.597136 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-config-data" (OuterVolumeSpecName: "config-data") pod "778a1f81-32bc-4c34-b77d-3675d24d5467" (UID: "778a1f81-32bc-4c34-b77d-3675d24d5467"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:21 crc kubenswrapper[4997]: I1205 07:21:21.599665 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "778a1f81-32bc-4c34-b77d-3675d24d5467" (UID: "778a1f81-32bc-4c34-b77d-3675d24d5467"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:21 crc kubenswrapper[4997]: I1205 07:21:21.669403 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:21 crc kubenswrapper[4997]: I1205 07:21:21.669478 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:21 crc kubenswrapper[4997]: I1205 07:21:21.669498 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778a1f81-32bc-4c34-b77d-3675d24d5467-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:21 crc kubenswrapper[4997]: I1205 07:21:21.669519 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vjb2\" (UniqueName: \"kubernetes.io/projected/778a1f81-32bc-4c34-b77d-3675d24d5467-kube-api-access-5vjb2\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.069057 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-qzrvc" event={"ID":"778a1f81-32bc-4c34-b77d-3675d24d5467","Type":"ContainerDied","Data":"31601a6b0af246491220bd2c26af1630dbd55e6abc42e0acb34fbca7b310b7ff"} Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.069511 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31601a6b0af246491220bd2c26af1630dbd55e6abc42e0acb34fbca7b310b7ff" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.069356 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-qzrvc" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.177805 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 07:21:22 crc kubenswrapper[4997]: E1205 07:21:22.178332 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="778a1f81-32bc-4c34-b77d-3675d24d5467" containerName="nova-cell0-conductor-db-sync" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.178357 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="778a1f81-32bc-4c34-b77d-3675d24d5467" containerName="nova-cell0-conductor-db-sync" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.178679 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="778a1f81-32bc-4c34-b77d-3675d24d5467" containerName="nova-cell0-conductor-db-sync" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.179588 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.182693 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-qwgjd" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.182736 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.191262 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.285326 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c706bba7-965c-497f-ae7d-b7087d37d70e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c706bba7-965c-497f-ae7d-b7087d37d70e\") " pod="openstack/nova-cell0-conductor-0" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.285562 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-929vr\" (UniqueName: \"kubernetes.io/projected/c706bba7-965c-497f-ae7d-b7087d37d70e-kube-api-access-929vr\") pod \"nova-cell0-conductor-0\" (UID: \"c706bba7-965c-497f-ae7d-b7087d37d70e\") " pod="openstack/nova-cell0-conductor-0" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.285877 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c706bba7-965c-497f-ae7d-b7087d37d70e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c706bba7-965c-497f-ae7d-b7087d37d70e\") " pod="openstack/nova-cell0-conductor-0" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.388481 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c706bba7-965c-497f-ae7d-b7087d37d70e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c706bba7-965c-497f-ae7d-b7087d37d70e\") " pod="openstack/nova-cell0-conductor-0" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.388781 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c706bba7-965c-497f-ae7d-b7087d37d70e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c706bba7-965c-497f-ae7d-b7087d37d70e\") " pod="openstack/nova-cell0-conductor-0" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.388967 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-929vr\" (UniqueName: \"kubernetes.io/projected/c706bba7-965c-497f-ae7d-b7087d37d70e-kube-api-access-929vr\") pod \"nova-cell0-conductor-0\" (UID: \"c706bba7-965c-497f-ae7d-b7087d37d70e\") " pod="openstack/nova-cell0-conductor-0" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.394931 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c706bba7-965c-497f-ae7d-b7087d37d70e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c706bba7-965c-497f-ae7d-b7087d37d70e\") " pod="openstack/nova-cell0-conductor-0" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.400422 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c706bba7-965c-497f-ae7d-b7087d37d70e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c706bba7-965c-497f-ae7d-b7087d37d70e\") " pod="openstack/nova-cell0-conductor-0" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.419591 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-929vr\" (UniqueName: \"kubernetes.io/projected/c706bba7-965c-497f-ae7d-b7087d37d70e-kube-api-access-929vr\") pod \"nova-cell0-conductor-0\" (UID: \"c706bba7-965c-497f-ae7d-b7087d37d70e\") " pod="openstack/nova-cell0-conductor-0" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.500748 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 07:21:22 crc kubenswrapper[4997]: I1205 07:21:22.979510 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 07:21:22 crc kubenswrapper[4997]: W1205 07:21:22.983895 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc706bba7_965c_497f_ae7d_b7087d37d70e.slice/crio-94189baad8432517afd7770bd384f234f9e285019dcb23f7f78b04388f24dadd WatchSource:0}: Error finding container 94189baad8432517afd7770bd384f234f9e285019dcb23f7f78b04388f24dadd: Status 404 returned error can't find the container with id 94189baad8432517afd7770bd384f234f9e285019dcb23f7f78b04388f24dadd Dec 05 07:21:23 crc kubenswrapper[4997]: I1205 07:21:23.082346 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c706bba7-965c-497f-ae7d-b7087d37d70e","Type":"ContainerStarted","Data":"94189baad8432517afd7770bd384f234f9e285019dcb23f7f78b04388f24dadd"} Dec 05 07:21:24 crc kubenswrapper[4997]: I1205 07:21:24.100189 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c706bba7-965c-497f-ae7d-b7087d37d70e","Type":"ContainerStarted","Data":"b81d606450f4ec2753efb3ba9b01b31144b20d257dc2ed0c1a6d4b204cf562b4"} Dec 05 07:21:24 crc kubenswrapper[4997]: I1205 07:21:24.101559 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 05 07:21:24 crc kubenswrapper[4997]: I1205 07:21:24.346075 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-txfw9" Dec 05 07:21:24 crc kubenswrapper[4997]: I1205 07:21:24.378057 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.378012922 podStartE2EDuration="2.378012922s" podCreationTimestamp="2025-12-05 07:21:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:21:24.137686347 +0000 UTC m=+1584.666593698" watchObservedRunningTime="2025-12-05 07:21:24.378012922 +0000 UTC m=+1584.906920193" Dec 05 07:21:24 crc kubenswrapper[4997]: I1205 07:21:24.399305 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-txfw9" Dec 05 07:21:24 crc kubenswrapper[4997]: I1205 07:21:24.593789 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-txfw9"] Dec 05 07:21:26 crc kubenswrapper[4997]: I1205 07:21:26.125052 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-txfw9" podUID="517be6bb-125b-4649-a4a1-764b628c1f25" containerName="registry-server" containerID="cri-o://3739f7c6336aa11acde4ea7624bd022a7df58589f6491a955cd7be2ce2cf3ebc" gracePeriod=2 Dec 05 07:21:26 crc kubenswrapper[4997]: I1205 07:21:26.690197 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-txfw9" Dec 05 07:21:26 crc kubenswrapper[4997]: I1205 07:21:26.803790 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/517be6bb-125b-4649-a4a1-764b628c1f25-catalog-content\") pod \"517be6bb-125b-4649-a4a1-764b628c1f25\" (UID: \"517be6bb-125b-4649-a4a1-764b628c1f25\") " Dec 05 07:21:26 crc kubenswrapper[4997]: I1205 07:21:26.803958 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9zv7\" (UniqueName: \"kubernetes.io/projected/517be6bb-125b-4649-a4a1-764b628c1f25-kube-api-access-d9zv7\") pod \"517be6bb-125b-4649-a4a1-764b628c1f25\" (UID: \"517be6bb-125b-4649-a4a1-764b628c1f25\") " Dec 05 07:21:26 crc kubenswrapper[4997]: I1205 07:21:26.804011 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/517be6bb-125b-4649-a4a1-764b628c1f25-utilities\") pod \"517be6bb-125b-4649-a4a1-764b628c1f25\" (UID: \"517be6bb-125b-4649-a4a1-764b628c1f25\") " Dec 05 07:21:26 crc kubenswrapper[4997]: I1205 07:21:26.805370 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/517be6bb-125b-4649-a4a1-764b628c1f25-utilities" (OuterVolumeSpecName: "utilities") pod "517be6bb-125b-4649-a4a1-764b628c1f25" (UID: "517be6bb-125b-4649-a4a1-764b628c1f25"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:21:26 crc kubenswrapper[4997]: I1205 07:21:26.813225 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/517be6bb-125b-4649-a4a1-764b628c1f25-kube-api-access-d9zv7" (OuterVolumeSpecName: "kube-api-access-d9zv7") pod "517be6bb-125b-4649-a4a1-764b628c1f25" (UID: "517be6bb-125b-4649-a4a1-764b628c1f25"). InnerVolumeSpecName "kube-api-access-d9zv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:21:26 crc kubenswrapper[4997]: I1205 07:21:26.907424 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9zv7\" (UniqueName: \"kubernetes.io/projected/517be6bb-125b-4649-a4a1-764b628c1f25-kube-api-access-d9zv7\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:26 crc kubenswrapper[4997]: I1205 07:21:26.907463 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/517be6bb-125b-4649-a4a1-764b628c1f25-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:26 crc kubenswrapper[4997]: I1205 07:21:26.913351 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/517be6bb-125b-4649-a4a1-764b628c1f25-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "517be6bb-125b-4649-a4a1-764b628c1f25" (UID: "517be6bb-125b-4649-a4a1-764b628c1f25"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.009822 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/517be6bb-125b-4649-a4a1-764b628c1f25-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.096950 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.140321 4997 generic.go:334] "Generic (PLEG): container finished" podID="517be6bb-125b-4649-a4a1-764b628c1f25" containerID="3739f7c6336aa11acde4ea7624bd022a7df58589f6491a955cd7be2ce2cf3ebc" exitCode=0 Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.140393 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txfw9" event={"ID":"517be6bb-125b-4649-a4a1-764b628c1f25","Type":"ContainerDied","Data":"3739f7c6336aa11acde4ea7624bd022a7df58589f6491a955cd7be2ce2cf3ebc"} Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.140463 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txfw9" event={"ID":"517be6bb-125b-4649-a4a1-764b628c1f25","Type":"ContainerDied","Data":"1017cd5b4dbabdf101e8ca16cd7aa92834815c9ba38a2e91f34c549938b01fb7"} Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.140494 4997 scope.go:117] "RemoveContainer" containerID="3739f7c6336aa11acde4ea7624bd022a7df58589f6491a955cd7be2ce2cf3ebc" Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.140761 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-txfw9" Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.189497 4997 scope.go:117] "RemoveContainer" containerID="233f9f67946a8d324702ffea0d5561d6cb11f83092f9441f3eef275bb113d6bd" Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.194045 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-txfw9"] Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.210295 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-txfw9"] Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.226244 4997 scope.go:117] "RemoveContainer" containerID="6680aacd156af62c8359f41f4b22b8cc1b010dc7eb05d3bbe0ba8985d562b989" Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.266410 4997 scope.go:117] "RemoveContainer" containerID="3739f7c6336aa11acde4ea7624bd022a7df58589f6491a955cd7be2ce2cf3ebc" Dec 05 07:21:27 crc kubenswrapper[4997]: E1205 07:21:27.267203 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3739f7c6336aa11acde4ea7624bd022a7df58589f6491a955cd7be2ce2cf3ebc\": container with ID starting with 3739f7c6336aa11acde4ea7624bd022a7df58589f6491a955cd7be2ce2cf3ebc not found: ID does not exist" containerID="3739f7c6336aa11acde4ea7624bd022a7df58589f6491a955cd7be2ce2cf3ebc" Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.267238 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3739f7c6336aa11acde4ea7624bd022a7df58589f6491a955cd7be2ce2cf3ebc"} err="failed to get container status \"3739f7c6336aa11acde4ea7624bd022a7df58589f6491a955cd7be2ce2cf3ebc\": rpc error: code = NotFound desc = could not find container \"3739f7c6336aa11acde4ea7624bd022a7df58589f6491a955cd7be2ce2cf3ebc\": container with ID starting with 3739f7c6336aa11acde4ea7624bd022a7df58589f6491a955cd7be2ce2cf3ebc not found: ID does not exist" Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.267268 4997 scope.go:117] "RemoveContainer" containerID="233f9f67946a8d324702ffea0d5561d6cb11f83092f9441f3eef275bb113d6bd" Dec 05 07:21:27 crc kubenswrapper[4997]: E1205 07:21:27.269493 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"233f9f67946a8d324702ffea0d5561d6cb11f83092f9441f3eef275bb113d6bd\": container with ID starting with 233f9f67946a8d324702ffea0d5561d6cb11f83092f9441f3eef275bb113d6bd not found: ID does not exist" containerID="233f9f67946a8d324702ffea0d5561d6cb11f83092f9441f3eef275bb113d6bd" Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.269530 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"233f9f67946a8d324702ffea0d5561d6cb11f83092f9441f3eef275bb113d6bd"} err="failed to get container status \"233f9f67946a8d324702ffea0d5561d6cb11f83092f9441f3eef275bb113d6bd\": rpc error: code = NotFound desc = could not find container \"233f9f67946a8d324702ffea0d5561d6cb11f83092f9441f3eef275bb113d6bd\": container with ID starting with 233f9f67946a8d324702ffea0d5561d6cb11f83092f9441f3eef275bb113d6bd not found: ID does not exist" Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.269552 4997 scope.go:117] "RemoveContainer" containerID="6680aacd156af62c8359f41f4b22b8cc1b010dc7eb05d3bbe0ba8985d562b989" Dec 05 07:21:27 crc kubenswrapper[4997]: E1205 07:21:27.270330 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6680aacd156af62c8359f41f4b22b8cc1b010dc7eb05d3bbe0ba8985d562b989\": container with ID starting with 6680aacd156af62c8359f41f4b22b8cc1b010dc7eb05d3bbe0ba8985d562b989 not found: ID does not exist" containerID="6680aacd156af62c8359f41f4b22b8cc1b010dc7eb05d3bbe0ba8985d562b989" Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.270377 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6680aacd156af62c8359f41f4b22b8cc1b010dc7eb05d3bbe0ba8985d562b989"} err="failed to get container status \"6680aacd156af62c8359f41f4b22b8cc1b010dc7eb05d3bbe0ba8985d562b989\": rpc error: code = NotFound desc = could not find container \"6680aacd156af62c8359f41f4b22b8cc1b010dc7eb05d3bbe0ba8985d562b989\": container with ID starting with 6680aacd156af62c8359f41f4b22b8cc1b010dc7eb05d3bbe0ba8985d562b989 not found: ID does not exist" Dec 05 07:21:27 crc kubenswrapper[4997]: I1205 07:21:27.764860 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="517be6bb-125b-4649-a4a1-764b628c1f25" path="/var/lib/kubelet/pods/517be6bb-125b-4649-a4a1-764b628c1f25/volumes" Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.011794 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rs4fg"] Dec 05 07:21:30 crc kubenswrapper[4997]: E1205 07:21:30.013097 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="517be6bb-125b-4649-a4a1-764b628c1f25" containerName="extract-utilities" Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.013127 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="517be6bb-125b-4649-a4a1-764b628c1f25" containerName="extract-utilities" Dec 05 07:21:30 crc kubenswrapper[4997]: E1205 07:21:30.013144 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="517be6bb-125b-4649-a4a1-764b628c1f25" containerName="extract-content" Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.013158 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="517be6bb-125b-4649-a4a1-764b628c1f25" containerName="extract-content" Dec 05 07:21:30 crc kubenswrapper[4997]: E1205 07:21:30.013196 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="517be6bb-125b-4649-a4a1-764b628c1f25" containerName="registry-server" Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.013210 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="517be6bb-125b-4649-a4a1-764b628c1f25" containerName="registry-server" Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.013669 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="517be6bb-125b-4649-a4a1-764b628c1f25" containerName="registry-server" Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.016899 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rs4fg" Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.025199 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rs4fg"] Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.075835 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-catalog-content\") pod \"redhat-marketplace-rs4fg\" (UID: \"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3\") " pod="openshift-marketplace/redhat-marketplace-rs4fg" Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.076101 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nlmd\" (UniqueName: \"kubernetes.io/projected/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-kube-api-access-4nlmd\") pod \"redhat-marketplace-rs4fg\" (UID: \"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3\") " pod="openshift-marketplace/redhat-marketplace-rs4fg" Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.076433 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-utilities\") pod \"redhat-marketplace-rs4fg\" (UID: \"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3\") " pod="openshift-marketplace/redhat-marketplace-rs4fg" Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.179525 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-utilities\") pod \"redhat-marketplace-rs4fg\" (UID: \"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3\") " pod="openshift-marketplace/redhat-marketplace-rs4fg" Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.180229 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-utilities\") pod \"redhat-marketplace-rs4fg\" (UID: \"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3\") " pod="openshift-marketplace/redhat-marketplace-rs4fg" Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.180392 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-catalog-content\") pod \"redhat-marketplace-rs4fg\" (UID: \"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3\") " pod="openshift-marketplace/redhat-marketplace-rs4fg" Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.180651 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nlmd\" (UniqueName: \"kubernetes.io/projected/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-kube-api-access-4nlmd\") pod \"redhat-marketplace-rs4fg\" (UID: \"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3\") " pod="openshift-marketplace/redhat-marketplace-rs4fg" Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.180683 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-catalog-content\") pod \"redhat-marketplace-rs4fg\" (UID: \"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3\") " pod="openshift-marketplace/redhat-marketplace-rs4fg" Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.211799 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nlmd\" (UniqueName: \"kubernetes.io/projected/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-kube-api-access-4nlmd\") pod \"redhat-marketplace-rs4fg\" (UID: \"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3\") " pod="openshift-marketplace/redhat-marketplace-rs4fg" Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.360494 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rs4fg" Dec 05 07:21:30 crc kubenswrapper[4997]: I1205 07:21:30.913503 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rs4fg"] Dec 05 07:21:31 crc kubenswrapper[4997]: I1205 07:21:31.219747 4997 generic.go:334] "Generic (PLEG): container finished" podID="69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3" containerID="d87697661c0663232a7cd7d5bb17c32eb364912a6d887de778a20b4d36f8a5b0" exitCode=0 Dec 05 07:21:31 crc kubenswrapper[4997]: I1205 07:21:31.219895 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rs4fg" event={"ID":"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3","Type":"ContainerDied","Data":"d87697661c0663232a7cd7d5bb17c32eb364912a6d887de778a20b4d36f8a5b0"} Dec 05 07:21:31 crc kubenswrapper[4997]: I1205 07:21:31.220300 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rs4fg" event={"ID":"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3","Type":"ContainerStarted","Data":"5f40c1f5c852310c1efdc626d4264912e0769a3b32acbda922fd4e1524f3bc4a"} Dec 05 07:21:32 crc kubenswrapper[4997]: I1205 07:21:32.541712 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.082042 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-w6b8v"] Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.083783 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-w6b8v" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.086874 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.087749 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.098251 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-w6b8v"] Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.151460 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-scripts\") pod \"nova-cell0-cell-mapping-w6b8v\" (UID: \"b2325c5c-6d81-45b1-90ae-d3570c255bea\") " pod="openstack/nova-cell0-cell-mapping-w6b8v" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.151529 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-w6b8v\" (UID: \"b2325c5c-6d81-45b1-90ae-d3570c255bea\") " pod="openstack/nova-cell0-cell-mapping-w6b8v" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.151730 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6bvc\" (UniqueName: \"kubernetes.io/projected/b2325c5c-6d81-45b1-90ae-d3570c255bea-kube-api-access-v6bvc\") pod \"nova-cell0-cell-mapping-w6b8v\" (UID: \"b2325c5c-6d81-45b1-90ae-d3570c255bea\") " pod="openstack/nova-cell0-cell-mapping-w6b8v" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.151828 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-config-data\") pod \"nova-cell0-cell-mapping-w6b8v\" (UID: \"b2325c5c-6d81-45b1-90ae-d3570c255bea\") " pod="openstack/nova-cell0-cell-mapping-w6b8v" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.253910 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6bvc\" (UniqueName: \"kubernetes.io/projected/b2325c5c-6d81-45b1-90ae-d3570c255bea-kube-api-access-v6bvc\") pod \"nova-cell0-cell-mapping-w6b8v\" (UID: \"b2325c5c-6d81-45b1-90ae-d3570c255bea\") " pod="openstack/nova-cell0-cell-mapping-w6b8v" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.254943 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-config-data\") pod \"nova-cell0-cell-mapping-w6b8v\" (UID: \"b2325c5c-6d81-45b1-90ae-d3570c255bea\") " pod="openstack/nova-cell0-cell-mapping-w6b8v" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.255152 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-scripts\") pod \"nova-cell0-cell-mapping-w6b8v\" (UID: \"b2325c5c-6d81-45b1-90ae-d3570c255bea\") " pod="openstack/nova-cell0-cell-mapping-w6b8v" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.255284 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-w6b8v\" (UID: \"b2325c5c-6d81-45b1-90ae-d3570c255bea\") " pod="openstack/nova-cell0-cell-mapping-w6b8v" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.261339 4997 generic.go:334] "Generic (PLEG): container finished" podID="69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3" containerID="d662f99f7d137a1f54c5ab371f2664323c289c917544886b96361647c074c3e1" exitCode=0 Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.263269 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-config-data\") pod \"nova-cell0-cell-mapping-w6b8v\" (UID: \"b2325c5c-6d81-45b1-90ae-d3570c255bea\") " pod="openstack/nova-cell0-cell-mapping-w6b8v" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.263340 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rs4fg" event={"ID":"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3","Type":"ContainerDied","Data":"d662f99f7d137a1f54c5ab371f2664323c289c917544886b96361647c074c3e1"} Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.265334 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-scripts\") pod \"nova-cell0-cell-mapping-w6b8v\" (UID: \"b2325c5c-6d81-45b1-90ae-d3570c255bea\") " pod="openstack/nova-cell0-cell-mapping-w6b8v" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.266242 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-w6b8v\" (UID: \"b2325c5c-6d81-45b1-90ae-d3570c255bea\") " pod="openstack/nova-cell0-cell-mapping-w6b8v" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.284877 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6bvc\" (UniqueName: \"kubernetes.io/projected/b2325c5c-6d81-45b1-90ae-d3570c255bea-kube-api-access-v6bvc\") pod \"nova-cell0-cell-mapping-w6b8v\" (UID: \"b2325c5c-6d81-45b1-90ae-d3570c255bea\") " pod="openstack/nova-cell0-cell-mapping-w6b8v" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.378077 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.382572 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.387372 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.407478 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-w6b8v" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.409398 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.460587 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7de6ebed-df18-4907-ba33-3f4674373b43-logs\") pod \"nova-api-0\" (UID: \"7de6ebed-df18-4907-ba33-3f4674373b43\") " pod="openstack/nova-api-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.460976 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6tsj\" (UniqueName: \"kubernetes.io/projected/7de6ebed-df18-4907-ba33-3f4674373b43-kube-api-access-b6tsj\") pod \"nova-api-0\" (UID: \"7de6ebed-df18-4907-ba33-3f4674373b43\") " pod="openstack/nova-api-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.461236 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7de6ebed-df18-4907-ba33-3f4674373b43-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7de6ebed-df18-4907-ba33-3f4674373b43\") " pod="openstack/nova-api-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.461373 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7de6ebed-df18-4907-ba33-3f4674373b43-config-data\") pod \"nova-api-0\" (UID: \"7de6ebed-df18-4907-ba33-3f4674373b43\") " pod="openstack/nova-api-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.509071 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.511457 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.522804 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.530775 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.570529 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r78t8\" (UniqueName: \"kubernetes.io/projected/198455d4-8878-4c1c-b462-bcc050051e04-kube-api-access-r78t8\") pod \"nova-metadata-0\" (UID: \"198455d4-8878-4c1c-b462-bcc050051e04\") " pod="openstack/nova-metadata-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.570860 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7de6ebed-df18-4907-ba33-3f4674373b43-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7de6ebed-df18-4907-ba33-3f4674373b43\") " pod="openstack/nova-api-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.570896 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/198455d4-8878-4c1c-b462-bcc050051e04-logs\") pod \"nova-metadata-0\" (UID: \"198455d4-8878-4c1c-b462-bcc050051e04\") " pod="openstack/nova-metadata-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.570918 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7de6ebed-df18-4907-ba33-3f4674373b43-config-data\") pod \"nova-api-0\" (UID: \"7de6ebed-df18-4907-ba33-3f4674373b43\") " pod="openstack/nova-api-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.570949 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7de6ebed-df18-4907-ba33-3f4674373b43-logs\") pod \"nova-api-0\" (UID: \"7de6ebed-df18-4907-ba33-3f4674373b43\") " pod="openstack/nova-api-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.570979 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198455d4-8878-4c1c-b462-bcc050051e04-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"198455d4-8878-4c1c-b462-bcc050051e04\") " pod="openstack/nova-metadata-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.571009 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198455d4-8878-4c1c-b462-bcc050051e04-config-data\") pod \"nova-metadata-0\" (UID: \"198455d4-8878-4c1c-b462-bcc050051e04\") " pod="openstack/nova-metadata-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.571040 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6tsj\" (UniqueName: \"kubernetes.io/projected/7de6ebed-df18-4907-ba33-3f4674373b43-kube-api-access-b6tsj\") pod \"nova-api-0\" (UID: \"7de6ebed-df18-4907-ba33-3f4674373b43\") " pod="openstack/nova-api-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.572707 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7de6ebed-df18-4907-ba33-3f4674373b43-logs\") pod \"nova-api-0\" (UID: \"7de6ebed-df18-4907-ba33-3f4674373b43\") " pod="openstack/nova-api-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.578309 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7de6ebed-df18-4907-ba33-3f4674373b43-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7de6ebed-df18-4907-ba33-3f4674373b43\") " pod="openstack/nova-api-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.578973 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7de6ebed-df18-4907-ba33-3f4674373b43-config-data\") pod \"nova-api-0\" (UID: \"7de6ebed-df18-4907-ba33-3f4674373b43\") " pod="openstack/nova-api-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.674942 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198455d4-8878-4c1c-b462-bcc050051e04-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"198455d4-8878-4c1c-b462-bcc050051e04\") " pod="openstack/nova-metadata-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.675462 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198455d4-8878-4c1c-b462-bcc050051e04-config-data\") pod \"nova-metadata-0\" (UID: \"198455d4-8878-4c1c-b462-bcc050051e04\") " pod="openstack/nova-metadata-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.675527 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r78t8\" (UniqueName: \"kubernetes.io/projected/198455d4-8878-4c1c-b462-bcc050051e04-kube-api-access-r78t8\") pod \"nova-metadata-0\" (UID: \"198455d4-8878-4c1c-b462-bcc050051e04\") " pod="openstack/nova-metadata-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.675635 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/198455d4-8878-4c1c-b462-bcc050051e04-logs\") pod \"nova-metadata-0\" (UID: \"198455d4-8878-4c1c-b462-bcc050051e04\") " pod="openstack/nova-metadata-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.685834 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/198455d4-8878-4c1c-b462-bcc050051e04-logs\") pod \"nova-metadata-0\" (UID: \"198455d4-8878-4c1c-b462-bcc050051e04\") " pod="openstack/nova-metadata-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.691095 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.692061 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198455d4-8878-4c1c-b462-bcc050051e04-config-data\") pod \"nova-metadata-0\" (UID: \"198455d4-8878-4c1c-b462-bcc050051e04\") " pod="openstack/nova-metadata-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.699520 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198455d4-8878-4c1c-b462-bcc050051e04-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"198455d4-8878-4c1c-b462-bcc050051e04\") " pod="openstack/nova-metadata-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.703963 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.710720 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.717392 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6tsj\" (UniqueName: \"kubernetes.io/projected/7de6ebed-df18-4907-ba33-3f4674373b43-kube-api-access-b6tsj\") pod \"nova-api-0\" (UID: \"7de6ebed-df18-4907-ba33-3f4674373b43\") " pod="openstack/nova-api-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.725284 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.792480 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r78t8\" (UniqueName: \"kubernetes.io/projected/198455d4-8878-4c1c-b462-bcc050051e04-kube-api-access-r78t8\") pod \"nova-metadata-0\" (UID: \"198455d4-8878-4c1c-b462-bcc050051e04\") " pod="openstack/nova-metadata-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.794588 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d92859-2d10-43da-a8f2-64cdd0f4ada6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"18d92859-2d10-43da-a8f2-64cdd0f4ada6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.794921 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18d92859-2d10-43da-a8f2-64cdd0f4ada6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"18d92859-2d10-43da-a8f2-64cdd0f4ada6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.795214 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78tzh\" (UniqueName: \"kubernetes.io/projected/18d92859-2d10-43da-a8f2-64cdd0f4ada6-kube-api-access-78tzh\") pod \"nova-cell1-novncproxy-0\" (UID: \"18d92859-2d10-43da-a8f2-64cdd0f4ada6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.837138 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.912263 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18d92859-2d10-43da-a8f2-64cdd0f4ada6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"18d92859-2d10-43da-a8f2-64cdd0f4ada6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.912588 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78tzh\" (UniqueName: \"kubernetes.io/projected/18d92859-2d10-43da-a8f2-64cdd0f4ada6-kube-api-access-78tzh\") pod \"nova-cell1-novncproxy-0\" (UID: \"18d92859-2d10-43da-a8f2-64cdd0f4ada6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.912787 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d92859-2d10-43da-a8f2-64cdd0f4ada6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"18d92859-2d10-43da-a8f2-64cdd0f4ada6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.915910 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.940669 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d92859-2d10-43da-a8f2-64cdd0f4ada6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"18d92859-2d10-43da-a8f2-64cdd0f4ada6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.941270 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18d92859-2d10-43da-a8f2-64cdd0f4ada6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"18d92859-2d10-43da-a8f2-64cdd0f4ada6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.957424 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:21:33 crc kubenswrapper[4997]: I1205 07:21:33.959141 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.003661 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.017064 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-config-data\") pod \"nova-scheduler-0\" (UID: \"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.017414 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.017573 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jkkw\" (UniqueName: \"kubernetes.io/projected/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-kube-api-access-4jkkw\") pod \"nova-scheduler-0\" (UID: \"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.053111 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-xrxqk"] Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.055036 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.059467 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78tzh\" (UniqueName: \"kubernetes.io/projected/18d92859-2d10-43da-a8f2-64cdd0f4ada6-kube-api-access-78tzh\") pod \"nova-cell1-novncproxy-0\" (UID: \"18d92859-2d10-43da-a8f2-64cdd0f4ada6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.075696 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.121356 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-ovsdbserver-nb\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.121474 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.121644 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-dns-swift-storage-0\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.122298 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jkkw\" (UniqueName: \"kubernetes.io/projected/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-kube-api-access-4jkkw\") pod \"nova-scheduler-0\" (UID: \"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.122438 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-dns-svc\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.122494 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-config-data\") pod \"nova-scheduler-0\" (UID: \"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.122577 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkbfr\" (UniqueName: \"kubernetes.io/projected/09e38b0e-164a-4927-bbfa-87d8f2023a6c-kube-api-access-lkbfr\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.122705 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-config\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.122736 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-ovsdbserver-sb\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.132078 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-xrxqk"] Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.134583 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.137805 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-config-data\") pod \"nova-scheduler-0\" (UID: \"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.206562 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jkkw\" (UniqueName: \"kubernetes.io/projected/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-kube-api-access-4jkkw\") pod \"nova-scheduler-0\" (UID: \"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.282195 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.294429 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-ovsdbserver-nb\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.294695 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-dns-swift-storage-0\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.294930 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-dns-svc\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.295085 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkbfr\" (UniqueName: \"kubernetes.io/projected/09e38b0e-164a-4927-bbfa-87d8f2023a6c-kube-api-access-lkbfr\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.295120 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-config\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.295151 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-ovsdbserver-sb\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.298708 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-ovsdbserver-sb\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.299199 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-ovsdbserver-nb\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.300175 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-dns-swift-storage-0\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.302727 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-dns-svc\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.303021 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-config\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.338380 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkbfr\" (UniqueName: \"kubernetes.io/projected/09e38b0e-164a-4927-bbfa-87d8f2023a6c-kube-api-access-lkbfr\") pod \"dnsmasq-dns-7d75688ddc-xrxqk\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.378230 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.410492 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.724378 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-w6b8v"] Dec 05 07:21:34 crc kubenswrapper[4997]: I1205 07:21:34.741004 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.006983 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.146742 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xg2jl"] Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.148635 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xg2jl" Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.154029 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.154117 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.165832 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xg2jl"] Dec 05 07:21:35 crc kubenswrapper[4997]: W1205 07:21:35.220159 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3319cb2_6a5d_4806_a244_fa4f4b54e3e7.slice/crio-e8db2791d530516f345b458d52544bb65161db1aeb004902dc8f34de4080b7ac WatchSource:0}: Error finding container e8db2791d530516f345b458d52544bb65161db1aeb004902dc8f34de4080b7ac: Status 404 returned error can't find the container with id e8db2791d530516f345b458d52544bb65161db1aeb004902dc8f34de4080b7ac Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.229226 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.231167 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xg2jl\" (UID: \"e6066f8c-fc22-4592-b88e-603063a5a06a\") " pod="openstack/nova-cell1-conductor-db-sync-xg2jl" Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.231204 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-scripts\") pod \"nova-cell1-conductor-db-sync-xg2jl\" (UID: \"e6066f8c-fc22-4592-b88e-603063a5a06a\") " pod="openstack/nova-cell1-conductor-db-sync-xg2jl" Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.231328 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtd88\" (UniqueName: \"kubernetes.io/projected/e6066f8c-fc22-4592-b88e-603063a5a06a-kube-api-access-xtd88\") pod \"nova-cell1-conductor-db-sync-xg2jl\" (UID: \"e6066f8c-fc22-4592-b88e-603063a5a06a\") " pod="openstack/nova-cell1-conductor-db-sync-xg2jl" Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.231526 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-config-data\") pod \"nova-cell1-conductor-db-sync-xg2jl\" (UID: \"e6066f8c-fc22-4592-b88e-603063a5a06a\") " pod="openstack/nova-cell1-conductor-db-sync-xg2jl" Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.238079 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 07:21:35 crc kubenswrapper[4997]: W1205 07:21:35.239154 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18d92859_2d10_43da_a8f2_64cdd0f4ada6.slice/crio-304f79cb6bc4e772c2d902f2a1a65d629cf887f5d4cf5ea87918efb30859678a WatchSource:0}: Error finding container 304f79cb6bc4e772c2d902f2a1a65d629cf887f5d4cf5ea87918efb30859678a: Status 404 returned error can't find the container with id 304f79cb6bc4e772c2d902f2a1a65d629cf887f5d4cf5ea87918efb30859678a Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.312160 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-xrxqk"] Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.334206 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xg2jl\" (UID: \"e6066f8c-fc22-4592-b88e-603063a5a06a\") " pod="openstack/nova-cell1-conductor-db-sync-xg2jl" Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.334599 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-scripts\") pod \"nova-cell1-conductor-db-sync-xg2jl\" (UID: \"e6066f8c-fc22-4592-b88e-603063a5a06a\") " pod="openstack/nova-cell1-conductor-db-sync-xg2jl" Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.334723 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtd88\" (UniqueName: \"kubernetes.io/projected/e6066f8c-fc22-4592-b88e-603063a5a06a-kube-api-access-xtd88\") pod \"nova-cell1-conductor-db-sync-xg2jl\" (UID: \"e6066f8c-fc22-4592-b88e-603063a5a06a\") " pod="openstack/nova-cell1-conductor-db-sync-xg2jl" Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.334828 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-config-data\") pod \"nova-cell1-conductor-db-sync-xg2jl\" (UID: \"e6066f8c-fc22-4592-b88e-603063a5a06a\") " pod="openstack/nova-cell1-conductor-db-sync-xg2jl" Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.340769 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-scripts\") pod \"nova-cell1-conductor-db-sync-xg2jl\" (UID: \"e6066f8c-fc22-4592-b88e-603063a5a06a\") " pod="openstack/nova-cell1-conductor-db-sync-xg2jl" Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.341077 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xg2jl\" (UID: \"e6066f8c-fc22-4592-b88e-603063a5a06a\") " pod="openstack/nova-cell1-conductor-db-sync-xg2jl" Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.348742 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-config-data\") pod \"nova-cell1-conductor-db-sync-xg2jl\" (UID: \"e6066f8c-fc22-4592-b88e-603063a5a06a\") " pod="openstack/nova-cell1-conductor-db-sync-xg2jl" Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.351304 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtd88\" (UniqueName: \"kubernetes.io/projected/e6066f8c-fc22-4592-b88e-603063a5a06a-kube-api-access-xtd88\") pod \"nova-cell1-conductor-db-sync-xg2jl\" (UID: \"e6066f8c-fc22-4592-b88e-603063a5a06a\") " pod="openstack/nova-cell1-conductor-db-sync-xg2jl" Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.375439 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"198455d4-8878-4c1c-b462-bcc050051e04","Type":"ContainerStarted","Data":"a1b35aa3a07c94f303d0f3049aece5288b7afd1bc9628f4a9ff861aa7d8e11ca"} Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.381586 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" event={"ID":"09e38b0e-164a-4927-bbfa-87d8f2023a6c","Type":"ContainerStarted","Data":"02f13bc024bad3e76377fb724c2468380b7712120a238ab81278b40d1c74aa55"} Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.386561 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7de6ebed-df18-4907-ba33-3f4674373b43","Type":"ContainerStarted","Data":"e345191317c58142ae65b88ff19bc8d0c0291b804036e075123602fea4342f20"} Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.388248 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"18d92859-2d10-43da-a8f2-64cdd0f4ada6","Type":"ContainerStarted","Data":"304f79cb6bc4e772c2d902f2a1a65d629cf887f5d4cf5ea87918efb30859678a"} Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.390630 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rs4fg" event={"ID":"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3","Type":"ContainerStarted","Data":"4388c645e6ae612a32c56d44e288c44b6832d73edf59318c16e851015285c073"} Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.397710 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-w6b8v" event={"ID":"b2325c5c-6d81-45b1-90ae-d3570c255bea","Type":"ContainerStarted","Data":"159ce5f13204b37468ae891f1ae556c7d3978744ca70fe3f088c098ee1f00ac9"} Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.397778 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-w6b8v" event={"ID":"b2325c5c-6d81-45b1-90ae-d3570c255bea","Type":"ContainerStarted","Data":"85fa7f5f7759121444cc579bc813ab8d38c3c51d5a8e4c58d549d3e61c9406e1"} Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.413391 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7","Type":"ContainerStarted","Data":"e8db2791d530516f345b458d52544bb65161db1aeb004902dc8f34de4080b7ac"} Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.437197 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rs4fg" podStartSLOduration=3.551341258 podStartE2EDuration="6.437151464s" podCreationTimestamp="2025-12-05 07:21:29 +0000 UTC" firstStartedPulling="2025-12-05 07:21:31.222263536 +0000 UTC m=+1591.751170797" lastFinishedPulling="2025-12-05 07:21:34.108073742 +0000 UTC m=+1594.636981003" observedRunningTime="2025-12-05 07:21:35.419458505 +0000 UTC m=+1595.948365766" watchObservedRunningTime="2025-12-05 07:21:35.437151464 +0000 UTC m=+1595.966058725" Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.449941 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-w6b8v" podStartSLOduration=2.449920419 podStartE2EDuration="2.449920419s" podCreationTimestamp="2025-12-05 07:21:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:21:35.445483209 +0000 UTC m=+1595.974390490" watchObservedRunningTime="2025-12-05 07:21:35.449920419 +0000 UTC m=+1595.978827680" Dec 05 07:21:35 crc kubenswrapper[4997]: I1205 07:21:35.497964 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xg2jl" Dec 05 07:21:36 crc kubenswrapper[4997]: I1205 07:21:36.001249 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xg2jl"] Dec 05 07:21:36 crc kubenswrapper[4997]: W1205 07:21:36.035541 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6066f8c_fc22_4592_b88e_603063a5a06a.slice/crio-92048610850aee717b2be8f6d84cb00455b51496d3d2db0d0c2903c4619bb7e9 WatchSource:0}: Error finding container 92048610850aee717b2be8f6d84cb00455b51496d3d2db0d0c2903c4619bb7e9: Status 404 returned error can't find the container with id 92048610850aee717b2be8f6d84cb00455b51496d3d2db0d0c2903c4619bb7e9 Dec 05 07:21:36 crc kubenswrapper[4997]: I1205 07:21:36.472220 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xg2jl" event={"ID":"e6066f8c-fc22-4592-b88e-603063a5a06a","Type":"ContainerStarted","Data":"92048610850aee717b2be8f6d84cb00455b51496d3d2db0d0c2903c4619bb7e9"} Dec 05 07:21:36 crc kubenswrapper[4997]: I1205 07:21:36.522350 4997 generic.go:334] "Generic (PLEG): container finished" podID="09e38b0e-164a-4927-bbfa-87d8f2023a6c" containerID="8262a289c6746a6163cf170184afb1a33d7c8739d56d972fe6ac01db023fe654" exitCode=0 Dec 05 07:21:36 crc kubenswrapper[4997]: I1205 07:21:36.522655 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" event={"ID":"09e38b0e-164a-4927-bbfa-87d8f2023a6c","Type":"ContainerDied","Data":"8262a289c6746a6163cf170184afb1a33d7c8739d56d972fe6ac01db023fe654"} Dec 05 07:21:37 crc kubenswrapper[4997]: I1205 07:21:37.557840 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xg2jl" event={"ID":"e6066f8c-fc22-4592-b88e-603063a5a06a","Type":"ContainerStarted","Data":"6c7e34087d4bfaea3711d104c62c9dd6772529963b74b9d73df375927d1f000e"} Dec 05 07:21:37 crc kubenswrapper[4997]: I1205 07:21:37.571154 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" event={"ID":"09e38b0e-164a-4927-bbfa-87d8f2023a6c","Type":"ContainerStarted","Data":"bd53b88774db5676c0ecbd766a894c0787029e64044e69e6c842f108a8b45f92"} Dec 05 07:21:37 crc kubenswrapper[4997]: I1205 07:21:37.572185 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:37 crc kubenswrapper[4997]: I1205 07:21:37.594184 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-xg2jl" podStartSLOduration=2.594159753 podStartE2EDuration="2.594159753s" podCreationTimestamp="2025-12-05 07:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:21:37.574748288 +0000 UTC m=+1598.103655569" watchObservedRunningTime="2025-12-05 07:21:37.594159753 +0000 UTC m=+1598.123067014" Dec 05 07:21:37 crc kubenswrapper[4997]: I1205 07:21:37.600367 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" podStartSLOduration=4.600293518 podStartE2EDuration="4.600293518s" podCreationTimestamp="2025-12-05 07:21:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:21:37.597640476 +0000 UTC m=+1598.126547757" watchObservedRunningTime="2025-12-05 07:21:37.600293518 +0000 UTC m=+1598.129200799" Dec 05 07:21:38 crc kubenswrapper[4997]: I1205 07:21:38.452293 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:21:38 crc kubenswrapper[4997]: I1205 07:21:38.474594 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.352334 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.361753 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rs4fg" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.362356 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rs4fg" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.434206 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rs4fg" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.494245 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2b7a1c8e-7026-4860-9ab9-0f66cc187559-log-httpd\") pod \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.494657 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-sg-core-conf-yaml\") pod \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.494769 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-config-data\") pod \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.494822 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2b7a1c8e-7026-4860-9ab9-0f66cc187559-run-httpd\") pod \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.494881 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-combined-ca-bundle\") pod \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.494951 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-scripts\") pod \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.494979 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8jg8\" (UniqueName: \"kubernetes.io/projected/2b7a1c8e-7026-4860-9ab9-0f66cc187559-kube-api-access-q8jg8\") pod \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.495558 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b7a1c8e-7026-4860-9ab9-0f66cc187559-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2b7a1c8e-7026-4860-9ab9-0f66cc187559" (UID: "2b7a1c8e-7026-4860-9ab9-0f66cc187559"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.496747 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b7a1c8e-7026-4860-9ab9-0f66cc187559-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2b7a1c8e-7026-4860-9ab9-0f66cc187559" (UID: "2b7a1c8e-7026-4860-9ab9-0f66cc187559"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.510113 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b7a1c8e-7026-4860-9ab9-0f66cc187559-kube-api-access-q8jg8" (OuterVolumeSpecName: "kube-api-access-q8jg8") pod "2b7a1c8e-7026-4860-9ab9-0f66cc187559" (UID: "2b7a1c8e-7026-4860-9ab9-0f66cc187559"). InnerVolumeSpecName "kube-api-access-q8jg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.522801 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-scripts" (OuterVolumeSpecName: "scripts") pod "2b7a1c8e-7026-4860-9ab9-0f66cc187559" (UID: "2b7a1c8e-7026-4860-9ab9-0f66cc187559"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.560629 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2b7a1c8e-7026-4860-9ab9-0f66cc187559" (UID: "2b7a1c8e-7026-4860-9ab9-0f66cc187559"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.597641 4997 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2b7a1c8e-7026-4860-9ab9-0f66cc187559-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.597689 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.597700 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8jg8\" (UniqueName: \"kubernetes.io/projected/2b7a1c8e-7026-4860-9ab9-0f66cc187559-kube-api-access-q8jg8\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.597713 4997 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2b7a1c8e-7026-4860-9ab9-0f66cc187559-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.597722 4997 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.602758 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2b7a1c8e-7026-4860-9ab9-0f66cc187559" (UID: "2b7a1c8e-7026-4860-9ab9-0f66cc187559"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.626543 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"198455d4-8878-4c1c-b462-bcc050051e04","Type":"ContainerStarted","Data":"ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3"} Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.626606 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"198455d4-8878-4c1c-b462-bcc050051e04","Type":"ContainerStarted","Data":"8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620"} Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.626782 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="198455d4-8878-4c1c-b462-bcc050051e04" containerName="nova-metadata-log" containerID="cri-o://8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620" gracePeriod=30 Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.627416 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="198455d4-8878-4c1c-b462-bcc050051e04" containerName="nova-metadata-metadata" containerID="cri-o://ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3" gracePeriod=30 Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.638658 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7de6ebed-df18-4907-ba33-3f4674373b43","Type":"ContainerStarted","Data":"18d8f6410fbb4278bb56d59b7897b54be92fec395556cbd50421cb373abffb78"} Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.638717 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7de6ebed-df18-4907-ba33-3f4674373b43","Type":"ContainerStarted","Data":"7dff80056d298e81740e2d7124e576f7c27286c0a6688de6c08d8ca22980e18d"} Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.648801 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"18d92859-2d10-43da-a8f2-64cdd0f4ada6","Type":"ContainerStarted","Data":"0eeef8fbc295733a014cd992a3da5e4c4d7155017de2751ca5933ede672d5eab"} Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.648972 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="18d92859-2d10-43da-a8f2-64cdd0f4ada6" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://0eeef8fbc295733a014cd992a3da5e4c4d7155017de2751ca5933ede672d5eab" gracePeriod=30 Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.661466 4997 generic.go:334] "Generic (PLEG): container finished" podID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerID="04962c1512e9695617fdd16a154e03c3004c5cdbbec285021d959c5476bf76fc" exitCode=137 Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.661536 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2b7a1c8e-7026-4860-9ab9-0f66cc187559","Type":"ContainerDied","Data":"04962c1512e9695617fdd16a154e03c3004c5cdbbec285021d959c5476bf76fc"} Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.661570 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2b7a1c8e-7026-4860-9ab9-0f66cc187559","Type":"ContainerDied","Data":"f769606626428f7693ced509c300ba94f2fcfa44d6724b2861c2742206928d2c"} Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.661590 4997 scope.go:117] "RemoveContainer" containerID="04962c1512e9695617fdd16a154e03c3004c5cdbbec285021d959c5476bf76fc" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.661787 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.667577 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7","Type":"ContainerStarted","Data":"21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196"} Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.669917 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.148572782 podStartE2EDuration="7.669888622s" podCreationTimestamp="2025-12-05 07:21:33 +0000 UTC" firstStartedPulling="2025-12-05 07:21:35.023795932 +0000 UTC m=+1595.552703193" lastFinishedPulling="2025-12-05 07:21:39.545111772 +0000 UTC m=+1600.074019033" observedRunningTime="2025-12-05 07:21:40.646682195 +0000 UTC m=+1601.175589466" watchObservedRunningTime="2025-12-05 07:21:40.669888622 +0000 UTC m=+1601.198795893" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.681286 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.9429819950000002 podStartE2EDuration="7.681264949s" podCreationTimestamp="2025-12-05 07:21:33 +0000 UTC" firstStartedPulling="2025-12-05 07:21:34.800213739 +0000 UTC m=+1595.329121000" lastFinishedPulling="2025-12-05 07:21:39.538496693 +0000 UTC m=+1600.067403954" observedRunningTime="2025-12-05 07:21:40.668114824 +0000 UTC m=+1601.197022085" watchObservedRunningTime="2025-12-05 07:21:40.681264949 +0000 UTC m=+1601.210172210" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.698330 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.395334932 podStartE2EDuration="7.698309271s" podCreationTimestamp="2025-12-05 07:21:33 +0000 UTC" firstStartedPulling="2025-12-05 07:21:35.242067111 +0000 UTC m=+1595.770974372" lastFinishedPulling="2025-12-05 07:21:39.54504145 +0000 UTC m=+1600.073948711" observedRunningTime="2025-12-05 07:21:40.691959768 +0000 UTC m=+1601.220867039" watchObservedRunningTime="2025-12-05 07:21:40.698309271 +0000 UTC m=+1601.227216532" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.702873 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-config-data" (OuterVolumeSpecName: "config-data") pod "2b7a1c8e-7026-4860-9ab9-0f66cc187559" (UID: "2b7a1c8e-7026-4860-9ab9-0f66cc187559"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.704956 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-config-data\") pod \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\" (UID: \"2b7a1c8e-7026-4860-9ab9-0f66cc187559\") " Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.706116 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:40 crc kubenswrapper[4997]: W1205 07:21:40.706231 4997 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/2b7a1c8e-7026-4860-9ab9-0f66cc187559/volumes/kubernetes.io~secret/config-data Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.706250 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-config-data" (OuterVolumeSpecName: "config-data") pod "2b7a1c8e-7026-4860-9ab9-0f66cc187559" (UID: "2b7a1c8e-7026-4860-9ab9-0f66cc187559"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.724339 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.413713379 podStartE2EDuration="7.724316973s" podCreationTimestamp="2025-12-05 07:21:33 +0000 UTC" firstStartedPulling="2025-12-05 07:21:35.229572054 +0000 UTC m=+1595.758479315" lastFinishedPulling="2025-12-05 07:21:39.540175648 +0000 UTC m=+1600.069082909" observedRunningTime="2025-12-05 07:21:40.720501739 +0000 UTC m=+1601.249409010" watchObservedRunningTime="2025-12-05 07:21:40.724316973 +0000 UTC m=+1601.253224234" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.734700 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rs4fg" Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.801390 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rs4fg"] Dec 05 07:21:40 crc kubenswrapper[4997]: I1205 07:21:40.809657 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b7a1c8e-7026-4860-9ab9-0f66cc187559-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.011650 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.030491 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.052216 4997 scope.go:117] "RemoveContainer" containerID="c81fec47b7d1f800bb7239bb0f25c51ec857bc80d34743f312b791b92369ad13" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.070461 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:21:41 crc kubenswrapper[4997]: E1205 07:21:41.071040 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerName="sg-core" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.071056 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerName="sg-core" Dec 05 07:21:41 crc kubenswrapper[4997]: E1205 07:21:41.071075 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerName="ceilometer-notification-agent" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.071082 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerName="ceilometer-notification-agent" Dec 05 07:21:41 crc kubenswrapper[4997]: E1205 07:21:41.071092 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerName="ceilometer-central-agent" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.071099 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerName="ceilometer-central-agent" Dec 05 07:21:41 crc kubenswrapper[4997]: E1205 07:21:41.071110 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerName="proxy-httpd" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.071118 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerName="proxy-httpd" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.071324 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerName="proxy-httpd" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.071340 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerName="ceilometer-central-agent" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.071354 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerName="ceilometer-notification-agent" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.071363 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" containerName="sg-core" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.073697 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.080794 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.081843 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.086759 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.094588 4997 scope.go:117] "RemoveContainer" containerID="26f28ed47a68cff357f468ae8e1db8ac2a479eb104daa85ba40c09b316f06904" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.140020 4997 scope.go:117] "RemoveContainer" containerID="ac59ac472762c63070b59bce0b71ce6e17b3f41b388a25c53f2bd606a010ee10" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.177905 4997 scope.go:117] "RemoveContainer" containerID="04962c1512e9695617fdd16a154e03c3004c5cdbbec285021d959c5476bf76fc" Dec 05 07:21:41 crc kubenswrapper[4997]: E1205 07:21:41.178588 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04962c1512e9695617fdd16a154e03c3004c5cdbbec285021d959c5476bf76fc\": container with ID starting with 04962c1512e9695617fdd16a154e03c3004c5cdbbec285021d959c5476bf76fc not found: ID does not exist" containerID="04962c1512e9695617fdd16a154e03c3004c5cdbbec285021d959c5476bf76fc" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.178668 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04962c1512e9695617fdd16a154e03c3004c5cdbbec285021d959c5476bf76fc"} err="failed to get container status \"04962c1512e9695617fdd16a154e03c3004c5cdbbec285021d959c5476bf76fc\": rpc error: code = NotFound desc = could not find container \"04962c1512e9695617fdd16a154e03c3004c5cdbbec285021d959c5476bf76fc\": container with ID starting with 04962c1512e9695617fdd16a154e03c3004c5cdbbec285021d959c5476bf76fc not found: ID does not exist" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.178712 4997 scope.go:117] "RemoveContainer" containerID="c81fec47b7d1f800bb7239bb0f25c51ec857bc80d34743f312b791b92369ad13" Dec 05 07:21:41 crc kubenswrapper[4997]: E1205 07:21:41.179074 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c81fec47b7d1f800bb7239bb0f25c51ec857bc80d34743f312b791b92369ad13\": container with ID starting with c81fec47b7d1f800bb7239bb0f25c51ec857bc80d34743f312b791b92369ad13 not found: ID does not exist" containerID="c81fec47b7d1f800bb7239bb0f25c51ec857bc80d34743f312b791b92369ad13" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.179111 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c81fec47b7d1f800bb7239bb0f25c51ec857bc80d34743f312b791b92369ad13"} err="failed to get container status \"c81fec47b7d1f800bb7239bb0f25c51ec857bc80d34743f312b791b92369ad13\": rpc error: code = NotFound desc = could not find container \"c81fec47b7d1f800bb7239bb0f25c51ec857bc80d34743f312b791b92369ad13\": container with ID starting with c81fec47b7d1f800bb7239bb0f25c51ec857bc80d34743f312b791b92369ad13 not found: ID does not exist" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.179143 4997 scope.go:117] "RemoveContainer" containerID="26f28ed47a68cff357f468ae8e1db8ac2a479eb104daa85ba40c09b316f06904" Dec 05 07:21:41 crc kubenswrapper[4997]: E1205 07:21:41.179526 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26f28ed47a68cff357f468ae8e1db8ac2a479eb104daa85ba40c09b316f06904\": container with ID starting with 26f28ed47a68cff357f468ae8e1db8ac2a479eb104daa85ba40c09b316f06904 not found: ID does not exist" containerID="26f28ed47a68cff357f468ae8e1db8ac2a479eb104daa85ba40c09b316f06904" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.179582 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26f28ed47a68cff357f468ae8e1db8ac2a479eb104daa85ba40c09b316f06904"} err="failed to get container status \"26f28ed47a68cff357f468ae8e1db8ac2a479eb104daa85ba40c09b316f06904\": rpc error: code = NotFound desc = could not find container \"26f28ed47a68cff357f468ae8e1db8ac2a479eb104daa85ba40c09b316f06904\": container with ID starting with 26f28ed47a68cff357f468ae8e1db8ac2a479eb104daa85ba40c09b316f06904 not found: ID does not exist" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.179640 4997 scope.go:117] "RemoveContainer" containerID="ac59ac472762c63070b59bce0b71ce6e17b3f41b388a25c53f2bd606a010ee10" Dec 05 07:21:41 crc kubenswrapper[4997]: E1205 07:21:41.182119 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac59ac472762c63070b59bce0b71ce6e17b3f41b388a25c53f2bd606a010ee10\": container with ID starting with ac59ac472762c63070b59bce0b71ce6e17b3f41b388a25c53f2bd606a010ee10 not found: ID does not exist" containerID="ac59ac472762c63070b59bce0b71ce6e17b3f41b388a25c53f2bd606a010ee10" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.182151 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac59ac472762c63070b59bce0b71ce6e17b3f41b388a25c53f2bd606a010ee10"} err="failed to get container status \"ac59ac472762c63070b59bce0b71ce6e17b3f41b388a25c53f2bd606a010ee10\": rpc error: code = NotFound desc = could not find container \"ac59ac472762c63070b59bce0b71ce6e17b3f41b388a25c53f2bd606a010ee10\": container with ID starting with ac59ac472762c63070b59bce0b71ce6e17b3f41b388a25c53f2bd606a010ee10 not found: ID does not exist" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.220952 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.221145 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.221176 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wkxp\" (UniqueName: \"kubernetes.io/projected/14ae6221-a8c8-43b0-ac51-2a592e9dd726-kube-api-access-9wkxp\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.221211 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14ae6221-a8c8-43b0-ac51-2a592e9dd726-run-httpd\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.221251 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-config-data\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.221321 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14ae6221-a8c8-43b0-ac51-2a592e9dd726-log-httpd\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.221368 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-scripts\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.238217 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.322510 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198455d4-8878-4c1c-b462-bcc050051e04-combined-ca-bundle\") pod \"198455d4-8878-4c1c-b462-bcc050051e04\" (UID: \"198455d4-8878-4c1c-b462-bcc050051e04\") " Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.322734 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198455d4-8878-4c1c-b462-bcc050051e04-config-data\") pod \"198455d4-8878-4c1c-b462-bcc050051e04\" (UID: \"198455d4-8878-4c1c-b462-bcc050051e04\") " Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.322795 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/198455d4-8878-4c1c-b462-bcc050051e04-logs\") pod \"198455d4-8878-4c1c-b462-bcc050051e04\" (UID: \"198455d4-8878-4c1c-b462-bcc050051e04\") " Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.322878 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r78t8\" (UniqueName: \"kubernetes.io/projected/198455d4-8878-4c1c-b462-bcc050051e04-kube-api-access-r78t8\") pod \"198455d4-8878-4c1c-b462-bcc050051e04\" (UID: \"198455d4-8878-4c1c-b462-bcc050051e04\") " Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.323177 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wkxp\" (UniqueName: \"kubernetes.io/projected/14ae6221-a8c8-43b0-ac51-2a592e9dd726-kube-api-access-9wkxp\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.323220 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14ae6221-a8c8-43b0-ac51-2a592e9dd726-run-httpd\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.323255 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-config-data\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.323310 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14ae6221-a8c8-43b0-ac51-2a592e9dd726-log-httpd\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.323343 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-scripts\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.323407 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.323486 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.324926 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14ae6221-a8c8-43b0-ac51-2a592e9dd726-log-httpd\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.325101 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/198455d4-8878-4c1c-b462-bcc050051e04-logs" (OuterVolumeSpecName: "logs") pod "198455d4-8878-4c1c-b462-bcc050051e04" (UID: "198455d4-8878-4c1c-b462-bcc050051e04"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.325469 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14ae6221-a8c8-43b0-ac51-2a592e9dd726-run-httpd\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.328664 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.329191 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-scripts\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.333097 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/198455d4-8878-4c1c-b462-bcc050051e04-kube-api-access-r78t8" (OuterVolumeSpecName: "kube-api-access-r78t8") pod "198455d4-8878-4c1c-b462-bcc050051e04" (UID: "198455d4-8878-4c1c-b462-bcc050051e04"). InnerVolumeSpecName "kube-api-access-r78t8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.333299 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.333874 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-config-data\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.344411 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wkxp\" (UniqueName: \"kubernetes.io/projected/14ae6221-a8c8-43b0-ac51-2a592e9dd726-kube-api-access-9wkxp\") pod \"ceilometer-0\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.364130 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/198455d4-8878-4c1c-b462-bcc050051e04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "198455d4-8878-4c1c-b462-bcc050051e04" (UID: "198455d4-8878-4c1c-b462-bcc050051e04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.366331 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/198455d4-8878-4c1c-b462-bcc050051e04-config-data" (OuterVolumeSpecName: "config-data") pod "198455d4-8878-4c1c-b462-bcc050051e04" (UID: "198455d4-8878-4c1c-b462-bcc050051e04"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.423921 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.426058 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198455d4-8878-4c1c-b462-bcc050051e04-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.426092 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/198455d4-8878-4c1c-b462-bcc050051e04-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.426106 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r78t8\" (UniqueName: \"kubernetes.io/projected/198455d4-8878-4c1c-b462-bcc050051e04-kube-api-access-r78t8\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.426123 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198455d4-8878-4c1c-b462-bcc050051e04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.683052 4997 generic.go:334] "Generic (PLEG): container finished" podID="198455d4-8878-4c1c-b462-bcc050051e04" containerID="ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3" exitCode=0 Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.683401 4997 generic.go:334] "Generic (PLEG): container finished" podID="198455d4-8878-4c1c-b462-bcc050051e04" containerID="8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620" exitCode=143 Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.683301 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.683164 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"198455d4-8878-4c1c-b462-bcc050051e04","Type":"ContainerDied","Data":"ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3"} Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.683562 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"198455d4-8878-4c1c-b462-bcc050051e04","Type":"ContainerDied","Data":"8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620"} Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.683605 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"198455d4-8878-4c1c-b462-bcc050051e04","Type":"ContainerDied","Data":"a1b35aa3a07c94f303d0f3049aece5288b7afd1bc9628f4a9ff861aa7d8e11ca"} Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.683634 4997 scope.go:117] "RemoveContainer" containerID="ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.726669 4997 scope.go:117] "RemoveContainer" containerID="8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.751856 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.787544 4997 scope.go:117] "RemoveContainer" containerID="ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3" Dec 05 07:21:41 crc kubenswrapper[4997]: E1205 07:21:41.788658 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3\": container with ID starting with ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3 not found: ID does not exist" containerID="ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.788732 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3"} err="failed to get container status \"ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3\": rpc error: code = NotFound desc = could not find container \"ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3\": container with ID starting with ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3 not found: ID does not exist" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.788784 4997 scope.go:117] "RemoveContainer" containerID="8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620" Dec 05 07:21:41 crc kubenswrapper[4997]: E1205 07:21:41.789510 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620\": container with ID starting with 8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620 not found: ID does not exist" containerID="8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.789549 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620"} err="failed to get container status \"8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620\": rpc error: code = NotFound desc = could not find container \"8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620\": container with ID starting with 8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620 not found: ID does not exist" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.789578 4997 scope.go:117] "RemoveContainer" containerID="ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.790246 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3"} err="failed to get container status \"ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3\": rpc error: code = NotFound desc = could not find container \"ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3\": container with ID starting with ac4d30d70ed2e3708e2f8a67073b9c7afe0ea39bc914214d9df19526d52112f3 not found: ID does not exist" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.790271 4997 scope.go:117] "RemoveContainer" containerID="8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.790647 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620"} err="failed to get container status \"8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620\": rpc error: code = NotFound desc = could not find container \"8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620\": container with ID starting with 8dd12ff41dbc8b4c4ee6cae4a8b67c722580d99ff0f36966d10104934a615620 not found: ID does not exist" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.819885 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b7a1c8e-7026-4860-9ab9-0f66cc187559" path="/var/lib/kubelet/pods/2b7a1c8e-7026-4860-9ab9-0f66cc187559/volumes" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.820738 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.820773 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:21:41 crc kubenswrapper[4997]: E1205 07:21:41.821129 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="198455d4-8878-4c1c-b462-bcc050051e04" containerName="nova-metadata-log" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.821148 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="198455d4-8878-4c1c-b462-bcc050051e04" containerName="nova-metadata-log" Dec 05 07:21:41 crc kubenswrapper[4997]: E1205 07:21:41.821165 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="198455d4-8878-4c1c-b462-bcc050051e04" containerName="nova-metadata-metadata" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.821173 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="198455d4-8878-4c1c-b462-bcc050051e04" containerName="nova-metadata-metadata" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.821362 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="198455d4-8878-4c1c-b462-bcc050051e04" containerName="nova-metadata-log" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.821374 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="198455d4-8878-4c1c-b462-bcc050051e04" containerName="nova-metadata-metadata" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.822743 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.822886 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.827397 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.827427 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.943712 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ab6d059-6774-4664-be65-7ca5f45a2037-logs\") pod \"nova-metadata-0\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " pod="openstack/nova-metadata-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.944353 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " pod="openstack/nova-metadata-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.944382 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvfz2\" (UniqueName: \"kubernetes.io/projected/4ab6d059-6774-4664-be65-7ca5f45a2037-kube-api-access-dvfz2\") pod \"nova-metadata-0\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " pod="openstack/nova-metadata-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.944437 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " pod="openstack/nova-metadata-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.944495 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-config-data\") pod \"nova-metadata-0\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " pod="openstack/nova-metadata-0" Dec 05 07:21:41 crc kubenswrapper[4997]: I1205 07:21:41.973798 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:21:42 crc kubenswrapper[4997]: I1205 07:21:42.047532 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " pod="openstack/nova-metadata-0" Dec 05 07:21:42 crc kubenswrapper[4997]: I1205 07:21:42.047628 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-config-data\") pod \"nova-metadata-0\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " pod="openstack/nova-metadata-0" Dec 05 07:21:42 crc kubenswrapper[4997]: I1205 07:21:42.047699 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ab6d059-6774-4664-be65-7ca5f45a2037-logs\") pod \"nova-metadata-0\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " pod="openstack/nova-metadata-0" Dec 05 07:21:42 crc kubenswrapper[4997]: I1205 07:21:42.047765 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " pod="openstack/nova-metadata-0" Dec 05 07:21:42 crc kubenswrapper[4997]: I1205 07:21:42.047782 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvfz2\" (UniqueName: \"kubernetes.io/projected/4ab6d059-6774-4664-be65-7ca5f45a2037-kube-api-access-dvfz2\") pod \"nova-metadata-0\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " pod="openstack/nova-metadata-0" Dec 05 07:21:42 crc kubenswrapper[4997]: I1205 07:21:42.049423 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ab6d059-6774-4664-be65-7ca5f45a2037-logs\") pod \"nova-metadata-0\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " pod="openstack/nova-metadata-0" Dec 05 07:21:42 crc kubenswrapper[4997]: I1205 07:21:42.055087 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-config-data\") pod \"nova-metadata-0\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " pod="openstack/nova-metadata-0" Dec 05 07:21:42 crc kubenswrapper[4997]: I1205 07:21:42.055453 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " pod="openstack/nova-metadata-0" Dec 05 07:21:42 crc kubenswrapper[4997]: I1205 07:21:42.056038 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " pod="openstack/nova-metadata-0" Dec 05 07:21:42 crc kubenswrapper[4997]: I1205 07:21:42.071087 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvfz2\" (UniqueName: \"kubernetes.io/projected/4ab6d059-6774-4664-be65-7ca5f45a2037-kube-api-access-dvfz2\") pod \"nova-metadata-0\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " pod="openstack/nova-metadata-0" Dec 05 07:21:42 crc kubenswrapper[4997]: I1205 07:21:42.149186 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 07:21:42 crc kubenswrapper[4997]: I1205 07:21:42.454221 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:21:42 crc kubenswrapper[4997]: I1205 07:21:42.709225 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14ae6221-a8c8-43b0-ac51-2a592e9dd726","Type":"ContainerStarted","Data":"a7bd73ccb82c71dadb40dad6d9c0c99ca61d76861fb2cb4dfade74f8b1456580"} Dec 05 07:21:42 crc kubenswrapper[4997]: I1205 07:21:42.709887 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14ae6221-a8c8-43b0-ac51-2a592e9dd726","Type":"ContainerStarted","Data":"3c418c0e77dad4a64e83851dc85bd355831038ca14c60bf2f8f83e3a8e83b28a"} Dec 05 07:21:42 crc kubenswrapper[4997]: I1205 07:21:42.713378 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4ab6d059-6774-4664-be65-7ca5f45a2037","Type":"ContainerStarted","Data":"8f47c899abbb6af9b3320b15af9c51283ce75a6cdc6d54963859d436e40677b4"} Dec 05 07:21:42 crc kubenswrapper[4997]: I1205 07:21:42.713589 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rs4fg" podUID="69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3" containerName="registry-server" containerID="cri-o://4388c645e6ae612a32c56d44e288c44b6832d73edf59318c16e851015285c073" gracePeriod=2 Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.458017 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rs4fg" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.489571 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-catalog-content\") pod \"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3\" (UID: \"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3\") " Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.489769 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-utilities\") pod \"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3\" (UID: \"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3\") " Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.489835 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nlmd\" (UniqueName: \"kubernetes.io/projected/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-kube-api-access-4nlmd\") pod \"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3\" (UID: \"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3\") " Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.497776 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-utilities" (OuterVolumeSpecName: "utilities") pod "69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3" (UID: "69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.508941 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-kube-api-access-4nlmd" (OuterVolumeSpecName: "kube-api-access-4nlmd") pod "69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3" (UID: "69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3"). InnerVolumeSpecName "kube-api-access-4nlmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.517795 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3" (UID: "69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.592921 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.593418 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nlmd\" (UniqueName: \"kubernetes.io/projected/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-kube-api-access-4nlmd\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.593430 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.730134 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4ab6d059-6774-4664-be65-7ca5f45a2037","Type":"ContainerStarted","Data":"651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2"} Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.730191 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4ab6d059-6774-4664-be65-7ca5f45a2037","Type":"ContainerStarted","Data":"ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9"} Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.732513 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.732601 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.738035 4997 generic.go:334] "Generic (PLEG): container finished" podID="69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3" containerID="4388c645e6ae612a32c56d44e288c44b6832d73edf59318c16e851015285c073" exitCode=0 Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.738109 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rs4fg" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.738126 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rs4fg" event={"ID":"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3","Type":"ContainerDied","Data":"4388c645e6ae612a32c56d44e288c44b6832d73edf59318c16e851015285c073"} Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.738165 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rs4fg" event={"ID":"69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3","Type":"ContainerDied","Data":"5f40c1f5c852310c1efdc626d4264912e0769a3b32acbda922fd4e1524f3bc4a"} Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.738188 4997 scope.go:117] "RemoveContainer" containerID="4388c645e6ae612a32c56d44e288c44b6832d73edf59318c16e851015285c073" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.742936 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14ae6221-a8c8-43b0-ac51-2a592e9dd726","Type":"ContainerStarted","Data":"b01ea2023bffc9dd49bdaf3a6b38449c2ab85e91386ff4a605523494c7a5cba3"} Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.769718 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="198455d4-8878-4c1c-b462-bcc050051e04" path="/var/lib/kubelet/pods/198455d4-8878-4c1c-b462-bcc050051e04/volumes" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.777389 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.7773603700000002 podStartE2EDuration="2.77736037s" podCreationTimestamp="2025-12-05 07:21:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:21:43.760306429 +0000 UTC m=+1604.289213690" watchObservedRunningTime="2025-12-05 07:21:43.77736037 +0000 UTC m=+1604.306267631" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.787939 4997 scope.go:117] "RemoveContainer" containerID="d662f99f7d137a1f54c5ab371f2664323c289c917544886b96361647c074c3e1" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.810858 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rs4fg"] Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.829348 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rs4fg"] Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.844967 4997 scope.go:117] "RemoveContainer" containerID="d87697661c0663232a7cd7d5bb17c32eb364912a6d887de778a20b4d36f8a5b0" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.895400 4997 scope.go:117] "RemoveContainer" containerID="4388c645e6ae612a32c56d44e288c44b6832d73edf59318c16e851015285c073" Dec 05 07:21:43 crc kubenswrapper[4997]: E1205 07:21:43.896325 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4388c645e6ae612a32c56d44e288c44b6832d73edf59318c16e851015285c073\": container with ID starting with 4388c645e6ae612a32c56d44e288c44b6832d73edf59318c16e851015285c073 not found: ID does not exist" containerID="4388c645e6ae612a32c56d44e288c44b6832d73edf59318c16e851015285c073" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.896408 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4388c645e6ae612a32c56d44e288c44b6832d73edf59318c16e851015285c073"} err="failed to get container status \"4388c645e6ae612a32c56d44e288c44b6832d73edf59318c16e851015285c073\": rpc error: code = NotFound desc = could not find container \"4388c645e6ae612a32c56d44e288c44b6832d73edf59318c16e851015285c073\": container with ID starting with 4388c645e6ae612a32c56d44e288c44b6832d73edf59318c16e851015285c073 not found: ID does not exist" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.896445 4997 scope.go:117] "RemoveContainer" containerID="d662f99f7d137a1f54c5ab371f2664323c289c917544886b96361647c074c3e1" Dec 05 07:21:43 crc kubenswrapper[4997]: E1205 07:21:43.897029 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d662f99f7d137a1f54c5ab371f2664323c289c917544886b96361647c074c3e1\": container with ID starting with d662f99f7d137a1f54c5ab371f2664323c289c917544886b96361647c074c3e1 not found: ID does not exist" containerID="d662f99f7d137a1f54c5ab371f2664323c289c917544886b96361647c074c3e1" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.897059 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d662f99f7d137a1f54c5ab371f2664323c289c917544886b96361647c074c3e1"} err="failed to get container status \"d662f99f7d137a1f54c5ab371f2664323c289c917544886b96361647c074c3e1\": rpc error: code = NotFound desc = could not find container \"d662f99f7d137a1f54c5ab371f2664323c289c917544886b96361647c074c3e1\": container with ID starting with d662f99f7d137a1f54c5ab371f2664323c289c917544886b96361647c074c3e1 not found: ID does not exist" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.897080 4997 scope.go:117] "RemoveContainer" containerID="d87697661c0663232a7cd7d5bb17c32eb364912a6d887de778a20b4d36f8a5b0" Dec 05 07:21:43 crc kubenswrapper[4997]: E1205 07:21:43.897683 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d87697661c0663232a7cd7d5bb17c32eb364912a6d887de778a20b4d36f8a5b0\": container with ID starting with d87697661c0663232a7cd7d5bb17c32eb364912a6d887de778a20b4d36f8a5b0 not found: ID does not exist" containerID="d87697661c0663232a7cd7d5bb17c32eb364912a6d887de778a20b4d36f8a5b0" Dec 05 07:21:43 crc kubenswrapper[4997]: I1205 07:21:43.897730 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d87697661c0663232a7cd7d5bb17c32eb364912a6d887de778a20b4d36f8a5b0"} err="failed to get container status \"d87697661c0663232a7cd7d5bb17c32eb364912a6d887de778a20b4d36f8a5b0\": rpc error: code = NotFound desc = could not find container \"d87697661c0663232a7cd7d5bb17c32eb364912a6d887de778a20b4d36f8a5b0\": container with ID starting with d87697661c0663232a7cd7d5bb17c32eb364912a6d887de778a20b4d36f8a5b0 not found: ID does not exist" Dec 05 07:21:44 crc kubenswrapper[4997]: I1205 07:21:44.284588 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:21:44 crc kubenswrapper[4997]: I1205 07:21:44.379222 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 07:21:44 crc kubenswrapper[4997]: I1205 07:21:44.379280 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 07:21:44 crc kubenswrapper[4997]: I1205 07:21:44.413525 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:21:44 crc kubenswrapper[4997]: I1205 07:21:44.416535 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 07:21:44 crc kubenswrapper[4997]: I1205 07:21:44.534170 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-62wl6"] Dec 05 07:21:44 crc kubenswrapper[4997]: I1205 07:21:44.541972 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" podUID="e4b49169-e268-4a11-81fc-fe32226a01fd" containerName="dnsmasq-dns" containerID="cri-o://bafe6e887ce62394ebc58d484ef2d5496cf300ba94f50eae3ce35d5603d9ea1c" gracePeriod=10 Dec 05 07:21:44 crc kubenswrapper[4997]: I1205 07:21:44.763208 4997 generic.go:334] "Generic (PLEG): container finished" podID="e4b49169-e268-4a11-81fc-fe32226a01fd" containerID="bafe6e887ce62394ebc58d484ef2d5496cf300ba94f50eae3ce35d5603d9ea1c" exitCode=0 Dec 05 07:21:44 crc kubenswrapper[4997]: I1205 07:21:44.763330 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" event={"ID":"e4b49169-e268-4a11-81fc-fe32226a01fd","Type":"ContainerDied","Data":"bafe6e887ce62394ebc58d484ef2d5496cf300ba94f50eae3ce35d5603d9ea1c"} Dec 05 07:21:44 crc kubenswrapper[4997]: I1205 07:21:44.814768 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 07:21:44 crc kubenswrapper[4997]: I1205 07:21:44.817226 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7de6ebed-df18-4907-ba33-3f4674373b43" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 07:21:44 crc kubenswrapper[4997]: I1205 07:21:44.817233 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7de6ebed-df18-4907-ba33-3f4674373b43" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.095813 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" podUID="e4b49169-e268-4a11-81fc-fe32226a01fd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.161:5353: connect: connection refused" Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.617721 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.683794 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-dns-swift-storage-0\") pod \"e4b49169-e268-4a11-81fc-fe32226a01fd\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.683990 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-dns-svc\") pod \"e4b49169-e268-4a11-81fc-fe32226a01fd\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.684085 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-ovsdbserver-nb\") pod \"e4b49169-e268-4a11-81fc-fe32226a01fd\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.684162 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-config\") pod \"e4b49169-e268-4a11-81fc-fe32226a01fd\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.684204 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgfqj\" (UniqueName: \"kubernetes.io/projected/e4b49169-e268-4a11-81fc-fe32226a01fd-kube-api-access-bgfqj\") pod \"e4b49169-e268-4a11-81fc-fe32226a01fd\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.684252 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-ovsdbserver-sb\") pod \"e4b49169-e268-4a11-81fc-fe32226a01fd\" (UID: \"e4b49169-e268-4a11-81fc-fe32226a01fd\") " Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.706829 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4b49169-e268-4a11-81fc-fe32226a01fd-kube-api-access-bgfqj" (OuterVolumeSpecName: "kube-api-access-bgfqj") pod "e4b49169-e268-4a11-81fc-fe32226a01fd" (UID: "e4b49169-e268-4a11-81fc-fe32226a01fd"). InnerVolumeSpecName "kube-api-access-bgfqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.774494 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3" path="/var/lib/kubelet/pods/69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3/volumes" Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.789820 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgfqj\" (UniqueName: \"kubernetes.io/projected/e4b49169-e268-4a11-81fc-fe32226a01fd-kube-api-access-bgfqj\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.790107 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e4b49169-e268-4a11-81fc-fe32226a01fd" (UID: "e4b49169-e268-4a11-81fc-fe32226a01fd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.803768 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.806508 4997 generic.go:334] "Generic (PLEG): container finished" podID="b2325c5c-6d81-45b1-90ae-d3570c255bea" containerID="159ce5f13204b37468ae891f1ae556c7d3978744ca70fe3f088c098ee1f00ac9" exitCode=0 Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.847449 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-config" (OuterVolumeSpecName: "config") pod "e4b49169-e268-4a11-81fc-fe32226a01fd" (UID: "e4b49169-e268-4a11-81fc-fe32226a01fd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.855598 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e4b49169-e268-4a11-81fc-fe32226a01fd" (UID: "e4b49169-e268-4a11-81fc-fe32226a01fd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.859352 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e4b49169-e268-4a11-81fc-fe32226a01fd" (UID: "e4b49169-e268-4a11-81fc-fe32226a01fd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.882172 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e4b49169-e268-4a11-81fc-fe32226a01fd" (UID: "e4b49169-e268-4a11-81fc-fe32226a01fd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.893849 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.893890 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.893900 4997 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.893912 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.893923 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e4b49169-e268-4a11-81fc-fe32226a01fd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.965357 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8c9c7-62wl6" event={"ID":"e4b49169-e268-4a11-81fc-fe32226a01fd","Type":"ContainerDied","Data":"6e425e4aeb21e62c46892cff8bf494c237226acd32bfd5d16f8667eb894e5924"} Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.965455 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-w6b8v" event={"ID":"b2325c5c-6d81-45b1-90ae-d3570c255bea","Type":"ContainerDied","Data":"159ce5f13204b37468ae891f1ae556c7d3978744ca70fe3f088c098ee1f00ac9"} Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.965482 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14ae6221-a8c8-43b0-ac51-2a592e9dd726","Type":"ContainerStarted","Data":"4d5258220aa8b4d6ed169649576d64364b57d578c9fff38f0952dbdcdf27a945"} Dec 05 07:21:45 crc kubenswrapper[4997]: I1205 07:21:45.965550 4997 scope.go:117] "RemoveContainer" containerID="bafe6e887ce62394ebc58d484ef2d5496cf300ba94f50eae3ce35d5603d9ea1c" Dec 05 07:21:46 crc kubenswrapper[4997]: I1205 07:21:46.005641 4997 scope.go:117] "RemoveContainer" containerID="eb4dacc8abb90bf9759f70da561b93049b2c7561f701c584106c80b19aabf364" Dec 05 07:21:46 crc kubenswrapper[4997]: I1205 07:21:46.175569 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-62wl6"] Dec 05 07:21:46 crc kubenswrapper[4997]: I1205 07:21:46.189153 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-62wl6"] Dec 05 07:21:46 crc kubenswrapper[4997]: I1205 07:21:46.829851 4997 generic.go:334] "Generic (PLEG): container finished" podID="e6066f8c-fc22-4592-b88e-603063a5a06a" containerID="6c7e34087d4bfaea3711d104c62c9dd6772529963b74b9d73df375927d1f000e" exitCode=0 Dec 05 07:21:46 crc kubenswrapper[4997]: I1205 07:21:46.829936 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xg2jl" event={"ID":"e6066f8c-fc22-4592-b88e-603063a5a06a","Type":"ContainerDied","Data":"6c7e34087d4bfaea3711d104c62c9dd6772529963b74b9d73df375927d1f000e"} Dec 05 07:21:46 crc kubenswrapper[4997]: I1205 07:21:46.834969 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14ae6221-a8c8-43b0-ac51-2a592e9dd726","Type":"ContainerStarted","Data":"3fdc67df65aa7ef91669a32b7905e294fd97b1083008a1f6e7d196d58e703279"} Dec 05 07:21:46 crc kubenswrapper[4997]: I1205 07:21:46.884200 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.765582653 podStartE2EDuration="5.884178149s" podCreationTimestamp="2025-12-05 07:21:41 +0000 UTC" firstStartedPulling="2025-12-05 07:21:41.967329269 +0000 UTC m=+1602.496236530" lastFinishedPulling="2025-12-05 07:21:46.085924765 +0000 UTC m=+1606.614832026" observedRunningTime="2025-12-05 07:21:46.883401328 +0000 UTC m=+1607.412308599" watchObservedRunningTime="2025-12-05 07:21:46.884178149 +0000 UTC m=+1607.413085410" Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.149699 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.151387 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.281729 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-w6b8v" Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.434465 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-combined-ca-bundle\") pod \"b2325c5c-6d81-45b1-90ae-d3570c255bea\" (UID: \"b2325c5c-6d81-45b1-90ae-d3570c255bea\") " Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.434560 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-config-data\") pod \"b2325c5c-6d81-45b1-90ae-d3570c255bea\" (UID: \"b2325c5c-6d81-45b1-90ae-d3570c255bea\") " Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.434677 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-scripts\") pod \"b2325c5c-6d81-45b1-90ae-d3570c255bea\" (UID: \"b2325c5c-6d81-45b1-90ae-d3570c255bea\") " Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.434827 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6bvc\" (UniqueName: \"kubernetes.io/projected/b2325c5c-6d81-45b1-90ae-d3570c255bea-kube-api-access-v6bvc\") pod \"b2325c5c-6d81-45b1-90ae-d3570c255bea\" (UID: \"b2325c5c-6d81-45b1-90ae-d3570c255bea\") " Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.446184 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2325c5c-6d81-45b1-90ae-d3570c255bea-kube-api-access-v6bvc" (OuterVolumeSpecName: "kube-api-access-v6bvc") pod "b2325c5c-6d81-45b1-90ae-d3570c255bea" (UID: "b2325c5c-6d81-45b1-90ae-d3570c255bea"). InnerVolumeSpecName "kube-api-access-v6bvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.462116 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-scripts" (OuterVolumeSpecName: "scripts") pod "b2325c5c-6d81-45b1-90ae-d3570c255bea" (UID: "b2325c5c-6d81-45b1-90ae-d3570c255bea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.493159 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2325c5c-6d81-45b1-90ae-d3570c255bea" (UID: "b2325c5c-6d81-45b1-90ae-d3570c255bea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.496116 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-config-data" (OuterVolumeSpecName: "config-data") pod "b2325c5c-6d81-45b1-90ae-d3570c255bea" (UID: "b2325c5c-6d81-45b1-90ae-d3570c255bea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.538870 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.540993 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6bvc\" (UniqueName: \"kubernetes.io/projected/b2325c5c-6d81-45b1-90ae-d3570c255bea-kube-api-access-v6bvc\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.541019 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.541035 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2325c5c-6d81-45b1-90ae-d3570c255bea-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.760394 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4b49169-e268-4a11-81fc-fe32226a01fd" path="/var/lib/kubelet/pods/e4b49169-e268-4a11-81fc-fe32226a01fd/volumes" Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.845217 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-w6b8v" Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.845291 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-w6b8v" event={"ID":"b2325c5c-6d81-45b1-90ae-d3570c255bea","Type":"ContainerDied","Data":"85fa7f5f7759121444cc579bc813ab8d38c3c51d5a8e4c58d549d3e61c9406e1"} Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.845360 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85fa7f5f7759121444cc579bc813ab8d38c3c51d5a8e4c58d549d3e61c9406e1" Dec 05 07:21:47 crc kubenswrapper[4997]: I1205 07:21:47.847675 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.191119 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.191446 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7de6ebed-df18-4907-ba33-3f4674373b43" containerName="nova-api-log" containerID="cri-o://7dff80056d298e81740e2d7124e576f7c27286c0a6688de6c08d8ca22980e18d" gracePeriod=30 Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.192098 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7de6ebed-df18-4907-ba33-3f4674373b43" containerName="nova-api-api" containerID="cri-o://18d8f6410fbb4278bb56d59b7897b54be92fec395556cbd50421cb373abffb78" gracePeriod=30 Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.213347 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.216097 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f3319cb2-6a5d-4806-a244-fa4f4b54e3e7" containerName="nova-scheduler-scheduler" containerID="cri-o://21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196" gracePeriod=30 Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.243722 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.268707 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xg2jl" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.360525 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-scripts\") pod \"e6066f8c-fc22-4592-b88e-603063a5a06a\" (UID: \"e6066f8c-fc22-4592-b88e-603063a5a06a\") " Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.360765 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtd88\" (UniqueName: \"kubernetes.io/projected/e6066f8c-fc22-4592-b88e-603063a5a06a-kube-api-access-xtd88\") pod \"e6066f8c-fc22-4592-b88e-603063a5a06a\" (UID: \"e6066f8c-fc22-4592-b88e-603063a5a06a\") " Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.360858 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-config-data\") pod \"e6066f8c-fc22-4592-b88e-603063a5a06a\" (UID: \"e6066f8c-fc22-4592-b88e-603063a5a06a\") " Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.360887 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-combined-ca-bundle\") pod \"e6066f8c-fc22-4592-b88e-603063a5a06a\" (UID: \"e6066f8c-fc22-4592-b88e-603063a5a06a\") " Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.375313 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-scripts" (OuterVolumeSpecName: "scripts") pod "e6066f8c-fc22-4592-b88e-603063a5a06a" (UID: "e6066f8c-fc22-4592-b88e-603063a5a06a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.392822 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6066f8c-fc22-4592-b88e-603063a5a06a-kube-api-access-xtd88" (OuterVolumeSpecName: "kube-api-access-xtd88") pod "e6066f8c-fc22-4592-b88e-603063a5a06a" (UID: "e6066f8c-fc22-4592-b88e-603063a5a06a"). InnerVolumeSpecName "kube-api-access-xtd88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.415345 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e6066f8c-fc22-4592-b88e-603063a5a06a" (UID: "e6066f8c-fc22-4592-b88e-603063a5a06a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.423909 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-config-data" (OuterVolumeSpecName: "config-data") pod "e6066f8c-fc22-4592-b88e-603063a5a06a" (UID: "e6066f8c-fc22-4592-b88e-603063a5a06a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.463267 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtd88\" (UniqueName: \"kubernetes.io/projected/e6066f8c-fc22-4592-b88e-603063a5a06a-kube-api-access-xtd88\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.463312 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.463322 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.463332 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6066f8c-fc22-4592-b88e-603063a5a06a-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.858988 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xg2jl" event={"ID":"e6066f8c-fc22-4592-b88e-603063a5a06a","Type":"ContainerDied","Data":"92048610850aee717b2be8f6d84cb00455b51496d3d2db0d0c2903c4619bb7e9"} Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.859057 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92048610850aee717b2be8f6d84cb00455b51496d3d2db0d0c2903c4619bb7e9" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.859129 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xg2jl" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.876082 4997 generic.go:334] "Generic (PLEG): container finished" podID="7de6ebed-df18-4907-ba33-3f4674373b43" containerID="7dff80056d298e81740e2d7124e576f7c27286c0a6688de6c08d8ca22980e18d" exitCode=143 Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.876157 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7de6ebed-df18-4907-ba33-3f4674373b43","Type":"ContainerDied","Data":"7dff80056d298e81740e2d7124e576f7c27286c0a6688de6c08d8ca22980e18d"} Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.876399 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4ab6d059-6774-4664-be65-7ca5f45a2037" containerName="nova-metadata-log" containerID="cri-o://651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2" gracePeriod=30 Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.876521 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4ab6d059-6774-4664-be65-7ca5f45a2037" containerName="nova-metadata-metadata" containerID="cri-o://ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9" gracePeriod=30 Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.997483 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 07:21:48 crc kubenswrapper[4997]: E1205 07:21:48.997975 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3" containerName="registry-server" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.997999 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3" containerName="registry-server" Dec 05 07:21:48 crc kubenswrapper[4997]: E1205 07:21:48.998012 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6066f8c-fc22-4592-b88e-603063a5a06a" containerName="nova-cell1-conductor-db-sync" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.998019 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6066f8c-fc22-4592-b88e-603063a5a06a" containerName="nova-cell1-conductor-db-sync" Dec 05 07:21:48 crc kubenswrapper[4997]: E1205 07:21:48.998033 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3" containerName="extract-content" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.998039 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3" containerName="extract-content" Dec 05 07:21:48 crc kubenswrapper[4997]: E1205 07:21:48.998054 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4b49169-e268-4a11-81fc-fe32226a01fd" containerName="dnsmasq-dns" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.998062 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4b49169-e268-4a11-81fc-fe32226a01fd" containerName="dnsmasq-dns" Dec 05 07:21:48 crc kubenswrapper[4997]: E1205 07:21:48.998083 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2325c5c-6d81-45b1-90ae-d3570c255bea" containerName="nova-manage" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.998100 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2325c5c-6d81-45b1-90ae-d3570c255bea" containerName="nova-manage" Dec 05 07:21:48 crc kubenswrapper[4997]: E1205 07:21:48.998117 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3" containerName="extract-utilities" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.998123 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3" containerName="extract-utilities" Dec 05 07:21:48 crc kubenswrapper[4997]: E1205 07:21:48.998133 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4b49169-e268-4a11-81fc-fe32226a01fd" containerName="init" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.998139 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4b49169-e268-4a11-81fc-fe32226a01fd" containerName="init" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.998325 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6066f8c-fc22-4592-b88e-603063a5a06a" containerName="nova-cell1-conductor-db-sync" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.998344 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4b49169-e268-4a11-81fc-fe32226a01fd" containerName="dnsmasq-dns" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.998357 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2325c5c-6d81-45b1-90ae-d3570c255bea" containerName="nova-manage" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.998365 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="69cbf37e-7d9b-4895-a0b2-e1bdf4a169c3" containerName="registry-server" Dec 05 07:21:48 crc kubenswrapper[4997]: I1205 07:21:48.999129 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.001931 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.024843 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.076452 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmh68\" (UniqueName: \"kubernetes.io/projected/8374cc67-e735-47f0-b310-6bebd608cece-kube-api-access-xmh68\") pod \"nova-cell1-conductor-0\" (UID: \"8374cc67-e735-47f0-b310-6bebd608cece\") " pod="openstack/nova-cell1-conductor-0" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.076512 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8374cc67-e735-47f0-b310-6bebd608cece-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8374cc67-e735-47f0-b310-6bebd608cece\") " pod="openstack/nova-cell1-conductor-0" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.076590 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8374cc67-e735-47f0-b310-6bebd608cece-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8374cc67-e735-47f0-b310-6bebd608cece\") " pod="openstack/nova-cell1-conductor-0" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.178539 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8374cc67-e735-47f0-b310-6bebd608cece-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8374cc67-e735-47f0-b310-6bebd608cece\") " pod="openstack/nova-cell1-conductor-0" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.178737 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmh68\" (UniqueName: \"kubernetes.io/projected/8374cc67-e735-47f0-b310-6bebd608cece-kube-api-access-xmh68\") pod \"nova-cell1-conductor-0\" (UID: \"8374cc67-e735-47f0-b310-6bebd608cece\") " pod="openstack/nova-cell1-conductor-0" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.178776 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8374cc67-e735-47f0-b310-6bebd608cece-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8374cc67-e735-47f0-b310-6bebd608cece\") " pod="openstack/nova-cell1-conductor-0" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.188512 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8374cc67-e735-47f0-b310-6bebd608cece-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8374cc67-e735-47f0-b310-6bebd608cece\") " pod="openstack/nova-cell1-conductor-0" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.202864 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8374cc67-e735-47f0-b310-6bebd608cece-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8374cc67-e735-47f0-b310-6bebd608cece\") " pod="openstack/nova-cell1-conductor-0" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.224586 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmh68\" (UniqueName: \"kubernetes.io/projected/8374cc67-e735-47f0-b310-6bebd608cece-kube-api-access-xmh68\") pod \"nova-cell1-conductor-0\" (UID: \"8374cc67-e735-47f0-b310-6bebd608cece\") " pod="openstack/nova-cell1-conductor-0" Dec 05 07:21:49 crc kubenswrapper[4997]: E1205 07:21:49.381067 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196 is running failed: container process not found" containerID="21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 07:21:49 crc kubenswrapper[4997]: E1205 07:21:49.385071 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196 is running failed: container process not found" containerID="21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 07:21:49 crc kubenswrapper[4997]: E1205 07:21:49.387001 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196 is running failed: container process not found" containerID="21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 07:21:49 crc kubenswrapper[4997]: E1205 07:21:49.387041 4997 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="f3319cb2-6a5d-4806-a244-fa4f4b54e3e7" containerName="nova-scheduler-scheduler" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.486568 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.511787 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.600769 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-config-data\") pod \"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7\" (UID: \"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7\") " Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.600914 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-combined-ca-bundle\") pod \"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7\" (UID: \"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7\") " Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.601035 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jkkw\" (UniqueName: \"kubernetes.io/projected/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-kube-api-access-4jkkw\") pod \"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7\" (UID: \"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7\") " Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.617170 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-kube-api-access-4jkkw" (OuterVolumeSpecName: "kube-api-access-4jkkw") pod "f3319cb2-6a5d-4806-a244-fa4f4b54e3e7" (UID: "f3319cb2-6a5d-4806-a244-fa4f4b54e3e7"). InnerVolumeSpecName "kube-api-access-4jkkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.657824 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-config-data" (OuterVolumeSpecName: "config-data") pod "f3319cb2-6a5d-4806-a244-fa4f4b54e3e7" (UID: "f3319cb2-6a5d-4806-a244-fa4f4b54e3e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.658572 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3319cb2-6a5d-4806-a244-fa4f4b54e3e7" (UID: "f3319cb2-6a5d-4806-a244-fa4f4b54e3e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.705462 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.705505 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.705536 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jkkw\" (UniqueName: \"kubernetes.io/projected/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7-kube-api-access-4jkkw\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.775528 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.775586 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.838945 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.905069 4997 generic.go:334] "Generic (PLEG): container finished" podID="4ab6d059-6774-4664-be65-7ca5f45a2037" containerID="ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9" exitCode=0 Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.905113 4997 generic.go:334] "Generic (PLEG): container finished" podID="4ab6d059-6774-4664-be65-7ca5f45a2037" containerID="651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2" exitCode=143 Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.905162 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4ab6d059-6774-4664-be65-7ca5f45a2037","Type":"ContainerDied","Data":"ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9"} Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.905198 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4ab6d059-6774-4664-be65-7ca5f45a2037","Type":"ContainerDied","Data":"651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2"} Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.905210 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4ab6d059-6774-4664-be65-7ca5f45a2037","Type":"ContainerDied","Data":"8f47c899abbb6af9b3320b15af9c51283ce75a6cdc6d54963859d436e40677b4"} Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.905227 4997 scope.go:117] "RemoveContainer" containerID="ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.905381 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.907860 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ab6d059-6774-4664-be65-7ca5f45a2037-logs\") pod \"4ab6d059-6774-4664-be65-7ca5f45a2037\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.907915 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-config-data\") pod \"4ab6d059-6774-4664-be65-7ca5f45a2037\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.907977 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvfz2\" (UniqueName: \"kubernetes.io/projected/4ab6d059-6774-4664-be65-7ca5f45a2037-kube-api-access-dvfz2\") pod \"4ab6d059-6774-4664-be65-7ca5f45a2037\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.908013 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-nova-metadata-tls-certs\") pod \"4ab6d059-6774-4664-be65-7ca5f45a2037\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.908140 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-combined-ca-bundle\") pod \"4ab6d059-6774-4664-be65-7ca5f45a2037\" (UID: \"4ab6d059-6774-4664-be65-7ca5f45a2037\") " Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.910285 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ab6d059-6774-4664-be65-7ca5f45a2037-logs" (OuterVolumeSpecName: "logs") pod "4ab6d059-6774-4664-be65-7ca5f45a2037" (UID: "4ab6d059-6774-4664-be65-7ca5f45a2037"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.912695 4997 generic.go:334] "Generic (PLEG): container finished" podID="f3319cb2-6a5d-4806-a244-fa4f4b54e3e7" containerID="21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196" exitCode=0 Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.912780 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7","Type":"ContainerDied","Data":"21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196"} Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.912819 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f3319cb2-6a5d-4806-a244-fa4f4b54e3e7","Type":"ContainerDied","Data":"e8db2791d530516f345b458d52544bb65161db1aeb004902dc8f34de4080b7ac"} Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.912894 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.917541 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ab6d059-6774-4664-be65-7ca5f45a2037-kube-api-access-dvfz2" (OuterVolumeSpecName: "kube-api-access-dvfz2") pod "4ab6d059-6774-4664-be65-7ca5f45a2037" (UID: "4ab6d059-6774-4664-be65-7ca5f45a2037"). InnerVolumeSpecName "kube-api-access-dvfz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.928098 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.980821 4997 scope.go:117] "RemoveContainer" containerID="651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.988287 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ab6d059-6774-4664-be65-7ca5f45a2037" (UID: "4ab6d059-6774-4664-be65-7ca5f45a2037"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:49 crc kubenswrapper[4997]: I1205 07:21:49.996959 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-config-data" (OuterVolumeSpecName: "config-data") pod "4ab6d059-6774-4664-be65-7ca5f45a2037" (UID: "4ab6d059-6774-4664-be65-7ca5f45a2037"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.010654 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.010853 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ab6d059-6774-4664-be65-7ca5f45a2037-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.010937 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.011013 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvfz2\" (UniqueName: \"kubernetes.io/projected/4ab6d059-6774-4664-be65-7ca5f45a2037-kube-api-access-dvfz2\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.015726 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "4ab6d059-6774-4664-be65-7ca5f45a2037" (UID: "4ab6d059-6774-4664-be65-7ca5f45a2037"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.088501 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.097383 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.100518 4997 scope.go:117] "RemoveContainer" containerID="ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9" Dec 05 07:21:50 crc kubenswrapper[4997]: E1205 07:21:50.102685 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9\": container with ID starting with ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9 not found: ID does not exist" containerID="ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.102753 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9"} err="failed to get container status \"ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9\": rpc error: code = NotFound desc = could not find container \"ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9\": container with ID starting with ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9 not found: ID does not exist" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.102809 4997 scope.go:117] "RemoveContainer" containerID="651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2" Dec 05 07:21:50 crc kubenswrapper[4997]: E1205 07:21:50.104629 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2\": container with ID starting with 651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2 not found: ID does not exist" containerID="651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.104675 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2"} err="failed to get container status \"651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2\": rpc error: code = NotFound desc = could not find container \"651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2\": container with ID starting with 651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2 not found: ID does not exist" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.104706 4997 scope.go:117] "RemoveContainer" containerID="ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.107983 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9"} err="failed to get container status \"ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9\": rpc error: code = NotFound desc = could not find container \"ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9\": container with ID starting with ab5a2cf2c8392f1cbdfd51a437137ad31ddd1e9d35e47161661a9474c164c9f9 not found: ID does not exist" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.108015 4997 scope.go:117] "RemoveContainer" containerID="651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.110217 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2"} err="failed to get container status \"651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2\": rpc error: code = NotFound desc = could not find container \"651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2\": container with ID starting with 651f9ea8c4439f8110a22db1c122aa8aa99b95bc531641939ed5607583b239c2 not found: ID does not exist" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.110283 4997 scope.go:117] "RemoveContainer" containerID="21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.115389 4997 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ab6d059-6774-4664-be65-7ca5f45a2037-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.121677 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:21:50 crc kubenswrapper[4997]: E1205 07:21:50.122223 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab6d059-6774-4664-be65-7ca5f45a2037" containerName="nova-metadata-metadata" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.122245 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab6d059-6774-4664-be65-7ca5f45a2037" containerName="nova-metadata-metadata" Dec 05 07:21:50 crc kubenswrapper[4997]: E1205 07:21:50.122256 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab6d059-6774-4664-be65-7ca5f45a2037" containerName="nova-metadata-log" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.122264 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab6d059-6774-4664-be65-7ca5f45a2037" containerName="nova-metadata-log" Dec 05 07:21:50 crc kubenswrapper[4997]: E1205 07:21:50.122275 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3319cb2-6a5d-4806-a244-fa4f4b54e3e7" containerName="nova-scheduler-scheduler" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.122281 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3319cb2-6a5d-4806-a244-fa4f4b54e3e7" containerName="nova-scheduler-scheduler" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.122470 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab6d059-6774-4664-be65-7ca5f45a2037" containerName="nova-metadata-metadata" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.122497 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3319cb2-6a5d-4806-a244-fa4f4b54e3e7" containerName="nova-scheduler-scheduler" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.122505 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab6d059-6774-4664-be65-7ca5f45a2037" containerName="nova-metadata-log" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.123284 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.130302 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.143531 4997 scope.go:117] "RemoveContainer" containerID="21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196" Dec 05 07:21:50 crc kubenswrapper[4997]: E1205 07:21:50.148315 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196\": container with ID starting with 21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196 not found: ID does not exist" containerID="21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.148543 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196"} err="failed to get container status \"21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196\": rpc error: code = NotFound desc = could not find container \"21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196\": container with ID starting with 21923fc7ed65b4791827c8c115ac2ea0b6db6e1df1dd2ed0a92cbb82f6501196 not found: ID does not exist" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.159664 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.217109 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-config-data\") pod \"nova-scheduler-0\" (UID: \"a4305f0a-669e-43e3-b6b8-d2f04a556d1a\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.217187 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p62zd\" (UniqueName: \"kubernetes.io/projected/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-kube-api-access-p62zd\") pod \"nova-scheduler-0\" (UID: \"a4305f0a-669e-43e3-b6b8-d2f04a556d1a\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.217218 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a4305f0a-669e-43e3-b6b8-d2f04a556d1a\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.244687 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.289068 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.328841 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.332170 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.333340 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-config-data\") pod \"nova-scheduler-0\" (UID: \"a4305f0a-669e-43e3-b6b8-d2f04a556d1a\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.333500 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p62zd\" (UniqueName: \"kubernetes.io/projected/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-kube-api-access-p62zd\") pod \"nova-scheduler-0\" (UID: \"a4305f0a-669e-43e3-b6b8-d2f04a556d1a\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.333549 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a4305f0a-669e-43e3-b6b8-d2f04a556d1a\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.335597 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.343173 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.343745 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a4305f0a-669e-43e3-b6b8-d2f04a556d1a\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.344062 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-config-data\") pod \"nova-scheduler-0\" (UID: \"a4305f0a-669e-43e3-b6b8-d2f04a556d1a\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.360092 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.360825 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p62zd\" (UniqueName: \"kubernetes.io/projected/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-kube-api-access-p62zd\") pod \"nova-scheduler-0\" (UID: \"a4305f0a-669e-43e3-b6b8-d2f04a556d1a\") " pod="openstack/nova-scheduler-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.435523 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-config-data\") pod \"nova-metadata-0\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " pod="openstack/nova-metadata-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.436035 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " pod="openstack/nova-metadata-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.436065 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9mzf\" (UniqueName: \"kubernetes.io/projected/ef4a9d91-c9b1-4157-b6e5-485063844959-kube-api-access-g9mzf\") pod \"nova-metadata-0\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " pod="openstack/nova-metadata-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.436119 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef4a9d91-c9b1-4157-b6e5-485063844959-logs\") pod \"nova-metadata-0\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " pod="openstack/nova-metadata-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.436201 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " pod="openstack/nova-metadata-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.469934 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.538372 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef4a9d91-c9b1-4157-b6e5-485063844959-logs\") pod \"nova-metadata-0\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " pod="openstack/nova-metadata-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.538543 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " pod="openstack/nova-metadata-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.538723 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-config-data\") pod \"nova-metadata-0\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " pod="openstack/nova-metadata-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.538819 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " pod="openstack/nova-metadata-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.538871 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9mzf\" (UniqueName: \"kubernetes.io/projected/ef4a9d91-c9b1-4157-b6e5-485063844959-kube-api-access-g9mzf\") pod \"nova-metadata-0\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " pod="openstack/nova-metadata-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.538884 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef4a9d91-c9b1-4157-b6e5-485063844959-logs\") pod \"nova-metadata-0\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " pod="openstack/nova-metadata-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.544783 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-config-data\") pod \"nova-metadata-0\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " pod="openstack/nova-metadata-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.545565 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " pod="openstack/nova-metadata-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.550288 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " pod="openstack/nova-metadata-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.564798 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9mzf\" (UniqueName: \"kubernetes.io/projected/ef4a9d91-c9b1-4157-b6e5-485063844959-kube-api-access-g9mzf\") pod \"nova-metadata-0\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " pod="openstack/nova-metadata-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.665787 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.933755 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8374cc67-e735-47f0-b310-6bebd608cece","Type":"ContainerStarted","Data":"175b0ae7934dc4fc48a3e2c7faa4ea18aa39089d2851c19371ca3e1b197fb37a"} Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.934278 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8374cc67-e735-47f0-b310-6bebd608cece","Type":"ContainerStarted","Data":"18cf304d9afa92bcfc6cfed42f01f81ce744f5e8753b217fbfb36d68b769eed7"} Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.935803 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 05 07:21:50 crc kubenswrapper[4997]: I1205 07:21:50.959967 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.959945088 podStartE2EDuration="2.959945088s" podCreationTimestamp="2025-12-05 07:21:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:21:50.959436433 +0000 UTC m=+1611.488343704" watchObservedRunningTime="2025-12-05 07:21:50.959945088 +0000 UTC m=+1611.488852349" Dec 05 07:21:51 crc kubenswrapper[4997]: W1205 07:21:51.000063 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4305f0a_669e_43e3_b6b8_d2f04a556d1a.slice/crio-d8a909c219ab5b0ba0c5c080e427dda4882ede15cd8de4d20a9e7e7df388ad83 WatchSource:0}: Error finding container d8a909c219ab5b0ba0c5c080e427dda4882ede15cd8de4d20a9e7e7df388ad83: Status 404 returned error can't find the container with id d8a909c219ab5b0ba0c5c080e427dda4882ede15cd8de4d20a9e7e7df388ad83 Dec 05 07:21:51 crc kubenswrapper[4997]: I1205 07:21:51.005641 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:21:51 crc kubenswrapper[4997]: I1205 07:21:51.015250 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:21:51 crc kubenswrapper[4997]: W1205 07:21:51.509049 4997 container.go:586] Failed to update stats for container "/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7de6ebed_df18_4907_ba33_3f4674373b43.slice/crio-e345191317c58142ae65b88ff19bc8d0c0291b804036e075123602fea4342f20": error while statting cgroup v2: [unable to parse /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7de6ebed_df18_4907_ba33_3f4674373b43.slice/crio-e345191317c58142ae65b88ff19bc8d0c0291b804036e075123602fea4342f20/memory.stat: read /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7de6ebed_df18_4907_ba33_3f4674373b43.slice/crio-e345191317c58142ae65b88ff19bc8d0c0291b804036e075123602fea4342f20/memory.stat: no such device], continuing to push stats Dec 05 07:21:51 crc kubenswrapper[4997]: I1205 07:21:51.763781 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ab6d059-6774-4664-be65-7ca5f45a2037" path="/var/lib/kubelet/pods/4ab6d059-6774-4664-be65-7ca5f45a2037/volumes" Dec 05 07:21:51 crc kubenswrapper[4997]: I1205 07:21:51.764827 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3319cb2-6a5d-4806-a244-fa4f4b54e3e7" path="/var/lib/kubelet/pods/f3319cb2-6a5d-4806-a244-fa4f4b54e3e7/volumes" Dec 05 07:21:51 crc kubenswrapper[4997]: I1205 07:21:51.941354 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 07:21:51 crc kubenswrapper[4997]: I1205 07:21:51.956397 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ef4a9d91-c9b1-4157-b6e5-485063844959","Type":"ContainerStarted","Data":"47da05d83f44ced35382e509b58b55a2a14d56cd32fca524955be754080be0e4"} Dec 05 07:21:51 crc kubenswrapper[4997]: I1205 07:21:51.956492 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ef4a9d91-c9b1-4157-b6e5-485063844959","Type":"ContainerStarted","Data":"a56a7ab1e0d538fd4235867f849354a7dd6ca9ef4109431edc74e29a0b53d67d"} Dec 05 07:21:51 crc kubenswrapper[4997]: I1205 07:21:51.956505 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ef4a9d91-c9b1-4157-b6e5-485063844959","Type":"ContainerStarted","Data":"ba7919f30ccfce0bdf5be6e5d34db245d5b1292c2c9ea2e80a10a2fce6fb5d08"} Dec 05 07:21:51 crc kubenswrapper[4997]: I1205 07:21:51.959469 4997 generic.go:334] "Generic (PLEG): container finished" podID="7de6ebed-df18-4907-ba33-3f4674373b43" containerID="18d8f6410fbb4278bb56d59b7897b54be92fec395556cbd50421cb373abffb78" exitCode=0 Dec 05 07:21:51 crc kubenswrapper[4997]: I1205 07:21:51.959517 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7de6ebed-df18-4907-ba33-3f4674373b43","Type":"ContainerDied","Data":"18d8f6410fbb4278bb56d59b7897b54be92fec395556cbd50421cb373abffb78"} Dec 05 07:21:51 crc kubenswrapper[4997]: I1205 07:21:51.959558 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7de6ebed-df18-4907-ba33-3f4674373b43","Type":"ContainerDied","Data":"e345191317c58142ae65b88ff19bc8d0c0291b804036e075123602fea4342f20"} Dec 05 07:21:51 crc kubenswrapper[4997]: I1205 07:21:51.959583 4997 scope.go:117] "RemoveContainer" containerID="18d8f6410fbb4278bb56d59b7897b54be92fec395556cbd50421cb373abffb78" Dec 05 07:21:51 crc kubenswrapper[4997]: I1205 07:21:51.959776 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 07:21:51 crc kubenswrapper[4997]: I1205 07:21:51.966846 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a4305f0a-669e-43e3-b6b8-d2f04a556d1a","Type":"ContainerStarted","Data":"3fb15335b8e9a5b14b24d7a4a3e397f45a0acca4d7ed2803cbc0e18cb9c51e7f"} Dec 05 07:21:51 crc kubenswrapper[4997]: I1205 07:21:51.966902 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a4305f0a-669e-43e3-b6b8-d2f04a556d1a","Type":"ContainerStarted","Data":"d8a909c219ab5b0ba0c5c080e427dda4882ede15cd8de4d20a9e7e7df388ad83"} Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.000544 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.000504771 podStartE2EDuration="2.000504771s" podCreationTimestamp="2025-12-05 07:21:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:21:51.989078602 +0000 UTC m=+1612.517985873" watchObservedRunningTime="2025-12-05 07:21:52.000504771 +0000 UTC m=+1612.529412032" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.007685 4997 scope.go:117] "RemoveContainer" containerID="7dff80056d298e81740e2d7124e576f7c27286c0a6688de6c08d8ca22980e18d" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.028601 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.02857601 podStartE2EDuration="2.02857601s" podCreationTimestamp="2025-12-05 07:21:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:21:52.020631765 +0000 UTC m=+1612.549539016" watchObservedRunningTime="2025-12-05 07:21:52.02857601 +0000 UTC m=+1612.557483271" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.037099 4997 scope.go:117] "RemoveContainer" containerID="18d8f6410fbb4278bb56d59b7897b54be92fec395556cbd50421cb373abffb78" Dec 05 07:21:52 crc kubenswrapper[4997]: E1205 07:21:52.037677 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18d8f6410fbb4278bb56d59b7897b54be92fec395556cbd50421cb373abffb78\": container with ID starting with 18d8f6410fbb4278bb56d59b7897b54be92fec395556cbd50421cb373abffb78 not found: ID does not exist" containerID="18d8f6410fbb4278bb56d59b7897b54be92fec395556cbd50421cb373abffb78" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.037709 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18d8f6410fbb4278bb56d59b7897b54be92fec395556cbd50421cb373abffb78"} err="failed to get container status \"18d8f6410fbb4278bb56d59b7897b54be92fec395556cbd50421cb373abffb78\": rpc error: code = NotFound desc = could not find container \"18d8f6410fbb4278bb56d59b7897b54be92fec395556cbd50421cb373abffb78\": container with ID starting with 18d8f6410fbb4278bb56d59b7897b54be92fec395556cbd50421cb373abffb78 not found: ID does not exist" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.037729 4997 scope.go:117] "RemoveContainer" containerID="7dff80056d298e81740e2d7124e576f7c27286c0a6688de6c08d8ca22980e18d" Dec 05 07:21:52 crc kubenswrapper[4997]: E1205 07:21:52.038410 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dff80056d298e81740e2d7124e576f7c27286c0a6688de6c08d8ca22980e18d\": container with ID starting with 7dff80056d298e81740e2d7124e576f7c27286c0a6688de6c08d8ca22980e18d not found: ID does not exist" containerID="7dff80056d298e81740e2d7124e576f7c27286c0a6688de6c08d8ca22980e18d" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.038468 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dff80056d298e81740e2d7124e576f7c27286c0a6688de6c08d8ca22980e18d"} err="failed to get container status \"7dff80056d298e81740e2d7124e576f7c27286c0a6688de6c08d8ca22980e18d\": rpc error: code = NotFound desc = could not find container \"7dff80056d298e81740e2d7124e576f7c27286c0a6688de6c08d8ca22980e18d\": container with ID starting with 7dff80056d298e81740e2d7124e576f7c27286c0a6688de6c08d8ca22980e18d not found: ID does not exist" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.090483 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7de6ebed-df18-4907-ba33-3f4674373b43-logs\") pod \"7de6ebed-df18-4907-ba33-3f4674373b43\" (UID: \"7de6ebed-df18-4907-ba33-3f4674373b43\") " Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.090981 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7de6ebed-df18-4907-ba33-3f4674373b43-logs" (OuterVolumeSpecName: "logs") pod "7de6ebed-df18-4907-ba33-3f4674373b43" (UID: "7de6ebed-df18-4907-ba33-3f4674373b43"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.091144 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6tsj\" (UniqueName: \"kubernetes.io/projected/7de6ebed-df18-4907-ba33-3f4674373b43-kube-api-access-b6tsj\") pod \"7de6ebed-df18-4907-ba33-3f4674373b43\" (UID: \"7de6ebed-df18-4907-ba33-3f4674373b43\") " Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.092211 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7de6ebed-df18-4907-ba33-3f4674373b43-combined-ca-bundle\") pod \"7de6ebed-df18-4907-ba33-3f4674373b43\" (UID: \"7de6ebed-df18-4907-ba33-3f4674373b43\") " Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.092601 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7de6ebed-df18-4907-ba33-3f4674373b43-config-data\") pod \"7de6ebed-df18-4907-ba33-3f4674373b43\" (UID: \"7de6ebed-df18-4907-ba33-3f4674373b43\") " Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.093777 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7de6ebed-df18-4907-ba33-3f4674373b43-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.096999 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7de6ebed-df18-4907-ba33-3f4674373b43-kube-api-access-b6tsj" (OuterVolumeSpecName: "kube-api-access-b6tsj") pod "7de6ebed-df18-4907-ba33-3f4674373b43" (UID: "7de6ebed-df18-4907-ba33-3f4674373b43"). InnerVolumeSpecName "kube-api-access-b6tsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.129312 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7de6ebed-df18-4907-ba33-3f4674373b43-config-data" (OuterVolumeSpecName: "config-data") pod "7de6ebed-df18-4907-ba33-3f4674373b43" (UID: "7de6ebed-df18-4907-ba33-3f4674373b43"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.133366 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7de6ebed-df18-4907-ba33-3f4674373b43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7de6ebed-df18-4907-ba33-3f4674373b43" (UID: "7de6ebed-df18-4907-ba33-3f4674373b43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.195538 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7de6ebed-df18-4907-ba33-3f4674373b43-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.195581 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6tsj\" (UniqueName: \"kubernetes.io/projected/7de6ebed-df18-4907-ba33-3f4674373b43-kube-api-access-b6tsj\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.195595 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7de6ebed-df18-4907-ba33-3f4674373b43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.304977 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.318413 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.333399 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 07:21:52 crc kubenswrapper[4997]: E1205 07:21:52.333955 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7de6ebed-df18-4907-ba33-3f4674373b43" containerName="nova-api-api" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.333977 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de6ebed-df18-4907-ba33-3f4674373b43" containerName="nova-api-api" Dec 05 07:21:52 crc kubenswrapper[4997]: E1205 07:21:52.334013 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7de6ebed-df18-4907-ba33-3f4674373b43" containerName="nova-api-log" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.334020 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de6ebed-df18-4907-ba33-3f4674373b43" containerName="nova-api-log" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.334220 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="7de6ebed-df18-4907-ba33-3f4674373b43" containerName="nova-api-api" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.334252 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="7de6ebed-df18-4907-ba33-3f4674373b43" containerName="nova-api-log" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.335386 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.337944 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.349096 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.401435 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4422dea-fe2c-463c-a3ef-ba129b466e52-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a4422dea-fe2c-463c-a3ef-ba129b466e52\") " pod="openstack/nova-api-0" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.401501 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rzpt\" (UniqueName: \"kubernetes.io/projected/a4422dea-fe2c-463c-a3ef-ba129b466e52-kube-api-access-9rzpt\") pod \"nova-api-0\" (UID: \"a4422dea-fe2c-463c-a3ef-ba129b466e52\") " pod="openstack/nova-api-0" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.401606 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4422dea-fe2c-463c-a3ef-ba129b466e52-config-data\") pod \"nova-api-0\" (UID: \"a4422dea-fe2c-463c-a3ef-ba129b466e52\") " pod="openstack/nova-api-0" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.401672 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4422dea-fe2c-463c-a3ef-ba129b466e52-logs\") pod \"nova-api-0\" (UID: \"a4422dea-fe2c-463c-a3ef-ba129b466e52\") " pod="openstack/nova-api-0" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.503408 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4422dea-fe2c-463c-a3ef-ba129b466e52-config-data\") pod \"nova-api-0\" (UID: \"a4422dea-fe2c-463c-a3ef-ba129b466e52\") " pod="openstack/nova-api-0" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.503953 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4422dea-fe2c-463c-a3ef-ba129b466e52-logs\") pod \"nova-api-0\" (UID: \"a4422dea-fe2c-463c-a3ef-ba129b466e52\") " pod="openstack/nova-api-0" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.504050 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4422dea-fe2c-463c-a3ef-ba129b466e52-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a4422dea-fe2c-463c-a3ef-ba129b466e52\") " pod="openstack/nova-api-0" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.504085 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rzpt\" (UniqueName: \"kubernetes.io/projected/a4422dea-fe2c-463c-a3ef-ba129b466e52-kube-api-access-9rzpt\") pod \"nova-api-0\" (UID: \"a4422dea-fe2c-463c-a3ef-ba129b466e52\") " pod="openstack/nova-api-0" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.505606 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4422dea-fe2c-463c-a3ef-ba129b466e52-logs\") pod \"nova-api-0\" (UID: \"a4422dea-fe2c-463c-a3ef-ba129b466e52\") " pod="openstack/nova-api-0" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.523570 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4422dea-fe2c-463c-a3ef-ba129b466e52-config-data\") pod \"nova-api-0\" (UID: \"a4422dea-fe2c-463c-a3ef-ba129b466e52\") " pod="openstack/nova-api-0" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.537244 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4422dea-fe2c-463c-a3ef-ba129b466e52-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a4422dea-fe2c-463c-a3ef-ba129b466e52\") " pod="openstack/nova-api-0" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.548271 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rzpt\" (UniqueName: \"kubernetes.io/projected/a4422dea-fe2c-463c-a3ef-ba129b466e52-kube-api-access-9rzpt\") pod \"nova-api-0\" (UID: \"a4422dea-fe2c-463c-a3ef-ba129b466e52\") " pod="openstack/nova-api-0" Dec 05 07:21:52 crc kubenswrapper[4997]: I1205 07:21:52.673292 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 07:21:53 crc kubenswrapper[4997]: I1205 07:21:53.208417 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:21:53 crc kubenswrapper[4997]: I1205 07:21:53.761919 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7de6ebed-df18-4907-ba33-3f4674373b43" path="/var/lib/kubelet/pods/7de6ebed-df18-4907-ba33-3f4674373b43/volumes" Dec 05 07:21:54 crc kubenswrapper[4997]: I1205 07:21:54.017291 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4422dea-fe2c-463c-a3ef-ba129b466e52","Type":"ContainerStarted","Data":"474f86cd2d7e52d69336a5aea4dcce8edaa23f3a6af211bedfe7cf096e48832e"} Dec 05 07:21:54 crc kubenswrapper[4997]: I1205 07:21:54.017363 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4422dea-fe2c-463c-a3ef-ba129b466e52","Type":"ContainerStarted","Data":"87256845007f9a3d1996b5209d06e27d72eafe8d7d53c8abfa8abb0a6cbecdb4"} Dec 05 07:21:54 crc kubenswrapper[4997]: I1205 07:21:54.017381 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4422dea-fe2c-463c-a3ef-ba129b466e52","Type":"ContainerStarted","Data":"fc2757f45936a947a04a0a871179632269771e4b259fb37681eb5012e799370b"} Dec 05 07:21:54 crc kubenswrapper[4997]: I1205 07:21:54.050247 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.05020388 podStartE2EDuration="2.05020388s" podCreationTimestamp="2025-12-05 07:21:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:21:54.038216735 +0000 UTC m=+1614.567124086" watchObservedRunningTime="2025-12-05 07:21:54.05020388 +0000 UTC m=+1614.579111181" Dec 05 07:21:55 crc kubenswrapper[4997]: I1205 07:21:55.470930 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 07:21:55 crc kubenswrapper[4997]: I1205 07:21:55.666896 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 07:21:55 crc kubenswrapper[4997]: I1205 07:21:55.666957 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 07:21:59 crc kubenswrapper[4997]: I1205 07:21:59.522489 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 05 07:22:00 crc kubenswrapper[4997]: I1205 07:22:00.470961 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 07:22:00 crc kubenswrapper[4997]: I1205 07:22:00.500634 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 07:22:00 crc kubenswrapper[4997]: I1205 07:22:00.667007 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 07:22:00 crc kubenswrapper[4997]: I1205 07:22:00.667108 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 07:22:01 crc kubenswrapper[4997]: I1205 07:22:01.129739 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 07:22:01 crc kubenswrapper[4997]: I1205 07:22:01.675314 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ef4a9d91-c9b1-4157-b6e5-485063844959" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 07:22:01 crc kubenswrapper[4997]: I1205 07:22:01.675310 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ef4a9d91-c9b1-4157-b6e5-485063844959" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 07:22:02 crc kubenswrapper[4997]: I1205 07:22:02.674232 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 07:22:02 crc kubenswrapper[4997]: I1205 07:22:02.674756 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 07:22:03 crc kubenswrapper[4997]: I1205 07:22:03.761372 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a4422dea-fe2c-463c-a3ef-ba129b466e52" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 07:22:03 crc kubenswrapper[4997]: I1205 07:22:03.761400 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a4422dea-fe2c-463c-a3ef-ba129b466e52" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 07:22:07 crc kubenswrapper[4997]: I1205 07:22:07.097355 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tdf5t"] Dec 05 07:22:07 crc kubenswrapper[4997]: I1205 07:22:07.100807 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdf5t" Dec 05 07:22:07 crc kubenswrapper[4997]: I1205 07:22:07.135902 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tdf5t"] Dec 05 07:22:07 crc kubenswrapper[4997]: I1205 07:22:07.164972 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/370b26bd-ba35-4485-a022-613df8a920f9-catalog-content\") pod \"community-operators-tdf5t\" (UID: \"370b26bd-ba35-4485-a022-613df8a920f9\") " pod="openshift-marketplace/community-operators-tdf5t" Dec 05 07:22:07 crc kubenswrapper[4997]: I1205 07:22:07.165036 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/370b26bd-ba35-4485-a022-613df8a920f9-utilities\") pod \"community-operators-tdf5t\" (UID: \"370b26bd-ba35-4485-a022-613df8a920f9\") " pod="openshift-marketplace/community-operators-tdf5t" Dec 05 07:22:07 crc kubenswrapper[4997]: I1205 07:22:07.165116 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw4ts\" (UniqueName: \"kubernetes.io/projected/370b26bd-ba35-4485-a022-613df8a920f9-kube-api-access-kw4ts\") pod \"community-operators-tdf5t\" (UID: \"370b26bd-ba35-4485-a022-613df8a920f9\") " pod="openshift-marketplace/community-operators-tdf5t" Dec 05 07:22:07 crc kubenswrapper[4997]: I1205 07:22:07.267885 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw4ts\" (UniqueName: \"kubernetes.io/projected/370b26bd-ba35-4485-a022-613df8a920f9-kube-api-access-kw4ts\") pod \"community-operators-tdf5t\" (UID: \"370b26bd-ba35-4485-a022-613df8a920f9\") " pod="openshift-marketplace/community-operators-tdf5t" Dec 05 07:22:07 crc kubenswrapper[4997]: I1205 07:22:07.269529 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/370b26bd-ba35-4485-a022-613df8a920f9-catalog-content\") pod \"community-operators-tdf5t\" (UID: \"370b26bd-ba35-4485-a022-613df8a920f9\") " pod="openshift-marketplace/community-operators-tdf5t" Dec 05 07:22:07 crc kubenswrapper[4997]: I1205 07:22:07.269744 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/370b26bd-ba35-4485-a022-613df8a920f9-utilities\") pod \"community-operators-tdf5t\" (UID: \"370b26bd-ba35-4485-a022-613df8a920f9\") " pod="openshift-marketplace/community-operators-tdf5t" Dec 05 07:22:07 crc kubenswrapper[4997]: I1205 07:22:07.270195 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/370b26bd-ba35-4485-a022-613df8a920f9-catalog-content\") pod \"community-operators-tdf5t\" (UID: \"370b26bd-ba35-4485-a022-613df8a920f9\") " pod="openshift-marketplace/community-operators-tdf5t" Dec 05 07:22:07 crc kubenswrapper[4997]: I1205 07:22:07.270313 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/370b26bd-ba35-4485-a022-613df8a920f9-utilities\") pod \"community-operators-tdf5t\" (UID: \"370b26bd-ba35-4485-a022-613df8a920f9\") " pod="openshift-marketplace/community-operators-tdf5t" Dec 05 07:22:07 crc kubenswrapper[4997]: I1205 07:22:07.292879 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw4ts\" (UniqueName: \"kubernetes.io/projected/370b26bd-ba35-4485-a022-613df8a920f9-kube-api-access-kw4ts\") pod \"community-operators-tdf5t\" (UID: \"370b26bd-ba35-4485-a022-613df8a920f9\") " pod="openshift-marketplace/community-operators-tdf5t" Dec 05 07:22:07 crc kubenswrapper[4997]: I1205 07:22:07.434460 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdf5t" Dec 05 07:22:08 crc kubenswrapper[4997]: I1205 07:22:08.049730 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tdf5t"] Dec 05 07:22:08 crc kubenswrapper[4997]: I1205 07:22:08.177499 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdf5t" event={"ID":"370b26bd-ba35-4485-a022-613df8a920f9","Type":"ContainerStarted","Data":"d47f8885cb4a9d73348004bfc5fd6eda1beb5adb6d7f2449a60daaf2b478542d"} Dec 05 07:22:09 crc kubenswrapper[4997]: I1205 07:22:09.193077 4997 generic.go:334] "Generic (PLEG): container finished" podID="370b26bd-ba35-4485-a022-613df8a920f9" containerID="1b0e4c587dd686a6a66d69afde2560006d6cea7e7c5fd970a06a0cdc446cdf37" exitCode=0 Dec 05 07:22:09 crc kubenswrapper[4997]: I1205 07:22:09.193180 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdf5t" event={"ID":"370b26bd-ba35-4485-a022-613df8a920f9","Type":"ContainerDied","Data":"1b0e4c587dd686a6a66d69afde2560006d6cea7e7c5fd970a06a0cdc446cdf37"} Dec 05 07:22:10 crc kubenswrapper[4997]: I1205 07:22:10.673748 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 07:22:10 crc kubenswrapper[4997]: I1205 07:22:10.676124 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 07:22:10 crc kubenswrapper[4997]: I1205 07:22:10.682934 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 07:22:10 crc kubenswrapper[4997]: I1205 07:22:10.684682 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.182854 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.234071 4997 generic.go:334] "Generic (PLEG): container finished" podID="18d92859-2d10-43da-a8f2-64cdd0f4ada6" containerID="0eeef8fbc295733a014cd992a3da5e4c4d7155017de2751ca5933ede672d5eab" exitCode=137 Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.235274 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.235299 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"18d92859-2d10-43da-a8f2-64cdd0f4ada6","Type":"ContainerDied","Data":"0eeef8fbc295733a014cd992a3da5e4c4d7155017de2751ca5933ede672d5eab"} Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.235354 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"18d92859-2d10-43da-a8f2-64cdd0f4ada6","Type":"ContainerDied","Data":"304f79cb6bc4e772c2d902f2a1a65d629cf887f5d4cf5ea87918efb30859678a"} Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.235376 4997 scope.go:117] "RemoveContainer" containerID="0eeef8fbc295733a014cd992a3da5e4c4d7155017de2751ca5933ede672d5eab" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.276524 4997 scope.go:117] "RemoveContainer" containerID="0eeef8fbc295733a014cd992a3da5e4c4d7155017de2751ca5933ede672d5eab" Dec 05 07:22:11 crc kubenswrapper[4997]: E1205 07:22:11.278902 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eeef8fbc295733a014cd992a3da5e4c4d7155017de2751ca5933ede672d5eab\": container with ID starting with 0eeef8fbc295733a014cd992a3da5e4c4d7155017de2751ca5933ede672d5eab not found: ID does not exist" containerID="0eeef8fbc295733a014cd992a3da5e4c4d7155017de2751ca5933ede672d5eab" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.278986 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eeef8fbc295733a014cd992a3da5e4c4d7155017de2751ca5933ede672d5eab"} err="failed to get container status \"0eeef8fbc295733a014cd992a3da5e4c4d7155017de2751ca5933ede672d5eab\": rpc error: code = NotFound desc = could not find container \"0eeef8fbc295733a014cd992a3da5e4c4d7155017de2751ca5933ede672d5eab\": container with ID starting with 0eeef8fbc295733a014cd992a3da5e4c4d7155017de2751ca5933ede672d5eab not found: ID does not exist" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.287506 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d92859-2d10-43da-a8f2-64cdd0f4ada6-combined-ca-bundle\") pod \"18d92859-2d10-43da-a8f2-64cdd0f4ada6\" (UID: \"18d92859-2d10-43da-a8f2-64cdd0f4ada6\") " Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.287961 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18d92859-2d10-43da-a8f2-64cdd0f4ada6-config-data\") pod \"18d92859-2d10-43da-a8f2-64cdd0f4ada6\" (UID: \"18d92859-2d10-43da-a8f2-64cdd0f4ada6\") " Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.288352 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78tzh\" (UniqueName: \"kubernetes.io/projected/18d92859-2d10-43da-a8f2-64cdd0f4ada6-kube-api-access-78tzh\") pod \"18d92859-2d10-43da-a8f2-64cdd0f4ada6\" (UID: \"18d92859-2d10-43da-a8f2-64cdd0f4ada6\") " Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.294663 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18d92859-2d10-43da-a8f2-64cdd0f4ada6-kube-api-access-78tzh" (OuterVolumeSpecName: "kube-api-access-78tzh") pod "18d92859-2d10-43da-a8f2-64cdd0f4ada6" (UID: "18d92859-2d10-43da-a8f2-64cdd0f4ada6"). InnerVolumeSpecName "kube-api-access-78tzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.324609 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18d92859-2d10-43da-a8f2-64cdd0f4ada6-config-data" (OuterVolumeSpecName: "config-data") pod "18d92859-2d10-43da-a8f2-64cdd0f4ada6" (UID: "18d92859-2d10-43da-a8f2-64cdd0f4ada6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.330855 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18d92859-2d10-43da-a8f2-64cdd0f4ada6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18d92859-2d10-43da-a8f2-64cdd0f4ada6" (UID: "18d92859-2d10-43da-a8f2-64cdd0f4ada6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.392930 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d92859-2d10-43da-a8f2-64cdd0f4ada6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.392979 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18d92859-2d10-43da-a8f2-64cdd0f4ada6-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.392992 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78tzh\" (UniqueName: \"kubernetes.io/projected/18d92859-2d10-43da-a8f2-64cdd0f4ada6-kube-api-access-78tzh\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.432335 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.584093 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.600383 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.612791 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 07:22:11 crc kubenswrapper[4997]: E1205 07:22:11.613446 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18d92859-2d10-43da-a8f2-64cdd0f4ada6" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.613468 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="18d92859-2d10-43da-a8f2-64cdd0f4ada6" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.613775 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="18d92859-2d10-43da-a8f2-64cdd0f4ada6" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.616104 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.621064 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.621310 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.621474 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.621713 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.703712 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.703790 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs6jl\" (UniqueName: \"kubernetes.io/projected/0abba971-b935-4cb0-865e-b364f8521f81-kube-api-access-gs6jl\") pod \"nova-cell1-novncproxy-0\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.703864 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.703911 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.703938 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.765263 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18d92859-2d10-43da-a8f2-64cdd0f4ada6" path="/var/lib/kubelet/pods/18d92859-2d10-43da-a8f2-64cdd0f4ada6/volumes" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.805769 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.805926 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.805998 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs6jl\" (UniqueName: \"kubernetes.io/projected/0abba971-b935-4cb0-865e-b364f8521f81-kube-api-access-gs6jl\") pod \"nova-cell1-novncproxy-0\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.806097 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.806172 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.810581 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.811540 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.811653 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.812953 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.844773 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs6jl\" (UniqueName: \"kubernetes.io/projected/0abba971-b935-4cb0-865e-b364f8521f81-kube-api-access-gs6jl\") pod \"nova-cell1-novncproxy-0\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:11 crc kubenswrapper[4997]: I1205 07:22:11.937258 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:12 crc kubenswrapper[4997]: I1205 07:22:12.677952 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 07:22:12 crc kubenswrapper[4997]: I1205 07:22:12.678865 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 07:22:12 crc kubenswrapper[4997]: I1205 07:22:12.680495 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 07:22:12 crc kubenswrapper[4997]: I1205 07:22:12.682717 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.266973 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.272978 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.487718 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-d6ftx"] Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.490172 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.492131 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-config\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.492179 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-ovsdbserver-sb\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.492212 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-dns-swift-storage-0\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.492254 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-ovsdbserver-nb\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.492319 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkgmc\" (UniqueName: \"kubernetes.io/projected/def8f706-0e1b-43ab-8d4e-2cab420b795d-kube-api-access-qkgmc\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.492352 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-dns-svc\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.542690 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-d6ftx"] Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.594625 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-ovsdbserver-nb\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.594753 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkgmc\" (UniqueName: \"kubernetes.io/projected/def8f706-0e1b-43ab-8d4e-2cab420b795d-kube-api-access-qkgmc\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.594794 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-dns-svc\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.594823 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-config\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.594859 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-ovsdbserver-sb\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.594890 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-dns-swift-storage-0\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.595974 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-dns-swift-storage-0\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.596589 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-ovsdbserver-nb\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.597182 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-config\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.597796 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-ovsdbserver-sb\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.598139 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-dns-svc\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.647984 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkgmc\" (UniqueName: \"kubernetes.io/projected/def8f706-0e1b-43ab-8d4e-2cab420b795d-kube-api-access-qkgmc\") pod \"dnsmasq-dns-cf4ff87b5-d6ftx\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:13 crc kubenswrapper[4997]: I1205 07:22:13.891923 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:14 crc kubenswrapper[4997]: I1205 07:22:14.303893 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 07:22:14 crc kubenswrapper[4997]: I1205 07:22:14.597894 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-d6ftx"] Dec 05 07:22:15 crc kubenswrapper[4997]: I1205 07:22:15.318759 4997 generic.go:334] "Generic (PLEG): container finished" podID="def8f706-0e1b-43ab-8d4e-2cab420b795d" containerID="10c39b4960112188aefd1d6e4212dc87504830944ea889c17f95beb408918e4c" exitCode=0 Dec 05 07:22:15 crc kubenswrapper[4997]: I1205 07:22:15.318967 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" event={"ID":"def8f706-0e1b-43ab-8d4e-2cab420b795d","Type":"ContainerDied","Data":"10c39b4960112188aefd1d6e4212dc87504830944ea889c17f95beb408918e4c"} Dec 05 07:22:15 crc kubenswrapper[4997]: I1205 07:22:15.319282 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" event={"ID":"def8f706-0e1b-43ab-8d4e-2cab420b795d","Type":"ContainerStarted","Data":"3d905d5b288a9f67b4075ad27d33434d8375b6a96e22194b534bb2fcbc98e05e"} Dec 05 07:22:15 crc kubenswrapper[4997]: I1205 07:22:15.326098 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0abba971-b935-4cb0-865e-b364f8521f81","Type":"ContainerStarted","Data":"d4ab663388d68e77285b0ecaf1e1e3e7993510fd57825fc77051658b8edf094e"} Dec 05 07:22:15 crc kubenswrapper[4997]: I1205 07:22:15.326339 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0abba971-b935-4cb0-865e-b364f8521f81","Type":"ContainerStarted","Data":"40db7048fdd40f1c31e882668f3781cab1f076430d89705b142af270080d6e41"} Dec 05 07:22:15 crc kubenswrapper[4997]: I1205 07:22:15.329012 4997 generic.go:334] "Generic (PLEG): container finished" podID="370b26bd-ba35-4485-a022-613df8a920f9" containerID="916c435a120557eb40452b62dde4f95b3fe9469ac55cc394c154cf5746df8ac6" exitCode=0 Dec 05 07:22:15 crc kubenswrapper[4997]: I1205 07:22:15.330560 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdf5t" event={"ID":"370b26bd-ba35-4485-a022-613df8a920f9","Type":"ContainerDied","Data":"916c435a120557eb40452b62dde4f95b3fe9469ac55cc394c154cf5746df8ac6"} Dec 05 07:22:15 crc kubenswrapper[4997]: I1205 07:22:15.398050 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=4.39802959 podStartE2EDuration="4.39802959s" podCreationTimestamp="2025-12-05 07:22:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:22:15.397035713 +0000 UTC m=+1635.925942964" watchObservedRunningTime="2025-12-05 07:22:15.39802959 +0000 UTC m=+1635.926936851" Dec 05 07:22:15 crc kubenswrapper[4997]: I1205 07:22:15.707794 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:22:15 crc kubenswrapper[4997]: I1205 07:22:15.708180 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerName="ceilometer-central-agent" containerID="cri-o://a7bd73ccb82c71dadb40dad6d9c0c99ca61d76861fb2cb4dfade74f8b1456580" gracePeriod=30 Dec 05 07:22:15 crc kubenswrapper[4997]: I1205 07:22:15.708983 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerName="sg-core" containerID="cri-o://4d5258220aa8b4d6ed169649576d64364b57d578c9fff38f0952dbdcdf27a945" gracePeriod=30 Dec 05 07:22:15 crc kubenswrapper[4997]: I1205 07:22:15.709000 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerName="proxy-httpd" containerID="cri-o://3fdc67df65aa7ef91669a32b7905e294fd97b1083008a1f6e7d196d58e703279" gracePeriod=30 Dec 05 07:22:15 crc kubenswrapper[4997]: I1205 07:22:15.709058 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerName="ceilometer-notification-agent" containerID="cri-o://b01ea2023bffc9dd49bdaf3a6b38449c2ab85e91386ff4a605523494c7a5cba3" gracePeriod=30 Dec 05 07:22:15 crc kubenswrapper[4997]: I1205 07:22:15.979806 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:22:16 crc kubenswrapper[4997]: I1205 07:22:16.342381 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" event={"ID":"def8f706-0e1b-43ab-8d4e-2cab420b795d","Type":"ContainerStarted","Data":"2ac514569d94e75e167d03cc51b57c157ef81c1e5e086d9812c749f5a5524ef7"} Dec 05 07:22:16 crc kubenswrapper[4997]: I1205 07:22:16.342569 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:16 crc kubenswrapper[4997]: I1205 07:22:16.348180 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdf5t" event={"ID":"370b26bd-ba35-4485-a022-613df8a920f9","Type":"ContainerStarted","Data":"e53254371f35aed52cd92c866695ce57e9363ca5fb3be536f5c33be9a92b3457"} Dec 05 07:22:16 crc kubenswrapper[4997]: I1205 07:22:16.354186 4997 generic.go:334] "Generic (PLEG): container finished" podID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerID="3fdc67df65aa7ef91669a32b7905e294fd97b1083008a1f6e7d196d58e703279" exitCode=0 Dec 05 07:22:16 crc kubenswrapper[4997]: I1205 07:22:16.354228 4997 generic.go:334] "Generic (PLEG): container finished" podID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerID="4d5258220aa8b4d6ed169649576d64364b57d578c9fff38f0952dbdcdf27a945" exitCode=2 Dec 05 07:22:16 crc kubenswrapper[4997]: I1205 07:22:16.354240 4997 generic.go:334] "Generic (PLEG): container finished" podID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerID="a7bd73ccb82c71dadb40dad6d9c0c99ca61d76861fb2cb4dfade74f8b1456580" exitCode=0 Dec 05 07:22:16 crc kubenswrapper[4997]: I1205 07:22:16.354264 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14ae6221-a8c8-43b0-ac51-2a592e9dd726","Type":"ContainerDied","Data":"3fdc67df65aa7ef91669a32b7905e294fd97b1083008a1f6e7d196d58e703279"} Dec 05 07:22:16 crc kubenswrapper[4997]: I1205 07:22:16.354308 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14ae6221-a8c8-43b0-ac51-2a592e9dd726","Type":"ContainerDied","Data":"4d5258220aa8b4d6ed169649576d64364b57d578c9fff38f0952dbdcdf27a945"} Dec 05 07:22:16 crc kubenswrapper[4997]: I1205 07:22:16.354323 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14ae6221-a8c8-43b0-ac51-2a592e9dd726","Type":"ContainerDied","Data":"a7bd73ccb82c71dadb40dad6d9c0c99ca61d76861fb2cb4dfade74f8b1456580"} Dec 05 07:22:16 crc kubenswrapper[4997]: I1205 07:22:16.354555 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a4422dea-fe2c-463c-a3ef-ba129b466e52" containerName="nova-api-log" containerID="cri-o://87256845007f9a3d1996b5209d06e27d72eafe8d7d53c8abfa8abb0a6cbecdb4" gracePeriod=30 Dec 05 07:22:16 crc kubenswrapper[4997]: I1205 07:22:16.354821 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a4422dea-fe2c-463c-a3ef-ba129b466e52" containerName="nova-api-api" containerID="cri-o://474f86cd2d7e52d69336a5aea4dcce8edaa23f3a6af211bedfe7cf096e48832e" gracePeriod=30 Dec 05 07:22:16 crc kubenswrapper[4997]: I1205 07:22:16.375913 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" podStartSLOduration=3.375876862 podStartE2EDuration="3.375876862s" podCreationTimestamp="2025-12-05 07:22:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:22:16.365798319 +0000 UTC m=+1636.894705590" watchObservedRunningTime="2025-12-05 07:22:16.375876862 +0000 UTC m=+1636.904784113" Dec 05 07:22:16 crc kubenswrapper[4997]: I1205 07:22:16.398878 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tdf5t" podStartSLOduration=2.845752604 podStartE2EDuration="9.398854684s" podCreationTimestamp="2025-12-05 07:22:07 +0000 UTC" firstStartedPulling="2025-12-05 07:22:09.196532583 +0000 UTC m=+1629.725439854" lastFinishedPulling="2025-12-05 07:22:15.749634673 +0000 UTC m=+1636.278541934" observedRunningTime="2025-12-05 07:22:16.388333479 +0000 UTC m=+1636.917240730" watchObservedRunningTime="2025-12-05 07:22:16.398854684 +0000 UTC m=+1636.927761955" Dec 05 07:22:16 crc kubenswrapper[4997]: I1205 07:22:16.822562 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:22:16 crc kubenswrapper[4997]: I1205 07:22:16.938308 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.008691 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-config-data\") pod \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.008795 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14ae6221-a8c8-43b0-ac51-2a592e9dd726-log-httpd\") pod \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.009022 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14ae6221-a8c8-43b0-ac51-2a592e9dd726-run-httpd\") pod \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.009076 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-sg-core-conf-yaml\") pod \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.009127 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-scripts\") pod \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.009199 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wkxp\" (UniqueName: \"kubernetes.io/projected/14ae6221-a8c8-43b0-ac51-2a592e9dd726-kube-api-access-9wkxp\") pod \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.009249 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-combined-ca-bundle\") pod \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\" (UID: \"14ae6221-a8c8-43b0-ac51-2a592e9dd726\") " Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.010304 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14ae6221-a8c8-43b0-ac51-2a592e9dd726-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "14ae6221-a8c8-43b0-ac51-2a592e9dd726" (UID: "14ae6221-a8c8-43b0-ac51-2a592e9dd726"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.010693 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14ae6221-a8c8-43b0-ac51-2a592e9dd726-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "14ae6221-a8c8-43b0-ac51-2a592e9dd726" (UID: "14ae6221-a8c8-43b0-ac51-2a592e9dd726"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.017833 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-scripts" (OuterVolumeSpecName: "scripts") pod "14ae6221-a8c8-43b0-ac51-2a592e9dd726" (UID: "14ae6221-a8c8-43b0-ac51-2a592e9dd726"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.024629 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14ae6221-a8c8-43b0-ac51-2a592e9dd726-kube-api-access-9wkxp" (OuterVolumeSpecName: "kube-api-access-9wkxp") pod "14ae6221-a8c8-43b0-ac51-2a592e9dd726" (UID: "14ae6221-a8c8-43b0-ac51-2a592e9dd726"). InnerVolumeSpecName "kube-api-access-9wkxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.052826 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "14ae6221-a8c8-43b0-ac51-2a592e9dd726" (UID: "14ae6221-a8c8-43b0-ac51-2a592e9dd726"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.111556 4997 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14ae6221-a8c8-43b0-ac51-2a592e9dd726-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.111594 4997 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.111603 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.111631 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wkxp\" (UniqueName: \"kubernetes.io/projected/14ae6221-a8c8-43b0-ac51-2a592e9dd726-kube-api-access-9wkxp\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.111644 4997 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14ae6221-a8c8-43b0-ac51-2a592e9dd726-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.127702 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14ae6221-a8c8-43b0-ac51-2a592e9dd726" (UID: "14ae6221-a8c8-43b0-ac51-2a592e9dd726"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.159441 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-config-data" (OuterVolumeSpecName: "config-data") pod "14ae6221-a8c8-43b0-ac51-2a592e9dd726" (UID: "14ae6221-a8c8-43b0-ac51-2a592e9dd726"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.213518 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.213562 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14ae6221-a8c8-43b0-ac51-2a592e9dd726-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.367559 4997 generic.go:334] "Generic (PLEG): container finished" podID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerID="b01ea2023bffc9dd49bdaf3a6b38449c2ab85e91386ff4a605523494c7a5cba3" exitCode=0 Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.367620 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.367640 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14ae6221-a8c8-43b0-ac51-2a592e9dd726","Type":"ContainerDied","Data":"b01ea2023bffc9dd49bdaf3a6b38449c2ab85e91386ff4a605523494c7a5cba3"} Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.367955 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14ae6221-a8c8-43b0-ac51-2a592e9dd726","Type":"ContainerDied","Data":"3c418c0e77dad4a64e83851dc85bd355831038ca14c60bf2f8f83e3a8e83b28a"} Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.367977 4997 scope.go:117] "RemoveContainer" containerID="3fdc67df65aa7ef91669a32b7905e294fd97b1083008a1f6e7d196d58e703279" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.375175 4997 generic.go:334] "Generic (PLEG): container finished" podID="a4422dea-fe2c-463c-a3ef-ba129b466e52" containerID="87256845007f9a3d1996b5209d06e27d72eafe8d7d53c8abfa8abb0a6cbecdb4" exitCode=143 Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.376484 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4422dea-fe2c-463c-a3ef-ba129b466e52","Type":"ContainerDied","Data":"87256845007f9a3d1996b5209d06e27d72eafe8d7d53c8abfa8abb0a6cbecdb4"} Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.395511 4997 scope.go:117] "RemoveContainer" containerID="4d5258220aa8b4d6ed169649576d64364b57d578c9fff38f0952dbdcdf27a945" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.422400 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.430982 4997 scope.go:117] "RemoveContainer" containerID="b01ea2023bffc9dd49bdaf3a6b38449c2ab85e91386ff4a605523494c7a5cba3" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.433691 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.435602 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tdf5t" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.435725 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tdf5t" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.460482 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:22:17 crc kubenswrapper[4997]: E1205 07:22:17.461455 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerName="sg-core" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.461481 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerName="sg-core" Dec 05 07:22:17 crc kubenswrapper[4997]: E1205 07:22:17.461526 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerName="ceilometer-notification-agent" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.461546 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerName="ceilometer-notification-agent" Dec 05 07:22:17 crc kubenswrapper[4997]: E1205 07:22:17.462051 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerName="ceilometer-central-agent" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.462070 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerName="ceilometer-central-agent" Dec 05 07:22:17 crc kubenswrapper[4997]: E1205 07:22:17.462096 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerName="proxy-httpd" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.462104 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerName="proxy-httpd" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.462320 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerName="proxy-httpd" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.462353 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerName="ceilometer-central-agent" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.462363 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerName="ceilometer-notification-agent" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.462375 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" containerName="sg-core" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.470501 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.481938 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.482145 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.510824 4997 scope.go:117] "RemoveContainer" containerID="a7bd73ccb82c71dadb40dad6d9c0c99ca61d76861fb2cb4dfade74f8b1456580" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.512191 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.540757 4997 scope.go:117] "RemoveContainer" containerID="3fdc67df65aa7ef91669a32b7905e294fd97b1083008a1f6e7d196d58e703279" Dec 05 07:22:17 crc kubenswrapper[4997]: E1205 07:22:17.541347 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fdc67df65aa7ef91669a32b7905e294fd97b1083008a1f6e7d196d58e703279\": container with ID starting with 3fdc67df65aa7ef91669a32b7905e294fd97b1083008a1f6e7d196d58e703279 not found: ID does not exist" containerID="3fdc67df65aa7ef91669a32b7905e294fd97b1083008a1f6e7d196d58e703279" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.541386 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fdc67df65aa7ef91669a32b7905e294fd97b1083008a1f6e7d196d58e703279"} err="failed to get container status \"3fdc67df65aa7ef91669a32b7905e294fd97b1083008a1f6e7d196d58e703279\": rpc error: code = NotFound desc = could not find container \"3fdc67df65aa7ef91669a32b7905e294fd97b1083008a1f6e7d196d58e703279\": container with ID starting with 3fdc67df65aa7ef91669a32b7905e294fd97b1083008a1f6e7d196d58e703279 not found: ID does not exist" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.541418 4997 scope.go:117] "RemoveContainer" containerID="4d5258220aa8b4d6ed169649576d64364b57d578c9fff38f0952dbdcdf27a945" Dec 05 07:22:17 crc kubenswrapper[4997]: E1205 07:22:17.541897 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d5258220aa8b4d6ed169649576d64364b57d578c9fff38f0952dbdcdf27a945\": container with ID starting with 4d5258220aa8b4d6ed169649576d64364b57d578c9fff38f0952dbdcdf27a945 not found: ID does not exist" containerID="4d5258220aa8b4d6ed169649576d64364b57d578c9fff38f0952dbdcdf27a945" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.541928 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d5258220aa8b4d6ed169649576d64364b57d578c9fff38f0952dbdcdf27a945"} err="failed to get container status \"4d5258220aa8b4d6ed169649576d64364b57d578c9fff38f0952dbdcdf27a945\": rpc error: code = NotFound desc = could not find container \"4d5258220aa8b4d6ed169649576d64364b57d578c9fff38f0952dbdcdf27a945\": container with ID starting with 4d5258220aa8b4d6ed169649576d64364b57d578c9fff38f0952dbdcdf27a945 not found: ID does not exist" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.541950 4997 scope.go:117] "RemoveContainer" containerID="b01ea2023bffc9dd49bdaf3a6b38449c2ab85e91386ff4a605523494c7a5cba3" Dec 05 07:22:17 crc kubenswrapper[4997]: E1205 07:22:17.542344 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b01ea2023bffc9dd49bdaf3a6b38449c2ab85e91386ff4a605523494c7a5cba3\": container with ID starting with b01ea2023bffc9dd49bdaf3a6b38449c2ab85e91386ff4a605523494c7a5cba3 not found: ID does not exist" containerID="b01ea2023bffc9dd49bdaf3a6b38449c2ab85e91386ff4a605523494c7a5cba3" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.542366 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b01ea2023bffc9dd49bdaf3a6b38449c2ab85e91386ff4a605523494c7a5cba3"} err="failed to get container status \"b01ea2023bffc9dd49bdaf3a6b38449c2ab85e91386ff4a605523494c7a5cba3\": rpc error: code = NotFound desc = could not find container \"b01ea2023bffc9dd49bdaf3a6b38449c2ab85e91386ff4a605523494c7a5cba3\": container with ID starting with b01ea2023bffc9dd49bdaf3a6b38449c2ab85e91386ff4a605523494c7a5cba3 not found: ID does not exist" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.542382 4997 scope.go:117] "RemoveContainer" containerID="a7bd73ccb82c71dadb40dad6d9c0c99ca61d76861fb2cb4dfade74f8b1456580" Dec 05 07:22:17 crc kubenswrapper[4997]: E1205 07:22:17.542861 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7bd73ccb82c71dadb40dad6d9c0c99ca61d76861fb2cb4dfade74f8b1456580\": container with ID starting with a7bd73ccb82c71dadb40dad6d9c0c99ca61d76861fb2cb4dfade74f8b1456580 not found: ID does not exist" containerID="a7bd73ccb82c71dadb40dad6d9c0c99ca61d76861fb2cb4dfade74f8b1456580" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.542915 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7bd73ccb82c71dadb40dad6d9c0c99ca61d76861fb2cb4dfade74f8b1456580"} err="failed to get container status \"a7bd73ccb82c71dadb40dad6d9c0c99ca61d76861fb2cb4dfade74f8b1456580\": rpc error: code = NotFound desc = could not find container \"a7bd73ccb82c71dadb40dad6d9c0c99ca61d76861fb2cb4dfade74f8b1456580\": container with ID starting with a7bd73ccb82c71dadb40dad6d9c0c99ca61d76861fb2cb4dfade74f8b1456580 not found: ID does not exist" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.622515 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/352614ac-912f-4b13-9514-f788a0f89589-log-httpd\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.622580 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.622663 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc4w8\" (UniqueName: \"kubernetes.io/projected/352614ac-912f-4b13-9514-f788a0f89589-kube-api-access-mc4w8\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.622803 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.622901 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/352614ac-912f-4b13-9514-f788a0f89589-run-httpd\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.623168 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-scripts\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.623299 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-config-data\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.725407 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-config-data\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.725545 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/352614ac-912f-4b13-9514-f788a0f89589-log-httpd\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.725573 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.725607 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc4w8\" (UniqueName: \"kubernetes.io/projected/352614ac-912f-4b13-9514-f788a0f89589-kube-api-access-mc4w8\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.725661 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.725680 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/352614ac-912f-4b13-9514-f788a0f89589-run-httpd\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.725729 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-scripts\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.726906 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/352614ac-912f-4b13-9514-f788a0f89589-run-httpd\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.728121 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/352614ac-912f-4b13-9514-f788a0f89589-log-httpd\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.737477 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-scripts\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.738059 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.738267 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.739163 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-config-data\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.747461 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc4w8\" (UniqueName: \"kubernetes.io/projected/352614ac-912f-4b13-9514-f788a0f89589-kube-api-access-mc4w8\") pod \"ceilometer-0\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.765676 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14ae6221-a8c8-43b0-ac51-2a592e9dd726" path="/var/lib/kubelet/pods/14ae6221-a8c8-43b0-ac51-2a592e9dd726/volumes" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.809571 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:22:17 crc kubenswrapper[4997]: I1205 07:22:17.937629 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:22:18 crc kubenswrapper[4997]: I1205 07:22:18.347965 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:22:18 crc kubenswrapper[4997]: W1205 07:22:18.353478 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod352614ac_912f_4b13_9514_f788a0f89589.slice/crio-270535577c75f16df35eabadad54e216ac1bafed26d4102f8f6da768d249f7c1 WatchSource:0}: Error finding container 270535577c75f16df35eabadad54e216ac1bafed26d4102f8f6da768d249f7c1: Status 404 returned error can't find the container with id 270535577c75f16df35eabadad54e216ac1bafed26d4102f8f6da768d249f7c1 Dec 05 07:22:18 crc kubenswrapper[4997]: I1205 07:22:18.389837 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"352614ac-912f-4b13-9514-f788a0f89589","Type":"ContainerStarted","Data":"270535577c75f16df35eabadad54e216ac1bafed26d4102f8f6da768d249f7c1"} Dec 05 07:22:18 crc kubenswrapper[4997]: I1205 07:22:18.513721 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-tdf5t" podUID="370b26bd-ba35-4485-a022-613df8a920f9" containerName="registry-server" probeResult="failure" output=< Dec 05 07:22:18 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 07:22:18 crc kubenswrapper[4997]: > Dec 05 07:22:19 crc kubenswrapper[4997]: I1205 07:22:19.405917 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"352614ac-912f-4b13-9514-f788a0f89589","Type":"ContainerStarted","Data":"5931d5aa7ad475c1925182f59f17bb91d4318e87e39668b1b4ecbb1548c319db"} Dec 05 07:22:19 crc kubenswrapper[4997]: I1205 07:22:19.771885 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:22:19 crc kubenswrapper[4997]: I1205 07:22:19.772492 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.396402 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.453969 4997 generic.go:334] "Generic (PLEG): container finished" podID="a4422dea-fe2c-463c-a3ef-ba129b466e52" containerID="474f86cd2d7e52d69336a5aea4dcce8edaa23f3a6af211bedfe7cf096e48832e" exitCode=0 Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.454076 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4422dea-fe2c-463c-a3ef-ba129b466e52","Type":"ContainerDied","Data":"474f86cd2d7e52d69336a5aea4dcce8edaa23f3a6af211bedfe7cf096e48832e"} Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.454112 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4422dea-fe2c-463c-a3ef-ba129b466e52","Type":"ContainerDied","Data":"fc2757f45936a947a04a0a871179632269771e4b259fb37681eb5012e799370b"} Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.454131 4997 scope.go:117] "RemoveContainer" containerID="474f86cd2d7e52d69336a5aea4dcce8edaa23f3a6af211bedfe7cf096e48832e" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.454119 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.461304 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"352614ac-912f-4b13-9514-f788a0f89589","Type":"ContainerStarted","Data":"10aeaa1da4aa6a1002dd033d477aac68a6da6595b1e71d67ef005ebcc04f2d7d"} Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.514291 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4422dea-fe2c-463c-a3ef-ba129b466e52-combined-ca-bundle\") pod \"a4422dea-fe2c-463c-a3ef-ba129b466e52\" (UID: \"a4422dea-fe2c-463c-a3ef-ba129b466e52\") " Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.514867 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rzpt\" (UniqueName: \"kubernetes.io/projected/a4422dea-fe2c-463c-a3ef-ba129b466e52-kube-api-access-9rzpt\") pod \"a4422dea-fe2c-463c-a3ef-ba129b466e52\" (UID: \"a4422dea-fe2c-463c-a3ef-ba129b466e52\") " Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.514977 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4422dea-fe2c-463c-a3ef-ba129b466e52-logs\") pod \"a4422dea-fe2c-463c-a3ef-ba129b466e52\" (UID: \"a4422dea-fe2c-463c-a3ef-ba129b466e52\") " Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.515143 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4422dea-fe2c-463c-a3ef-ba129b466e52-config-data\") pod \"a4422dea-fe2c-463c-a3ef-ba129b466e52\" (UID: \"a4422dea-fe2c-463c-a3ef-ba129b466e52\") " Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.518373 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4422dea-fe2c-463c-a3ef-ba129b466e52-logs" (OuterVolumeSpecName: "logs") pod "a4422dea-fe2c-463c-a3ef-ba129b466e52" (UID: "a4422dea-fe2c-463c-a3ef-ba129b466e52"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.518557 4997 scope.go:117] "RemoveContainer" containerID="87256845007f9a3d1996b5209d06e27d72eafe8d7d53c8abfa8abb0a6cbecdb4" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.526513 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4422dea-fe2c-463c-a3ef-ba129b466e52-kube-api-access-9rzpt" (OuterVolumeSpecName: "kube-api-access-9rzpt") pod "a4422dea-fe2c-463c-a3ef-ba129b466e52" (UID: "a4422dea-fe2c-463c-a3ef-ba129b466e52"). InnerVolumeSpecName "kube-api-access-9rzpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.569901 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4422dea-fe2c-463c-a3ef-ba129b466e52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4422dea-fe2c-463c-a3ef-ba129b466e52" (UID: "a4422dea-fe2c-463c-a3ef-ba129b466e52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.583853 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4422dea-fe2c-463c-a3ef-ba129b466e52-config-data" (OuterVolumeSpecName: "config-data") pod "a4422dea-fe2c-463c-a3ef-ba129b466e52" (UID: "a4422dea-fe2c-463c-a3ef-ba129b466e52"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.621038 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4422dea-fe2c-463c-a3ef-ba129b466e52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.621082 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rzpt\" (UniqueName: \"kubernetes.io/projected/a4422dea-fe2c-463c-a3ef-ba129b466e52-kube-api-access-9rzpt\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.621099 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4422dea-fe2c-463c-a3ef-ba129b466e52-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.621110 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4422dea-fe2c-463c-a3ef-ba129b466e52-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.674910 4997 scope.go:117] "RemoveContainer" containerID="474f86cd2d7e52d69336a5aea4dcce8edaa23f3a6af211bedfe7cf096e48832e" Dec 05 07:22:21 crc kubenswrapper[4997]: E1205 07:22:21.680743 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"474f86cd2d7e52d69336a5aea4dcce8edaa23f3a6af211bedfe7cf096e48832e\": container with ID starting with 474f86cd2d7e52d69336a5aea4dcce8edaa23f3a6af211bedfe7cf096e48832e not found: ID does not exist" containerID="474f86cd2d7e52d69336a5aea4dcce8edaa23f3a6af211bedfe7cf096e48832e" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.680790 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"474f86cd2d7e52d69336a5aea4dcce8edaa23f3a6af211bedfe7cf096e48832e"} err="failed to get container status \"474f86cd2d7e52d69336a5aea4dcce8edaa23f3a6af211bedfe7cf096e48832e\": rpc error: code = NotFound desc = could not find container \"474f86cd2d7e52d69336a5aea4dcce8edaa23f3a6af211bedfe7cf096e48832e\": container with ID starting with 474f86cd2d7e52d69336a5aea4dcce8edaa23f3a6af211bedfe7cf096e48832e not found: ID does not exist" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.680818 4997 scope.go:117] "RemoveContainer" containerID="87256845007f9a3d1996b5209d06e27d72eafe8d7d53c8abfa8abb0a6cbecdb4" Dec 05 07:22:21 crc kubenswrapper[4997]: E1205 07:22:21.684030 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87256845007f9a3d1996b5209d06e27d72eafe8d7d53c8abfa8abb0a6cbecdb4\": container with ID starting with 87256845007f9a3d1996b5209d06e27d72eafe8d7d53c8abfa8abb0a6cbecdb4 not found: ID does not exist" containerID="87256845007f9a3d1996b5209d06e27d72eafe8d7d53c8abfa8abb0a6cbecdb4" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.684057 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87256845007f9a3d1996b5209d06e27d72eafe8d7d53c8abfa8abb0a6cbecdb4"} err="failed to get container status \"87256845007f9a3d1996b5209d06e27d72eafe8d7d53c8abfa8abb0a6cbecdb4\": rpc error: code = NotFound desc = could not find container \"87256845007f9a3d1996b5209d06e27d72eafe8d7d53c8abfa8abb0a6cbecdb4\": container with ID starting with 87256845007f9a3d1996b5209d06e27d72eafe8d7d53c8abfa8abb0a6cbecdb4 not found: ID does not exist" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.882422 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.917318 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.938203 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 07:22:21 crc kubenswrapper[4997]: E1205 07:22:21.938834 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4422dea-fe2c-463c-a3ef-ba129b466e52" containerName="nova-api-log" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.938872 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4422dea-fe2c-463c-a3ef-ba129b466e52" containerName="nova-api-log" Dec 05 07:22:21 crc kubenswrapper[4997]: E1205 07:22:21.938895 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4422dea-fe2c-463c-a3ef-ba129b466e52" containerName="nova-api-api" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.938903 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4422dea-fe2c-463c-a3ef-ba129b466e52" containerName="nova-api-api" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.939178 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4422dea-fe2c-463c-a3ef-ba129b466e52" containerName="nova-api-api" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.939200 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4422dea-fe2c-463c-a3ef-ba129b466e52" containerName="nova-api-log" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.940489 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.942703 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.944345 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.944593 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 05 07:22:21 crc kubenswrapper[4997]: I1205 07:22:21.946955 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.001496 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.034287 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-public-tls-certs\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.034419 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.034510 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrjs8\" (UniqueName: \"kubernetes.io/projected/6673ee19-b22c-4a39-a6ea-bca48d529b4b-kube-api-access-nrjs8\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.034579 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-config-data\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.034682 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.034778 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6673ee19-b22c-4a39-a6ea-bca48d529b4b-logs\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.137219 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.137298 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrjs8\" (UniqueName: \"kubernetes.io/projected/6673ee19-b22c-4a39-a6ea-bca48d529b4b-kube-api-access-nrjs8\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.137339 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-config-data\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.137376 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.137425 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6673ee19-b22c-4a39-a6ea-bca48d529b4b-logs\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.137477 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-public-tls-certs\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.138291 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6673ee19-b22c-4a39-a6ea-bca48d529b4b-logs\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.142476 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.145805 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-public-tls-certs\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.153338 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-config-data\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.157121 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.168951 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrjs8\" (UniqueName: \"kubernetes.io/projected/6673ee19-b22c-4a39-a6ea-bca48d529b4b-kube-api-access-nrjs8\") pod \"nova-api-0\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.235920 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.303553 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.484877 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"352614ac-912f-4b13-9514-f788a0f89589","Type":"ContainerStarted","Data":"e827918c95883cb2b3ecc50516a17e2ffd55d05557594d758fc729a0342bd3fa"} Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.514814 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.776825 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-dv6nr"] Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.779389 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dv6nr" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.784084 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.784445 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.789412 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dv6nr"] Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.866196 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fjhg\" (UniqueName: \"kubernetes.io/projected/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-kube-api-access-5fjhg\") pod \"nova-cell1-cell-mapping-dv6nr\" (UID: \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\") " pod="openstack/nova-cell1-cell-mapping-dv6nr" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.866315 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-config-data\") pod \"nova-cell1-cell-mapping-dv6nr\" (UID: \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\") " pod="openstack/nova-cell1-cell-mapping-dv6nr" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.866391 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dv6nr\" (UID: \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\") " pod="openstack/nova-cell1-cell-mapping-dv6nr" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.866433 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-scripts\") pod \"nova-cell1-cell-mapping-dv6nr\" (UID: \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\") " pod="openstack/nova-cell1-cell-mapping-dv6nr" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.968917 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fjhg\" (UniqueName: \"kubernetes.io/projected/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-kube-api-access-5fjhg\") pod \"nova-cell1-cell-mapping-dv6nr\" (UID: \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\") " pod="openstack/nova-cell1-cell-mapping-dv6nr" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.969013 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-config-data\") pod \"nova-cell1-cell-mapping-dv6nr\" (UID: \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\") " pod="openstack/nova-cell1-cell-mapping-dv6nr" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.969086 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dv6nr\" (UID: \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\") " pod="openstack/nova-cell1-cell-mapping-dv6nr" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.969124 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-scripts\") pod \"nova-cell1-cell-mapping-dv6nr\" (UID: \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\") " pod="openstack/nova-cell1-cell-mapping-dv6nr" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.979676 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-scripts\") pod \"nova-cell1-cell-mapping-dv6nr\" (UID: \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\") " pod="openstack/nova-cell1-cell-mapping-dv6nr" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.979684 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dv6nr\" (UID: \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\") " pod="openstack/nova-cell1-cell-mapping-dv6nr" Dec 05 07:22:22 crc kubenswrapper[4997]: I1205 07:22:22.996298 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-config-data\") pod \"nova-cell1-cell-mapping-dv6nr\" (UID: \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\") " pod="openstack/nova-cell1-cell-mapping-dv6nr" Dec 05 07:22:23 crc kubenswrapper[4997]: I1205 07:22:23.000287 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fjhg\" (UniqueName: \"kubernetes.io/projected/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-kube-api-access-5fjhg\") pod \"nova-cell1-cell-mapping-dv6nr\" (UID: \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\") " pod="openstack/nova-cell1-cell-mapping-dv6nr" Dec 05 07:22:23 crc kubenswrapper[4997]: I1205 07:22:23.087070 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:22:23 crc kubenswrapper[4997]: I1205 07:22:23.118968 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dv6nr" Dec 05 07:22:23 crc kubenswrapper[4997]: I1205 07:22:23.514214 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6673ee19-b22c-4a39-a6ea-bca48d529b4b","Type":"ContainerStarted","Data":"bb219d357250fdac5860af9df506f620cd51f00862c6b7449d8ab0db68297cec"} Dec 05 07:22:23 crc kubenswrapper[4997]: I1205 07:22:23.515404 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6673ee19-b22c-4a39-a6ea-bca48d529b4b","Type":"ContainerStarted","Data":"5d1a871a19941cac1e50e94b897d7c66dc3485e4396828918865a90bff5b6c1a"} Dec 05 07:22:23 crc kubenswrapper[4997]: I1205 07:22:23.519942 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"352614ac-912f-4b13-9514-f788a0f89589","Type":"ContainerStarted","Data":"3d59be687e2d5e878c40ab9457e37b0983ab15aec7ffb998fdb45103e0317a60"} Dec 05 07:22:23 crc kubenswrapper[4997]: I1205 07:22:23.520119 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="352614ac-912f-4b13-9514-f788a0f89589" containerName="ceilometer-central-agent" containerID="cri-o://5931d5aa7ad475c1925182f59f17bb91d4318e87e39668b1b4ecbb1548c319db" gracePeriod=30 Dec 05 07:22:23 crc kubenswrapper[4997]: I1205 07:22:23.520179 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 07:22:23 crc kubenswrapper[4997]: I1205 07:22:23.520248 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="352614ac-912f-4b13-9514-f788a0f89589" containerName="proxy-httpd" containerID="cri-o://3d59be687e2d5e878c40ab9457e37b0983ab15aec7ffb998fdb45103e0317a60" gracePeriod=30 Dec 05 07:22:23 crc kubenswrapper[4997]: I1205 07:22:23.520291 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="352614ac-912f-4b13-9514-f788a0f89589" containerName="ceilometer-notification-agent" containerID="cri-o://10aeaa1da4aa6a1002dd033d477aac68a6da6595b1e71d67ef005ebcc04f2d7d" gracePeriod=30 Dec 05 07:22:23 crc kubenswrapper[4997]: I1205 07:22:23.520399 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="352614ac-912f-4b13-9514-f788a0f89589" containerName="sg-core" containerID="cri-o://e827918c95883cb2b3ecc50516a17e2ffd55d05557594d758fc729a0342bd3fa" gracePeriod=30 Dec 05 07:22:23 crc kubenswrapper[4997]: I1205 07:22:23.579320 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.903309346 podStartE2EDuration="6.579298072s" podCreationTimestamp="2025-12-05 07:22:17 +0000 UTC" firstStartedPulling="2025-12-05 07:22:18.35685276 +0000 UTC m=+1638.885760021" lastFinishedPulling="2025-12-05 07:22:23.032841486 +0000 UTC m=+1643.561748747" observedRunningTime="2025-12-05 07:22:23.565520309 +0000 UTC m=+1644.094427570" watchObservedRunningTime="2025-12-05 07:22:23.579298072 +0000 UTC m=+1644.108205323" Dec 05 07:22:23 crc kubenswrapper[4997]: I1205 07:22:23.731979 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dv6nr"] Dec 05 07:22:23 crc kubenswrapper[4997]: W1205 07:22:23.733003 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff349aa8_88f5_4c3d_bc2a_f4a9d89ba5a5.slice/crio-718e0076d2aa47cef346ef827cfa83b521f5cefe861beed2a3aa65bcb0e7a181 WatchSource:0}: Error finding container 718e0076d2aa47cef346ef827cfa83b521f5cefe861beed2a3aa65bcb0e7a181: Status 404 returned error can't find the container with id 718e0076d2aa47cef346ef827cfa83b521f5cefe861beed2a3aa65bcb0e7a181 Dec 05 07:22:23 crc kubenswrapper[4997]: I1205 07:22:23.772883 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4422dea-fe2c-463c-a3ef-ba129b466e52" path="/var/lib/kubelet/pods/a4422dea-fe2c-463c-a3ef-ba129b466e52/volumes" Dec 05 07:22:23 crc kubenswrapper[4997]: I1205 07:22:23.893781 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:22:23 crc kubenswrapper[4997]: I1205 07:22:23.987455 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-xrxqk"] Dec 05 07:22:23 crc kubenswrapper[4997]: I1205 07:22:23.987812 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" podUID="09e38b0e-164a-4927-bbfa-87d8f2023a6c" containerName="dnsmasq-dns" containerID="cri-o://bd53b88774db5676c0ecbd766a894c0787029e64044e69e6c842f108a8b45f92" gracePeriod=10 Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.538315 4997 generic.go:334] "Generic (PLEG): container finished" podID="09e38b0e-164a-4927-bbfa-87d8f2023a6c" containerID="bd53b88774db5676c0ecbd766a894c0787029e64044e69e6c842f108a8b45f92" exitCode=0 Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.538419 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" event={"ID":"09e38b0e-164a-4927-bbfa-87d8f2023a6c","Type":"ContainerDied","Data":"bd53b88774db5676c0ecbd766a894c0787029e64044e69e6c842f108a8b45f92"} Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.538831 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" event={"ID":"09e38b0e-164a-4927-bbfa-87d8f2023a6c","Type":"ContainerDied","Data":"02f13bc024bad3e76377fb724c2468380b7712120a238ab81278b40d1c74aa55"} Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.538882 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02f13bc024bad3e76377fb724c2468380b7712120a238ab81278b40d1c74aa55" Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.546348 4997 generic.go:334] "Generic (PLEG): container finished" podID="352614ac-912f-4b13-9514-f788a0f89589" containerID="3d59be687e2d5e878c40ab9457e37b0983ab15aec7ffb998fdb45103e0317a60" exitCode=0 Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.546393 4997 generic.go:334] "Generic (PLEG): container finished" podID="352614ac-912f-4b13-9514-f788a0f89589" containerID="e827918c95883cb2b3ecc50516a17e2ffd55d05557594d758fc729a0342bd3fa" exitCode=2 Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.546406 4997 generic.go:334] "Generic (PLEG): container finished" podID="352614ac-912f-4b13-9514-f788a0f89589" containerID="10aeaa1da4aa6a1002dd033d477aac68a6da6595b1e71d67ef005ebcc04f2d7d" exitCode=0 Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.546419 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"352614ac-912f-4b13-9514-f788a0f89589","Type":"ContainerDied","Data":"3d59be687e2d5e878c40ab9457e37b0983ab15aec7ffb998fdb45103e0317a60"} Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.546472 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"352614ac-912f-4b13-9514-f788a0f89589","Type":"ContainerDied","Data":"e827918c95883cb2b3ecc50516a17e2ffd55d05557594d758fc729a0342bd3fa"} Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.546484 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"352614ac-912f-4b13-9514-f788a0f89589","Type":"ContainerDied","Data":"10aeaa1da4aa6a1002dd033d477aac68a6da6595b1e71d67ef005ebcc04f2d7d"} Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.549818 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6673ee19-b22c-4a39-a6ea-bca48d529b4b","Type":"ContainerStarted","Data":"7287af1f418f2880f43bdca10a8b67ade52a90451e4e7185ef7e78c5b3a333b3"} Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.553792 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dv6nr" event={"ID":"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5","Type":"ContainerStarted","Data":"834d649b80dd8fee860f1e640b0ea6a36d58f4fc6fd2c37271b4e7ce17650c8d"} Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.553871 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dv6nr" event={"ID":"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5","Type":"ContainerStarted","Data":"718e0076d2aa47cef346ef827cfa83b521f5cefe861beed2a3aa65bcb0e7a181"} Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.570684 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.580700 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.580678732 podStartE2EDuration="3.580678732s" podCreationTimestamp="2025-12-05 07:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:22:24.57506724 +0000 UTC m=+1645.103974521" watchObservedRunningTime="2025-12-05 07:22:24.580678732 +0000 UTC m=+1645.109585993" Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.711024 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-dv6nr" podStartSLOduration=2.7109959249999998 podStartE2EDuration="2.710995925s" podCreationTimestamp="2025-12-05 07:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:22:24.64184708 +0000 UTC m=+1645.170754361" watchObservedRunningTime="2025-12-05 07:22:24.710995925 +0000 UTC m=+1645.239903186" Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.715903 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-dns-svc\") pod \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.716084 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-ovsdbserver-sb\") pod \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.716322 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-ovsdbserver-nb\") pod \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.716362 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-config\") pod \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.716393 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-dns-swift-storage-0\") pod \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.716426 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkbfr\" (UniqueName: \"kubernetes.io/projected/09e38b0e-164a-4927-bbfa-87d8f2023a6c-kube-api-access-lkbfr\") pod \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\" (UID: \"09e38b0e-164a-4927-bbfa-87d8f2023a6c\") " Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.724370 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09e38b0e-164a-4927-bbfa-87d8f2023a6c-kube-api-access-lkbfr" (OuterVolumeSpecName: "kube-api-access-lkbfr") pod "09e38b0e-164a-4927-bbfa-87d8f2023a6c" (UID: "09e38b0e-164a-4927-bbfa-87d8f2023a6c"). InnerVolumeSpecName "kube-api-access-lkbfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.782327 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-config" (OuterVolumeSpecName: "config") pod "09e38b0e-164a-4927-bbfa-87d8f2023a6c" (UID: "09e38b0e-164a-4927-bbfa-87d8f2023a6c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.783256 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "09e38b0e-164a-4927-bbfa-87d8f2023a6c" (UID: "09e38b0e-164a-4927-bbfa-87d8f2023a6c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.788136 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "09e38b0e-164a-4927-bbfa-87d8f2023a6c" (UID: "09e38b0e-164a-4927-bbfa-87d8f2023a6c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.797144 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "09e38b0e-164a-4927-bbfa-87d8f2023a6c" (UID: "09e38b0e-164a-4927-bbfa-87d8f2023a6c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.820739 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.820782 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.820792 4997 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.820814 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkbfr\" (UniqueName: \"kubernetes.io/projected/09e38b0e-164a-4927-bbfa-87d8f2023a6c-kube-api-access-lkbfr\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.820827 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.849873 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "09e38b0e-164a-4927-bbfa-87d8f2023a6c" (UID: "09e38b0e-164a-4927-bbfa-87d8f2023a6c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:22:24 crc kubenswrapper[4997]: I1205 07:22:24.929159 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09e38b0e-164a-4927-bbfa-87d8f2023a6c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:25 crc kubenswrapper[4997]: I1205 07:22:25.563668 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" Dec 05 07:22:25 crc kubenswrapper[4997]: I1205 07:22:25.606727 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-xrxqk"] Dec 05 07:22:25 crc kubenswrapper[4997]: I1205 07:22:25.618026 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-xrxqk"] Dec 05 07:22:25 crc kubenswrapper[4997]: I1205 07:22:25.762724 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09e38b0e-164a-4927-bbfa-87d8f2023a6c" path="/var/lib/kubelet/pods/09e38b0e-164a-4927-bbfa-87d8f2023a6c/volumes" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.285162 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.384922 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/352614ac-912f-4b13-9514-f788a0f89589-run-httpd\") pod \"352614ac-912f-4b13-9514-f788a0f89589\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.385011 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-sg-core-conf-yaml\") pod \"352614ac-912f-4b13-9514-f788a0f89589\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.385068 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-combined-ca-bundle\") pod \"352614ac-912f-4b13-9514-f788a0f89589\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.385108 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/352614ac-912f-4b13-9514-f788a0f89589-log-httpd\") pod \"352614ac-912f-4b13-9514-f788a0f89589\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.385154 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-scripts\") pod \"352614ac-912f-4b13-9514-f788a0f89589\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.385239 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc4w8\" (UniqueName: \"kubernetes.io/projected/352614ac-912f-4b13-9514-f788a0f89589-kube-api-access-mc4w8\") pod \"352614ac-912f-4b13-9514-f788a0f89589\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.385322 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-config-data\") pod \"352614ac-912f-4b13-9514-f788a0f89589\" (UID: \"352614ac-912f-4b13-9514-f788a0f89589\") " Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.385524 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/352614ac-912f-4b13-9514-f788a0f89589-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "352614ac-912f-4b13-9514-f788a0f89589" (UID: "352614ac-912f-4b13-9514-f788a0f89589"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.385794 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/352614ac-912f-4b13-9514-f788a0f89589-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "352614ac-912f-4b13-9514-f788a0f89589" (UID: "352614ac-912f-4b13-9514-f788a0f89589"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.386686 4997 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/352614ac-912f-4b13-9514-f788a0f89589-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.386716 4997 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/352614ac-912f-4b13-9514-f788a0f89589-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.394511 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/352614ac-912f-4b13-9514-f788a0f89589-kube-api-access-mc4w8" (OuterVolumeSpecName: "kube-api-access-mc4w8") pod "352614ac-912f-4b13-9514-f788a0f89589" (UID: "352614ac-912f-4b13-9514-f788a0f89589"). InnerVolumeSpecName "kube-api-access-mc4w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.394597 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-scripts" (OuterVolumeSpecName: "scripts") pod "352614ac-912f-4b13-9514-f788a0f89589" (UID: "352614ac-912f-4b13-9514-f788a0f89589"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.424818 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "352614ac-912f-4b13-9514-f788a0f89589" (UID: "352614ac-912f-4b13-9514-f788a0f89589"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.489689 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "352614ac-912f-4b13-9514-f788a0f89589" (UID: "352614ac-912f-4b13-9514-f788a0f89589"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.489750 4997 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.490131 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.490198 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc4w8\" (UniqueName: \"kubernetes.io/projected/352614ac-912f-4b13-9514-f788a0f89589-kube-api-access-mc4w8\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.492413 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tdf5t" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.529282 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-config-data" (OuterVolumeSpecName: "config-data") pod "352614ac-912f-4b13-9514-f788a0f89589" (UID: "352614ac-912f-4b13-9514-f788a0f89589"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.561792 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tdf5t" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.592150 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.592193 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352614ac-912f-4b13-9514-f788a0f89589-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.593762 4997 generic.go:334] "Generic (PLEG): container finished" podID="352614ac-912f-4b13-9514-f788a0f89589" containerID="5931d5aa7ad475c1925182f59f17bb91d4318e87e39668b1b4ecbb1548c319db" exitCode=0 Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.593981 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.594060 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"352614ac-912f-4b13-9514-f788a0f89589","Type":"ContainerDied","Data":"5931d5aa7ad475c1925182f59f17bb91d4318e87e39668b1b4ecbb1548c319db"} Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.594111 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"352614ac-912f-4b13-9514-f788a0f89589","Type":"ContainerDied","Data":"270535577c75f16df35eabadad54e216ac1bafed26d4102f8f6da768d249f7c1"} Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.594136 4997 scope.go:117] "RemoveContainer" containerID="3d59be687e2d5e878c40ab9457e37b0983ab15aec7ffb998fdb45103e0317a60" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.637487 4997 scope.go:117] "RemoveContainer" containerID="e827918c95883cb2b3ecc50516a17e2ffd55d05557594d758fc729a0342bd3fa" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.646773 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.662652 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.686173 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tdf5t"] Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.702745 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:22:27 crc kubenswrapper[4997]: E1205 07:22:27.703359 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="352614ac-912f-4b13-9514-f788a0f89589" containerName="ceilometer-central-agent" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.703384 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="352614ac-912f-4b13-9514-f788a0f89589" containerName="ceilometer-central-agent" Dec 05 07:22:27 crc kubenswrapper[4997]: E1205 07:22:27.703401 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="352614ac-912f-4b13-9514-f788a0f89589" containerName="sg-core" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.703413 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="352614ac-912f-4b13-9514-f788a0f89589" containerName="sg-core" Dec 05 07:22:27 crc kubenswrapper[4997]: E1205 07:22:27.703429 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09e38b0e-164a-4927-bbfa-87d8f2023a6c" containerName="dnsmasq-dns" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.703437 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="09e38b0e-164a-4927-bbfa-87d8f2023a6c" containerName="dnsmasq-dns" Dec 05 07:22:27 crc kubenswrapper[4997]: E1205 07:22:27.703452 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="352614ac-912f-4b13-9514-f788a0f89589" containerName="ceilometer-notification-agent" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.703460 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="352614ac-912f-4b13-9514-f788a0f89589" containerName="ceilometer-notification-agent" Dec 05 07:22:27 crc kubenswrapper[4997]: E1205 07:22:27.703479 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09e38b0e-164a-4927-bbfa-87d8f2023a6c" containerName="init" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.703516 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="09e38b0e-164a-4927-bbfa-87d8f2023a6c" containerName="init" Dec 05 07:22:27 crc kubenswrapper[4997]: E1205 07:22:27.703531 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="352614ac-912f-4b13-9514-f788a0f89589" containerName="proxy-httpd" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.703538 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="352614ac-912f-4b13-9514-f788a0f89589" containerName="proxy-httpd" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.703782 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="352614ac-912f-4b13-9514-f788a0f89589" containerName="ceilometer-notification-agent" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.703796 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="09e38b0e-164a-4927-bbfa-87d8f2023a6c" containerName="dnsmasq-dns" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.703808 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="352614ac-912f-4b13-9514-f788a0f89589" containerName="ceilometer-central-agent" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.703818 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="352614ac-912f-4b13-9514-f788a0f89589" containerName="proxy-httpd" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.703840 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="352614ac-912f-4b13-9514-f788a0f89589" containerName="sg-core" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.705888 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.709236 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.709547 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.734559 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.736951 4997 scope.go:117] "RemoveContainer" containerID="10aeaa1da4aa6a1002dd033d477aac68a6da6595b1e71d67ef005ebcc04f2d7d" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.769096 4997 scope.go:117] "RemoveContainer" containerID="5931d5aa7ad475c1925182f59f17bb91d4318e87e39668b1b4ecbb1548c319db" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.785802 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="352614ac-912f-4b13-9514-f788a0f89589" path="/var/lib/kubelet/pods/352614ac-912f-4b13-9514-f788a0f89589/volumes" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.786754 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zdk9s"] Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.791757 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zdk9s" podUID="d3274e3c-2ad5-4756-8d3f-bae751f1e719" containerName="registry-server" containerID="cri-o://9c2afaa8dfee20860b91a43a1030662f1087186da9f0d30e7927675d98f0bf48" gracePeriod=2 Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.798827 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-config-data\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.799007 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-scripts\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.799162 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-run-httpd\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.799271 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-log-httpd\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.799363 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.799440 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.799541 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtwl6\" (UniqueName: \"kubernetes.io/projected/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-kube-api-access-xtwl6\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.810319 4997 scope.go:117] "RemoveContainer" containerID="3d59be687e2d5e878c40ab9457e37b0983ab15aec7ffb998fdb45103e0317a60" Dec 05 07:22:27 crc kubenswrapper[4997]: E1205 07:22:27.810878 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d59be687e2d5e878c40ab9457e37b0983ab15aec7ffb998fdb45103e0317a60\": container with ID starting with 3d59be687e2d5e878c40ab9457e37b0983ab15aec7ffb998fdb45103e0317a60 not found: ID does not exist" containerID="3d59be687e2d5e878c40ab9457e37b0983ab15aec7ffb998fdb45103e0317a60" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.810928 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d59be687e2d5e878c40ab9457e37b0983ab15aec7ffb998fdb45103e0317a60"} err="failed to get container status \"3d59be687e2d5e878c40ab9457e37b0983ab15aec7ffb998fdb45103e0317a60\": rpc error: code = NotFound desc = could not find container \"3d59be687e2d5e878c40ab9457e37b0983ab15aec7ffb998fdb45103e0317a60\": container with ID starting with 3d59be687e2d5e878c40ab9457e37b0983ab15aec7ffb998fdb45103e0317a60 not found: ID does not exist" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.810955 4997 scope.go:117] "RemoveContainer" containerID="e827918c95883cb2b3ecc50516a17e2ffd55d05557594d758fc729a0342bd3fa" Dec 05 07:22:27 crc kubenswrapper[4997]: E1205 07:22:27.811334 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e827918c95883cb2b3ecc50516a17e2ffd55d05557594d758fc729a0342bd3fa\": container with ID starting with e827918c95883cb2b3ecc50516a17e2ffd55d05557594d758fc729a0342bd3fa not found: ID does not exist" containerID="e827918c95883cb2b3ecc50516a17e2ffd55d05557594d758fc729a0342bd3fa" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.811369 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e827918c95883cb2b3ecc50516a17e2ffd55d05557594d758fc729a0342bd3fa"} err="failed to get container status \"e827918c95883cb2b3ecc50516a17e2ffd55d05557594d758fc729a0342bd3fa\": rpc error: code = NotFound desc = could not find container \"e827918c95883cb2b3ecc50516a17e2ffd55d05557594d758fc729a0342bd3fa\": container with ID starting with e827918c95883cb2b3ecc50516a17e2ffd55d05557594d758fc729a0342bd3fa not found: ID does not exist" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.811390 4997 scope.go:117] "RemoveContainer" containerID="10aeaa1da4aa6a1002dd033d477aac68a6da6595b1e71d67ef005ebcc04f2d7d" Dec 05 07:22:27 crc kubenswrapper[4997]: E1205 07:22:27.811658 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10aeaa1da4aa6a1002dd033d477aac68a6da6595b1e71d67ef005ebcc04f2d7d\": container with ID starting with 10aeaa1da4aa6a1002dd033d477aac68a6da6595b1e71d67ef005ebcc04f2d7d not found: ID does not exist" containerID="10aeaa1da4aa6a1002dd033d477aac68a6da6595b1e71d67ef005ebcc04f2d7d" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.811692 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10aeaa1da4aa6a1002dd033d477aac68a6da6595b1e71d67ef005ebcc04f2d7d"} err="failed to get container status \"10aeaa1da4aa6a1002dd033d477aac68a6da6595b1e71d67ef005ebcc04f2d7d\": rpc error: code = NotFound desc = could not find container \"10aeaa1da4aa6a1002dd033d477aac68a6da6595b1e71d67ef005ebcc04f2d7d\": container with ID starting with 10aeaa1da4aa6a1002dd033d477aac68a6da6595b1e71d67ef005ebcc04f2d7d not found: ID does not exist" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.811710 4997 scope.go:117] "RemoveContainer" containerID="5931d5aa7ad475c1925182f59f17bb91d4318e87e39668b1b4ecbb1548c319db" Dec 05 07:22:27 crc kubenswrapper[4997]: E1205 07:22:27.811978 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5931d5aa7ad475c1925182f59f17bb91d4318e87e39668b1b4ecbb1548c319db\": container with ID starting with 5931d5aa7ad475c1925182f59f17bb91d4318e87e39668b1b4ecbb1548c319db not found: ID does not exist" containerID="5931d5aa7ad475c1925182f59f17bb91d4318e87e39668b1b4ecbb1548c319db" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.812002 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5931d5aa7ad475c1925182f59f17bb91d4318e87e39668b1b4ecbb1548c319db"} err="failed to get container status \"5931d5aa7ad475c1925182f59f17bb91d4318e87e39668b1b4ecbb1548c319db\": rpc error: code = NotFound desc = could not find container \"5931d5aa7ad475c1925182f59f17bb91d4318e87e39668b1b4ecbb1548c319db\": container with ID starting with 5931d5aa7ad475c1925182f59f17bb91d4318e87e39668b1b4ecbb1548c319db not found: ID does not exist" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.901522 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-run-httpd\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.901949 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-log-httpd\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.901973 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.902020 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.902046 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtwl6\" (UniqueName: \"kubernetes.io/projected/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-kube-api-access-xtwl6\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.902241 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-config-data\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.902269 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-scripts\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.904150 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-run-httpd\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.904835 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-log-httpd\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.914987 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.921499 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.921518 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-scripts\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.921829 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-config-data\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:27 crc kubenswrapper[4997]: I1205 07:22:27.926994 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtwl6\" (UniqueName: \"kubernetes.io/projected/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-kube-api-access-xtwl6\") pod \"ceilometer-0\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " pod="openstack/ceilometer-0" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.050746 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.340960 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zdk9s" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.411723 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3274e3c-2ad5-4756-8d3f-bae751f1e719-utilities\") pod \"d3274e3c-2ad5-4756-8d3f-bae751f1e719\" (UID: \"d3274e3c-2ad5-4756-8d3f-bae751f1e719\") " Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.411777 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3274e3c-2ad5-4756-8d3f-bae751f1e719-catalog-content\") pod \"d3274e3c-2ad5-4756-8d3f-bae751f1e719\" (UID: \"d3274e3c-2ad5-4756-8d3f-bae751f1e719\") " Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.411919 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rt55\" (UniqueName: \"kubernetes.io/projected/d3274e3c-2ad5-4756-8d3f-bae751f1e719-kube-api-access-9rt55\") pod \"d3274e3c-2ad5-4756-8d3f-bae751f1e719\" (UID: \"d3274e3c-2ad5-4756-8d3f-bae751f1e719\") " Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.413202 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3274e3c-2ad5-4756-8d3f-bae751f1e719-utilities" (OuterVolumeSpecName: "utilities") pod "d3274e3c-2ad5-4756-8d3f-bae751f1e719" (UID: "d3274e3c-2ad5-4756-8d3f-bae751f1e719"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.419055 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3274e3c-2ad5-4756-8d3f-bae751f1e719-kube-api-access-9rt55" (OuterVolumeSpecName: "kube-api-access-9rt55") pod "d3274e3c-2ad5-4756-8d3f-bae751f1e719" (UID: "d3274e3c-2ad5-4756-8d3f-bae751f1e719"). InnerVolumeSpecName "kube-api-access-9rt55". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.481843 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3274e3c-2ad5-4756-8d3f-bae751f1e719-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3274e3c-2ad5-4756-8d3f-bae751f1e719" (UID: "d3274e3c-2ad5-4756-8d3f-bae751f1e719"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.519089 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3274e3c-2ad5-4756-8d3f-bae751f1e719-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.519139 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3274e3c-2ad5-4756-8d3f-bae751f1e719-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.519151 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rt55\" (UniqueName: \"kubernetes.io/projected/d3274e3c-2ad5-4756-8d3f-bae751f1e719-kube-api-access-9rt55\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.612192 4997 generic.go:334] "Generic (PLEG): container finished" podID="d3274e3c-2ad5-4756-8d3f-bae751f1e719" containerID="9c2afaa8dfee20860b91a43a1030662f1087186da9f0d30e7927675d98f0bf48" exitCode=0 Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.612283 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zdk9s" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.612349 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zdk9s" event={"ID":"d3274e3c-2ad5-4756-8d3f-bae751f1e719","Type":"ContainerDied","Data":"9c2afaa8dfee20860b91a43a1030662f1087186da9f0d30e7927675d98f0bf48"} Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.612411 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zdk9s" event={"ID":"d3274e3c-2ad5-4756-8d3f-bae751f1e719","Type":"ContainerDied","Data":"af636190be01e54fa8360c220dce646c7204ad819c54aaba92271c59e2e10e67"} Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.612439 4997 scope.go:117] "RemoveContainer" containerID="9c2afaa8dfee20860b91a43a1030662f1087186da9f0d30e7927675d98f0bf48" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.652611 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zdk9s"] Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.656055 4997 scope.go:117] "RemoveContainer" containerID="f33f8296c34e4683e59da1ac65dc94102f8aa3f8cb8c40b4784a9cc92ecd4771" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.665725 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zdk9s"] Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.705844 4997 scope.go:117] "RemoveContainer" containerID="bc9de9eb01b3ac9c658ad129e5487b12cb3a6a4f387cee91a441aba69d50a5d9" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.779773 4997 scope.go:117] "RemoveContainer" containerID="9c2afaa8dfee20860b91a43a1030662f1087186da9f0d30e7927675d98f0bf48" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.781840 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:22:28 crc kubenswrapper[4997]: E1205 07:22:28.785793 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c2afaa8dfee20860b91a43a1030662f1087186da9f0d30e7927675d98f0bf48\": container with ID starting with 9c2afaa8dfee20860b91a43a1030662f1087186da9f0d30e7927675d98f0bf48 not found: ID does not exist" containerID="9c2afaa8dfee20860b91a43a1030662f1087186da9f0d30e7927675d98f0bf48" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.785839 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c2afaa8dfee20860b91a43a1030662f1087186da9f0d30e7927675d98f0bf48"} err="failed to get container status \"9c2afaa8dfee20860b91a43a1030662f1087186da9f0d30e7927675d98f0bf48\": rpc error: code = NotFound desc = could not find container \"9c2afaa8dfee20860b91a43a1030662f1087186da9f0d30e7927675d98f0bf48\": container with ID starting with 9c2afaa8dfee20860b91a43a1030662f1087186da9f0d30e7927675d98f0bf48 not found: ID does not exist" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.785870 4997 scope.go:117] "RemoveContainer" containerID="f33f8296c34e4683e59da1ac65dc94102f8aa3f8cb8c40b4784a9cc92ecd4771" Dec 05 07:22:28 crc kubenswrapper[4997]: E1205 07:22:28.792370 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f33f8296c34e4683e59da1ac65dc94102f8aa3f8cb8c40b4784a9cc92ecd4771\": container with ID starting with f33f8296c34e4683e59da1ac65dc94102f8aa3f8cb8c40b4784a9cc92ecd4771 not found: ID does not exist" containerID="f33f8296c34e4683e59da1ac65dc94102f8aa3f8cb8c40b4784a9cc92ecd4771" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.792449 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f33f8296c34e4683e59da1ac65dc94102f8aa3f8cb8c40b4784a9cc92ecd4771"} err="failed to get container status \"f33f8296c34e4683e59da1ac65dc94102f8aa3f8cb8c40b4784a9cc92ecd4771\": rpc error: code = NotFound desc = could not find container \"f33f8296c34e4683e59da1ac65dc94102f8aa3f8cb8c40b4784a9cc92ecd4771\": container with ID starting with f33f8296c34e4683e59da1ac65dc94102f8aa3f8cb8c40b4784a9cc92ecd4771 not found: ID does not exist" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.792487 4997 scope.go:117] "RemoveContainer" containerID="bc9de9eb01b3ac9c658ad129e5487b12cb3a6a4f387cee91a441aba69d50a5d9" Dec 05 07:22:28 crc kubenswrapper[4997]: E1205 07:22:28.796241 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc9de9eb01b3ac9c658ad129e5487b12cb3a6a4f387cee91a441aba69d50a5d9\": container with ID starting with bc9de9eb01b3ac9c658ad129e5487b12cb3a6a4f387cee91a441aba69d50a5d9 not found: ID does not exist" containerID="bc9de9eb01b3ac9c658ad129e5487b12cb3a6a4f387cee91a441aba69d50a5d9" Dec 05 07:22:28 crc kubenswrapper[4997]: I1205 07:22:28.796297 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc9de9eb01b3ac9c658ad129e5487b12cb3a6a4f387cee91a441aba69d50a5d9"} err="failed to get container status \"bc9de9eb01b3ac9c658ad129e5487b12cb3a6a4f387cee91a441aba69d50a5d9\": rpc error: code = NotFound desc = could not find container \"bc9de9eb01b3ac9c658ad129e5487b12cb3a6a4f387cee91a441aba69d50a5d9\": container with ID starting with bc9de9eb01b3ac9c658ad129e5487b12cb3a6a4f387cee91a441aba69d50a5d9 not found: ID does not exist" Dec 05 07:22:29 crc kubenswrapper[4997]: I1205 07:22:29.411980 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7d75688ddc-xrxqk" podUID="09e38b0e-164a-4927-bbfa-87d8f2023a6c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.188:5353: i/o timeout" Dec 05 07:22:29 crc kubenswrapper[4997]: I1205 07:22:29.639351 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ca00af9-cdab-474a-bc2d-8d57adaf6af3","Type":"ContainerStarted","Data":"699c4552c983b77f8ef82fe2f83784c6634322d0ae65846e6f6427f5d19efb66"} Dec 05 07:22:29 crc kubenswrapper[4997]: I1205 07:22:29.765518 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3274e3c-2ad5-4756-8d3f-bae751f1e719" path="/var/lib/kubelet/pods/d3274e3c-2ad5-4756-8d3f-bae751f1e719/volumes" Dec 05 07:22:30 crc kubenswrapper[4997]: I1205 07:22:30.665995 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ca00af9-cdab-474a-bc2d-8d57adaf6af3","Type":"ContainerStarted","Data":"3494f8420cb24505d7f938e7c164c46b5a8f381b4f3970efe627fc0e233bf507"} Dec 05 07:22:30 crc kubenswrapper[4997]: I1205 07:22:30.668144 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ca00af9-cdab-474a-bc2d-8d57adaf6af3","Type":"ContainerStarted","Data":"7c7e2b6ed58ca875f27f8c171f5355d4ba443fea9ebf25ae53a19570e4018b3d"} Dec 05 07:22:30 crc kubenswrapper[4997]: I1205 07:22:30.668258 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dv6nr" event={"ID":"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5","Type":"ContainerDied","Data":"834d649b80dd8fee860f1e640b0ea6a36d58f4fc6fd2c37271b4e7ce17650c8d"} Dec 05 07:22:30 crc kubenswrapper[4997]: I1205 07:22:30.668058 4997 generic.go:334] "Generic (PLEG): container finished" podID="ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5" containerID="834d649b80dd8fee860f1e640b0ea6a36d58f4fc6fd2c37271b4e7ce17650c8d" exitCode=0 Dec 05 07:22:31 crc kubenswrapper[4997]: I1205 07:22:31.683575 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ca00af9-cdab-474a-bc2d-8d57adaf6af3","Type":"ContainerStarted","Data":"b4bed8cd43908d62f851b25ef5e348d4883411917a38b3826aa5063962705041"} Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.125375 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dv6nr" Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.211287 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fjhg\" (UniqueName: \"kubernetes.io/projected/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-kube-api-access-5fjhg\") pod \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\" (UID: \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\") " Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.211361 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-config-data\") pod \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\" (UID: \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\") " Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.211398 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-scripts\") pod \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\" (UID: \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\") " Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.211569 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-combined-ca-bundle\") pod \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\" (UID: \"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5\") " Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.226814 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-scripts" (OuterVolumeSpecName: "scripts") pod "ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5" (UID: "ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.227039 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-kube-api-access-5fjhg" (OuterVolumeSpecName: "kube-api-access-5fjhg") pod "ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5" (UID: "ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5"). InnerVolumeSpecName "kube-api-access-5fjhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.249107 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-config-data" (OuterVolumeSpecName: "config-data") pod "ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5" (UID: "ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.250521 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5" (UID: "ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.304553 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.304641 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.317317 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.317369 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.317384 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.317403 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fjhg\" (UniqueName: \"kubernetes.io/projected/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5-kube-api-access-5fjhg\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.696201 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dv6nr" event={"ID":"ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5","Type":"ContainerDied","Data":"718e0076d2aa47cef346ef827cfa83b521f5cefe861beed2a3aa65bcb0e7a181"} Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.696728 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="718e0076d2aa47cef346ef827cfa83b521f5cefe861beed2a3aa65bcb0e7a181" Dec 05 07:22:32 crc kubenswrapper[4997]: I1205 07:22:32.696268 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dv6nr" Dec 05 07:22:33 crc kubenswrapper[4997]: I1205 07:22:33.089504 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:22:33 crc kubenswrapper[4997]: I1205 07:22:33.089897 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6673ee19-b22c-4a39-a6ea-bca48d529b4b" containerName="nova-api-log" containerID="cri-o://bb219d357250fdac5860af9df506f620cd51f00862c6b7449d8ab0db68297cec" gracePeriod=30 Dec 05 07:22:33 crc kubenswrapper[4997]: I1205 07:22:33.090044 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6673ee19-b22c-4a39-a6ea-bca48d529b4b" containerName="nova-api-api" containerID="cri-o://7287af1f418f2880f43bdca10a8b67ade52a90451e4e7185ef7e78c5b3a333b3" gracePeriod=30 Dec 05 07:22:33 crc kubenswrapper[4997]: I1205 07:22:33.099923 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6673ee19-b22c-4a39-a6ea-bca48d529b4b" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": EOF" Dec 05 07:22:33 crc kubenswrapper[4997]: I1205 07:22:33.100010 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6673ee19-b22c-4a39-a6ea-bca48d529b4b" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": EOF" Dec 05 07:22:33 crc kubenswrapper[4997]: I1205 07:22:33.120642 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:22:33 crc kubenswrapper[4997]: I1205 07:22:33.121044 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a4305f0a-669e-43e3-b6b8-d2f04a556d1a" containerName="nova-scheduler-scheduler" containerID="cri-o://3fb15335b8e9a5b14b24d7a4a3e397f45a0acca4d7ed2803cbc0e18cb9c51e7f" gracePeriod=30 Dec 05 07:22:33 crc kubenswrapper[4997]: I1205 07:22:33.153296 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:22:33 crc kubenswrapper[4997]: I1205 07:22:33.153651 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ef4a9d91-c9b1-4157-b6e5-485063844959" containerName="nova-metadata-log" containerID="cri-o://a56a7ab1e0d538fd4235867f849354a7dd6ca9ef4109431edc74e29a0b53d67d" gracePeriod=30 Dec 05 07:22:33 crc kubenswrapper[4997]: I1205 07:22:33.153770 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ef4a9d91-c9b1-4157-b6e5-485063844959" containerName="nova-metadata-metadata" containerID="cri-o://47da05d83f44ced35382e509b58b55a2a14d56cd32fca524955be754080be0e4" gracePeriod=30 Dec 05 07:22:33 crc kubenswrapper[4997]: I1205 07:22:33.743480 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ca00af9-cdab-474a-bc2d-8d57adaf6af3","Type":"ContainerStarted","Data":"dc16e86a99e27d23b43ef7dd79991efeb1c1797f058c1872f6bb5014709bbfd6"} Dec 05 07:22:33 crc kubenswrapper[4997]: I1205 07:22:33.743949 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 07:22:33 crc kubenswrapper[4997]: I1205 07:22:33.750517 4997 generic.go:334] "Generic (PLEG): container finished" podID="6673ee19-b22c-4a39-a6ea-bca48d529b4b" containerID="bb219d357250fdac5860af9df506f620cd51f00862c6b7449d8ab0db68297cec" exitCode=143 Dec 05 07:22:33 crc kubenswrapper[4997]: I1205 07:22:33.757240 4997 generic.go:334] "Generic (PLEG): container finished" podID="ef4a9d91-c9b1-4157-b6e5-485063844959" containerID="a56a7ab1e0d538fd4235867f849354a7dd6ca9ef4109431edc74e29a0b53d67d" exitCode=143 Dec 05 07:22:33 crc kubenswrapper[4997]: I1205 07:22:33.760546 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6673ee19-b22c-4a39-a6ea-bca48d529b4b","Type":"ContainerDied","Data":"bb219d357250fdac5860af9df506f620cd51f00862c6b7449d8ab0db68297cec"} Dec 05 07:22:33 crc kubenswrapper[4997]: I1205 07:22:33.760602 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ef4a9d91-c9b1-4157-b6e5-485063844959","Type":"ContainerDied","Data":"a56a7ab1e0d538fd4235867f849354a7dd6ca9ef4109431edc74e29a0b53d67d"} Dec 05 07:22:33 crc kubenswrapper[4997]: I1205 07:22:33.778795 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.655069719 podStartE2EDuration="6.778766151s" podCreationTimestamp="2025-12-05 07:22:27 +0000 UTC" firstStartedPulling="2025-12-05 07:22:28.807807589 +0000 UTC m=+1649.336714850" lastFinishedPulling="2025-12-05 07:22:32.931504021 +0000 UTC m=+1653.460411282" observedRunningTime="2025-12-05 07:22:33.769973233 +0000 UTC m=+1654.298880494" watchObservedRunningTime="2025-12-05 07:22:33.778766151 +0000 UTC m=+1654.307673412" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.291308 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.407767 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p62zd\" (UniqueName: \"kubernetes.io/projected/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-kube-api-access-p62zd\") pod \"a4305f0a-669e-43e3-b6b8-d2f04a556d1a\" (UID: \"a4305f0a-669e-43e3-b6b8-d2f04a556d1a\") " Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.408423 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-config-data\") pod \"a4305f0a-669e-43e3-b6b8-d2f04a556d1a\" (UID: \"a4305f0a-669e-43e3-b6b8-d2f04a556d1a\") " Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.408699 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-combined-ca-bundle\") pod \"a4305f0a-669e-43e3-b6b8-d2f04a556d1a\" (UID: \"a4305f0a-669e-43e3-b6b8-d2f04a556d1a\") " Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.431702 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-kube-api-access-p62zd" (OuterVolumeSpecName: "kube-api-access-p62zd") pod "a4305f0a-669e-43e3-b6b8-d2f04a556d1a" (UID: "a4305f0a-669e-43e3-b6b8-d2f04a556d1a"). InnerVolumeSpecName "kube-api-access-p62zd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.447280 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-config-data" (OuterVolumeSpecName: "config-data") pod "a4305f0a-669e-43e3-b6b8-d2f04a556d1a" (UID: "a4305f0a-669e-43e3-b6b8-d2f04a556d1a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.449638 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4305f0a-669e-43e3-b6b8-d2f04a556d1a" (UID: "a4305f0a-669e-43e3-b6b8-d2f04a556d1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.511768 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.511824 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p62zd\" (UniqueName: \"kubernetes.io/projected/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-kube-api-access-p62zd\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.511840 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4305f0a-669e-43e3-b6b8-d2f04a556d1a-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.781005 4997 generic.go:334] "Generic (PLEG): container finished" podID="a4305f0a-669e-43e3-b6b8-d2f04a556d1a" containerID="3fb15335b8e9a5b14b24d7a4a3e397f45a0acca4d7ed2803cbc0e18cb9c51e7f" exitCode=0 Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.781066 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a4305f0a-669e-43e3-b6b8-d2f04a556d1a","Type":"ContainerDied","Data":"3fb15335b8e9a5b14b24d7a4a3e397f45a0acca4d7ed2803cbc0e18cb9c51e7f"} Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.781078 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.781111 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a4305f0a-669e-43e3-b6b8-d2f04a556d1a","Type":"ContainerDied","Data":"d8a909c219ab5b0ba0c5c080e427dda4882ede15cd8de4d20a9e7e7df388ad83"} Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.781134 4997 scope.go:117] "RemoveContainer" containerID="3fb15335b8e9a5b14b24d7a4a3e397f45a0acca4d7ed2803cbc0e18cb9c51e7f" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.817729 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.821610 4997 scope.go:117] "RemoveContainer" containerID="3fb15335b8e9a5b14b24d7a4a3e397f45a0acca4d7ed2803cbc0e18cb9c51e7f" Dec 05 07:22:35 crc kubenswrapper[4997]: E1205 07:22:35.822223 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fb15335b8e9a5b14b24d7a4a3e397f45a0acca4d7ed2803cbc0e18cb9c51e7f\": container with ID starting with 3fb15335b8e9a5b14b24d7a4a3e397f45a0acca4d7ed2803cbc0e18cb9c51e7f not found: ID does not exist" containerID="3fb15335b8e9a5b14b24d7a4a3e397f45a0acca4d7ed2803cbc0e18cb9c51e7f" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.822285 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fb15335b8e9a5b14b24d7a4a3e397f45a0acca4d7ed2803cbc0e18cb9c51e7f"} err="failed to get container status \"3fb15335b8e9a5b14b24d7a4a3e397f45a0acca4d7ed2803cbc0e18cb9c51e7f\": rpc error: code = NotFound desc = could not find container \"3fb15335b8e9a5b14b24d7a4a3e397f45a0acca4d7ed2803cbc0e18cb9c51e7f\": container with ID starting with 3fb15335b8e9a5b14b24d7a4a3e397f45a0acca4d7ed2803cbc0e18cb9c51e7f not found: ID does not exist" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.844486 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.858816 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:22:35 crc kubenswrapper[4997]: E1205 07:22:35.859798 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3274e3c-2ad5-4756-8d3f-bae751f1e719" containerName="registry-server" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.859930 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3274e3c-2ad5-4756-8d3f-bae751f1e719" containerName="registry-server" Dec 05 07:22:35 crc kubenswrapper[4997]: E1205 07:22:35.860020 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3274e3c-2ad5-4756-8d3f-bae751f1e719" containerName="extract-content" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.860091 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3274e3c-2ad5-4756-8d3f-bae751f1e719" containerName="extract-content" Dec 05 07:22:35 crc kubenswrapper[4997]: E1205 07:22:35.860186 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4305f0a-669e-43e3-b6b8-d2f04a556d1a" containerName="nova-scheduler-scheduler" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.860255 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4305f0a-669e-43e3-b6b8-d2f04a556d1a" containerName="nova-scheduler-scheduler" Dec 05 07:22:35 crc kubenswrapper[4997]: E1205 07:22:35.860365 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3274e3c-2ad5-4756-8d3f-bae751f1e719" containerName="extract-utilities" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.860446 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3274e3c-2ad5-4756-8d3f-bae751f1e719" containerName="extract-utilities" Dec 05 07:22:35 crc kubenswrapper[4997]: E1205 07:22:35.860541 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5" containerName="nova-manage" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.860744 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5" containerName="nova-manage" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.861086 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5" containerName="nova-manage" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.861198 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3274e3c-2ad5-4756-8d3f-bae751f1e719" containerName="registry-server" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.861287 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4305f0a-669e-43e3-b6b8-d2f04a556d1a" containerName="nova-scheduler-scheduler" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.862452 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.867174 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 07:22:35 crc kubenswrapper[4997]: I1205 07:22:35.874120 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:22:36 crc kubenswrapper[4997]: I1205 07:22:36.026150 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed\") " pod="openstack/nova-scheduler-0" Dec 05 07:22:36 crc kubenswrapper[4997]: I1205 07:22:36.026319 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbz65\" (UniqueName: \"kubernetes.io/projected/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-kube-api-access-nbz65\") pod \"nova-scheduler-0\" (UID: \"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed\") " pod="openstack/nova-scheduler-0" Dec 05 07:22:36 crc kubenswrapper[4997]: I1205 07:22:36.026391 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-config-data\") pod \"nova-scheduler-0\" (UID: \"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed\") " pod="openstack/nova-scheduler-0" Dec 05 07:22:36 crc kubenswrapper[4997]: I1205 07:22:36.129026 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbz65\" (UniqueName: \"kubernetes.io/projected/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-kube-api-access-nbz65\") pod \"nova-scheduler-0\" (UID: \"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed\") " pod="openstack/nova-scheduler-0" Dec 05 07:22:36 crc kubenswrapper[4997]: I1205 07:22:36.129197 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-config-data\") pod \"nova-scheduler-0\" (UID: \"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed\") " pod="openstack/nova-scheduler-0" Dec 05 07:22:36 crc kubenswrapper[4997]: I1205 07:22:36.129286 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed\") " pod="openstack/nova-scheduler-0" Dec 05 07:22:36 crc kubenswrapper[4997]: I1205 07:22:36.135037 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-config-data\") pod \"nova-scheduler-0\" (UID: \"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed\") " pod="openstack/nova-scheduler-0" Dec 05 07:22:36 crc kubenswrapper[4997]: I1205 07:22:36.135471 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed\") " pod="openstack/nova-scheduler-0" Dec 05 07:22:36 crc kubenswrapper[4997]: I1205 07:22:36.153492 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbz65\" (UniqueName: \"kubernetes.io/projected/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-kube-api-access-nbz65\") pod \"nova-scheduler-0\" (UID: \"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed\") " pod="openstack/nova-scheduler-0" Dec 05 07:22:36 crc kubenswrapper[4997]: I1205 07:22:36.190107 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 07:22:36 crc kubenswrapper[4997]: I1205 07:22:36.609426 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ef4a9d91-c9b1-4157-b6e5-485063844959" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": read tcp 10.217.0.2:42020->10.217.0.194:8775: read: connection reset by peer" Dec 05 07:22:36 crc kubenswrapper[4997]: I1205 07:22:36.610267 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ef4a9d91-c9b1-4157-b6e5-485063844959" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": read tcp 10.217.0.2:42022->10.217.0.194:8775: read: connection reset by peer" Dec 05 07:22:36 crc kubenswrapper[4997]: I1205 07:22:36.797304 4997 generic.go:334] "Generic (PLEG): container finished" podID="ef4a9d91-c9b1-4157-b6e5-485063844959" containerID="47da05d83f44ced35382e509b58b55a2a14d56cd32fca524955be754080be0e4" exitCode=0 Dec 05 07:22:36 crc kubenswrapper[4997]: I1205 07:22:36.797383 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ef4a9d91-c9b1-4157-b6e5-485063844959","Type":"ContainerDied","Data":"47da05d83f44ced35382e509b58b55a2a14d56cd32fca524955be754080be0e4"} Dec 05 07:22:36 crc kubenswrapper[4997]: I1205 07:22:36.911511 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.044269 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.154905 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-nova-metadata-tls-certs\") pod \"ef4a9d91-c9b1-4157-b6e5-485063844959\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.154966 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef4a9d91-c9b1-4157-b6e5-485063844959-logs\") pod \"ef4a9d91-c9b1-4157-b6e5-485063844959\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.155072 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-config-data\") pod \"ef4a9d91-c9b1-4157-b6e5-485063844959\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.155269 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-combined-ca-bundle\") pod \"ef4a9d91-c9b1-4157-b6e5-485063844959\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.155366 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9mzf\" (UniqueName: \"kubernetes.io/projected/ef4a9d91-c9b1-4157-b6e5-485063844959-kube-api-access-g9mzf\") pod \"ef4a9d91-c9b1-4157-b6e5-485063844959\" (UID: \"ef4a9d91-c9b1-4157-b6e5-485063844959\") " Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.156837 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef4a9d91-c9b1-4157-b6e5-485063844959-logs" (OuterVolumeSpecName: "logs") pod "ef4a9d91-c9b1-4157-b6e5-485063844959" (UID: "ef4a9d91-c9b1-4157-b6e5-485063844959"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.167131 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef4a9d91-c9b1-4157-b6e5-485063844959-kube-api-access-g9mzf" (OuterVolumeSpecName: "kube-api-access-g9mzf") pod "ef4a9d91-c9b1-4157-b6e5-485063844959" (UID: "ef4a9d91-c9b1-4157-b6e5-485063844959"). InnerVolumeSpecName "kube-api-access-g9mzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.192632 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef4a9d91-c9b1-4157-b6e5-485063844959" (UID: "ef4a9d91-c9b1-4157-b6e5-485063844959"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.225296 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-config-data" (OuterVolumeSpecName: "config-data") pod "ef4a9d91-c9b1-4157-b6e5-485063844959" (UID: "ef4a9d91-c9b1-4157-b6e5-485063844959"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.233085 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "ef4a9d91-c9b1-4157-b6e5-485063844959" (UID: "ef4a9d91-c9b1-4157-b6e5-485063844959"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.262245 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.262277 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.262292 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9mzf\" (UniqueName: \"kubernetes.io/projected/ef4a9d91-c9b1-4157-b6e5-485063844959-kube-api-access-g9mzf\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.262302 4997 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef4a9d91-c9b1-4157-b6e5-485063844959-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.262311 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef4a9d91-c9b1-4157-b6e5-485063844959-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.780933 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4305f0a-669e-43e3-b6b8-d2f04a556d1a" path="/var/lib/kubelet/pods/a4305f0a-669e-43e3-b6b8-d2f04a556d1a/volumes" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.814092 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed","Type":"ContainerStarted","Data":"96f6c22ff0cd4cb36b8fadc88840e34b4298c41452961e008b5dd7fc23ef4889"} Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.814156 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed","Type":"ContainerStarted","Data":"8fcc5ac3680b52427d3eabdfd38a2e59e4209c08c59501e314ae1a992f0a4113"} Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.816792 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ef4a9d91-c9b1-4157-b6e5-485063844959","Type":"ContainerDied","Data":"ba7919f30ccfce0bdf5be6e5d34db245d5b1292c2c9ea2e80a10a2fce6fb5d08"} Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.816862 4997 scope.go:117] "RemoveContainer" containerID="47da05d83f44ced35382e509b58b55a2a14d56cd32fca524955be754080be0e4" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.817065 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.836038 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.836011863 podStartE2EDuration="2.836011863s" podCreationTimestamp="2025-12-05 07:22:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:22:37.835153129 +0000 UTC m=+1658.364060390" watchObservedRunningTime="2025-12-05 07:22:37.836011863 +0000 UTC m=+1658.364919124" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.859891 4997 scope.go:117] "RemoveContainer" containerID="a56a7ab1e0d538fd4235867f849354a7dd6ca9ef4109431edc74e29a0b53d67d" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.877535 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.924466 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.939195 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:22:37 crc kubenswrapper[4997]: E1205 07:22:37.940285 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef4a9d91-c9b1-4157-b6e5-485063844959" containerName="nova-metadata-log" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.940308 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef4a9d91-c9b1-4157-b6e5-485063844959" containerName="nova-metadata-log" Dec 05 07:22:37 crc kubenswrapper[4997]: E1205 07:22:37.940369 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef4a9d91-c9b1-4157-b6e5-485063844959" containerName="nova-metadata-metadata" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.940457 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef4a9d91-c9b1-4157-b6e5-485063844959" containerName="nova-metadata-metadata" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.940878 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef4a9d91-c9b1-4157-b6e5-485063844959" containerName="nova-metadata-log" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.940940 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef4a9d91-c9b1-4157-b6e5-485063844959" containerName="nova-metadata-metadata" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.944710 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.948297 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.950811 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 05 07:22:37 crc kubenswrapper[4997]: I1205 07:22:37.987052 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:22:38 crc kubenswrapper[4997]: I1205 07:22:38.112936 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-logs\") pod \"nova-metadata-0\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " pod="openstack/nova-metadata-0" Dec 05 07:22:38 crc kubenswrapper[4997]: I1205 07:22:38.113376 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb885\" (UniqueName: \"kubernetes.io/projected/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-kube-api-access-zb885\") pod \"nova-metadata-0\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " pod="openstack/nova-metadata-0" Dec 05 07:22:38 crc kubenswrapper[4997]: I1205 07:22:38.113481 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-config-data\") pod \"nova-metadata-0\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " pod="openstack/nova-metadata-0" Dec 05 07:22:38 crc kubenswrapper[4997]: I1205 07:22:38.113516 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " pod="openstack/nova-metadata-0" Dec 05 07:22:38 crc kubenswrapper[4997]: I1205 07:22:38.113577 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " pod="openstack/nova-metadata-0" Dec 05 07:22:38 crc kubenswrapper[4997]: I1205 07:22:38.218168 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-config-data\") pod \"nova-metadata-0\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " pod="openstack/nova-metadata-0" Dec 05 07:22:38 crc kubenswrapper[4997]: I1205 07:22:38.218234 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " pod="openstack/nova-metadata-0" Dec 05 07:22:38 crc kubenswrapper[4997]: I1205 07:22:38.218300 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " pod="openstack/nova-metadata-0" Dec 05 07:22:38 crc kubenswrapper[4997]: I1205 07:22:38.218383 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-logs\") pod \"nova-metadata-0\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " pod="openstack/nova-metadata-0" Dec 05 07:22:38 crc kubenswrapper[4997]: I1205 07:22:38.218404 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb885\" (UniqueName: \"kubernetes.io/projected/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-kube-api-access-zb885\") pod \"nova-metadata-0\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " pod="openstack/nova-metadata-0" Dec 05 07:22:38 crc kubenswrapper[4997]: I1205 07:22:38.219525 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-logs\") pod \"nova-metadata-0\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " pod="openstack/nova-metadata-0" Dec 05 07:22:38 crc kubenswrapper[4997]: I1205 07:22:38.226480 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " pod="openstack/nova-metadata-0" Dec 05 07:22:38 crc kubenswrapper[4997]: I1205 07:22:38.227790 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " pod="openstack/nova-metadata-0" Dec 05 07:22:38 crc kubenswrapper[4997]: I1205 07:22:38.227869 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-config-data\") pod \"nova-metadata-0\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " pod="openstack/nova-metadata-0" Dec 05 07:22:38 crc kubenswrapper[4997]: I1205 07:22:38.238274 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb885\" (UniqueName: \"kubernetes.io/projected/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-kube-api-access-zb885\") pod \"nova-metadata-0\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " pod="openstack/nova-metadata-0" Dec 05 07:22:38 crc kubenswrapper[4997]: I1205 07:22:38.304530 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 07:22:38 crc kubenswrapper[4997]: I1205 07:22:38.816398 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:22:39 crc kubenswrapper[4997]: I1205 07:22:39.767344 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef4a9d91-c9b1-4157-b6e5-485063844959" path="/var/lib/kubelet/pods/ef4a9d91-c9b1-4157-b6e5-485063844959/volumes" Dec 05 07:22:39 crc kubenswrapper[4997]: I1205 07:22:39.855223 4997 generic.go:334] "Generic (PLEG): container finished" podID="6673ee19-b22c-4a39-a6ea-bca48d529b4b" containerID="7287af1f418f2880f43bdca10a8b67ade52a90451e4e7185ef7e78c5b3a333b3" exitCode=0 Dec 05 07:22:39 crc kubenswrapper[4997]: I1205 07:22:39.855323 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6673ee19-b22c-4a39-a6ea-bca48d529b4b","Type":"ContainerDied","Data":"7287af1f418f2880f43bdca10a8b67ade52a90451e4e7185ef7e78c5b3a333b3"} Dec 05 07:22:39 crc kubenswrapper[4997]: I1205 07:22:39.859978 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bcbe26f4-fcd9-46c7-aa3a-5038b801741d","Type":"ContainerStarted","Data":"96ac136074ed0bcb85702f5bfe22b13a7895276ce80e0413118cb761bfafcad0"} Dec 05 07:22:39 crc kubenswrapper[4997]: I1205 07:22:39.860009 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bcbe26f4-fcd9-46c7-aa3a-5038b801741d","Type":"ContainerStarted","Data":"db7397da7d823a2ab29a061ca9792ba46bd6e1c49ac1ae528d2649ad3365c16a"} Dec 05 07:22:39 crc kubenswrapper[4997]: I1205 07:22:39.860023 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bcbe26f4-fcd9-46c7-aa3a-5038b801741d","Type":"ContainerStarted","Data":"c8cefb27923ca6a1b8c5bdd6a3af492257b89f880b6c87eec48e6a7441270d51"} Dec 05 07:22:39 crc kubenswrapper[4997]: I1205 07:22:39.890784 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.8907326209999997 podStartE2EDuration="2.890732621s" podCreationTimestamp="2025-12-05 07:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:22:39.882562789 +0000 UTC m=+1660.411470070" watchObservedRunningTime="2025-12-05 07:22:39.890732621 +0000 UTC m=+1660.419639882" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.034252 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.073545 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6673ee19-b22c-4a39-a6ea-bca48d529b4b-logs\") pod \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.073717 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrjs8\" (UniqueName: \"kubernetes.io/projected/6673ee19-b22c-4a39-a6ea-bca48d529b4b-kube-api-access-nrjs8\") pod \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.073791 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-config-data\") pod \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.073826 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-internal-tls-certs\") pod \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.073926 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-combined-ca-bundle\") pod \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.074820 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-public-tls-certs\") pod \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\" (UID: \"6673ee19-b22c-4a39-a6ea-bca48d529b4b\") " Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.074826 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6673ee19-b22c-4a39-a6ea-bca48d529b4b-logs" (OuterVolumeSpecName: "logs") pod "6673ee19-b22c-4a39-a6ea-bca48d529b4b" (UID: "6673ee19-b22c-4a39-a6ea-bca48d529b4b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.075152 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6673ee19-b22c-4a39-a6ea-bca48d529b4b-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.087966 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6673ee19-b22c-4a39-a6ea-bca48d529b4b-kube-api-access-nrjs8" (OuterVolumeSpecName: "kube-api-access-nrjs8") pod "6673ee19-b22c-4a39-a6ea-bca48d529b4b" (UID: "6673ee19-b22c-4a39-a6ea-bca48d529b4b"). InnerVolumeSpecName "kube-api-access-nrjs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.175804 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-config-data" (OuterVolumeSpecName: "config-data") pod "6673ee19-b22c-4a39-a6ea-bca48d529b4b" (UID: "6673ee19-b22c-4a39-a6ea-bca48d529b4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.177417 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrjs8\" (UniqueName: \"kubernetes.io/projected/6673ee19-b22c-4a39-a6ea-bca48d529b4b-kube-api-access-nrjs8\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.177438 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.199890 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6673ee19-b22c-4a39-a6ea-bca48d529b4b" (UID: "6673ee19-b22c-4a39-a6ea-bca48d529b4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.212915 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6673ee19-b22c-4a39-a6ea-bca48d529b4b" (UID: "6673ee19-b22c-4a39-a6ea-bca48d529b4b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.217036 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6673ee19-b22c-4a39-a6ea-bca48d529b4b" (UID: "6673ee19-b22c-4a39-a6ea-bca48d529b4b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.279340 4997 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.279394 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.279404 4997 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6673ee19-b22c-4a39-a6ea-bca48d529b4b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.881138 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6673ee19-b22c-4a39-a6ea-bca48d529b4b","Type":"ContainerDied","Data":"5d1a871a19941cac1e50e94b897d7c66dc3485e4396828918865a90bff5b6c1a"} Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.881643 4997 scope.go:117] "RemoveContainer" containerID="7287af1f418f2880f43bdca10a8b67ade52a90451e4e7185ef7e78c5b3a333b3" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.881178 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.915492 4997 scope.go:117] "RemoveContainer" containerID="bb219d357250fdac5860af9df506f620cd51f00862c6b7449d8ab0db68297cec" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.927413 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.944842 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.971508 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 07:22:40 crc kubenswrapper[4997]: E1205 07:22:40.972201 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6673ee19-b22c-4a39-a6ea-bca48d529b4b" containerName="nova-api-log" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.972233 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6673ee19-b22c-4a39-a6ea-bca48d529b4b" containerName="nova-api-log" Dec 05 07:22:40 crc kubenswrapper[4997]: E1205 07:22:40.972266 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6673ee19-b22c-4a39-a6ea-bca48d529b4b" containerName="nova-api-api" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.972274 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6673ee19-b22c-4a39-a6ea-bca48d529b4b" containerName="nova-api-api" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.972498 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6673ee19-b22c-4a39-a6ea-bca48d529b4b" containerName="nova-api-api" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.972539 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6673ee19-b22c-4a39-a6ea-bca48d529b4b" containerName="nova-api-log" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.974379 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.978005 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.979722 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.979852 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 05 07:22:40 crc kubenswrapper[4997]: I1205 07:22:40.983787 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.098767 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5afeb230-4a35-4872-bffa-c882fe927643-logs\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.098880 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5m8mj\" (UniqueName: \"kubernetes.io/projected/5afeb230-4a35-4872-bffa-c882fe927643-kube-api-access-5m8mj\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.099149 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.101034 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-config-data\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.101173 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-public-tls-certs\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.102015 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.190297 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.203783 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.203855 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5afeb230-4a35-4872-bffa-c882fe927643-logs\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.203924 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5m8mj\" (UniqueName: \"kubernetes.io/projected/5afeb230-4a35-4872-bffa-c882fe927643-kube-api-access-5m8mj\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.203966 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.204007 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-config-data\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.204056 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-public-tls-certs\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.204843 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5afeb230-4a35-4872-bffa-c882fe927643-logs\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.209982 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.210430 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.212907 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-public-tls-certs\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.216184 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-config-data\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.238924 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5m8mj\" (UniqueName: \"kubernetes.io/projected/5afeb230-4a35-4872-bffa-c882fe927643-kube-api-access-5m8mj\") pod \"nova-api-0\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.300833 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.599121 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.765323 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6673ee19-b22c-4a39-a6ea-bca48d529b4b" path="/var/lib/kubelet/pods/6673ee19-b22c-4a39-a6ea-bca48d529b4b/volumes" Dec 05 07:22:41 crc kubenswrapper[4997]: I1205 07:22:41.892899 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5afeb230-4a35-4872-bffa-c882fe927643","Type":"ContainerStarted","Data":"5082ac2136f164cf222572a1bf4fe1a02e45628512baa05a6250afdfaff826ff"} Dec 05 07:22:42 crc kubenswrapper[4997]: I1205 07:22:42.912475 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5afeb230-4a35-4872-bffa-c882fe927643","Type":"ContainerStarted","Data":"27d910165cbca8f6eb45e1a947fe3daf0e9d27a117dff110f3fc7aeef2a06961"} Dec 05 07:22:42 crc kubenswrapper[4997]: I1205 07:22:42.912554 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5afeb230-4a35-4872-bffa-c882fe927643","Type":"ContainerStarted","Data":"e1a5fbc1a91ff3e0d6d191e2796f352f99e1dd7df658b4c7a2370f2df1211989"} Dec 05 07:22:42 crc kubenswrapper[4997]: I1205 07:22:42.952114 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.952084151 podStartE2EDuration="2.952084151s" podCreationTimestamp="2025-12-05 07:22:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:22:42.938010219 +0000 UTC m=+1663.466917490" watchObservedRunningTime="2025-12-05 07:22:42.952084151 +0000 UTC m=+1663.480991422" Dec 05 07:22:43 crc kubenswrapper[4997]: I1205 07:22:43.306285 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 07:22:43 crc kubenswrapper[4997]: I1205 07:22:43.306658 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 07:22:46 crc kubenswrapper[4997]: I1205 07:22:46.190758 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 07:22:46 crc kubenswrapper[4997]: I1205 07:22:46.222111 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 07:22:46 crc kubenswrapper[4997]: I1205 07:22:46.997531 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 07:22:48 crc kubenswrapper[4997]: I1205 07:22:48.306514 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 07:22:48 crc kubenswrapper[4997]: I1205 07:22:48.306672 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 07:22:49 crc kubenswrapper[4997]: I1205 07:22:49.327824 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bcbe26f4-fcd9-46c7-aa3a-5038b801741d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 07:22:49 crc kubenswrapper[4997]: I1205 07:22:49.327931 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bcbe26f4-fcd9-46c7-aa3a-5038b801741d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 07:22:49 crc kubenswrapper[4997]: I1205 07:22:49.770047 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:22:49 crc kubenswrapper[4997]: I1205 07:22:49.770707 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:22:49 crc kubenswrapper[4997]: I1205 07:22:49.770893 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 07:22:49 crc kubenswrapper[4997]: I1205 07:22:49.772423 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 07:22:49 crc kubenswrapper[4997]: I1205 07:22:49.772753 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" gracePeriod=600 Dec 05 07:22:49 crc kubenswrapper[4997]: E1205 07:22:49.910773 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:22:50 crc kubenswrapper[4997]: I1205 07:22:50.010825 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" exitCode=0 Dec 05 07:22:50 crc kubenswrapper[4997]: I1205 07:22:50.010874 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08"} Dec 05 07:22:50 crc kubenswrapper[4997]: I1205 07:22:50.010927 4997 scope.go:117] "RemoveContainer" containerID="1b8d25609ffa42f5e50251e3eb59f7240e05fe03de714e77c29a98e4dc7b4510" Dec 05 07:22:50 crc kubenswrapper[4997]: I1205 07:22:50.011764 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:22:50 crc kubenswrapper[4997]: E1205 07:22:50.012050 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:22:50 crc kubenswrapper[4997]: I1205 07:22:50.355478 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8d8zv"] Dec 05 07:22:50 crc kubenswrapper[4997]: I1205 07:22:50.358448 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8d8zv" Dec 05 07:22:50 crc kubenswrapper[4997]: I1205 07:22:50.382049 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8d8zv"] Dec 05 07:22:50 crc kubenswrapper[4997]: I1205 07:22:50.476872 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f00e97b3-98c3-41b6-9949-2961dd9524a9-utilities\") pod \"certified-operators-8d8zv\" (UID: \"f00e97b3-98c3-41b6-9949-2961dd9524a9\") " pod="openshift-marketplace/certified-operators-8d8zv" Dec 05 07:22:50 crc kubenswrapper[4997]: I1205 07:22:50.477005 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szrlr\" (UniqueName: \"kubernetes.io/projected/f00e97b3-98c3-41b6-9949-2961dd9524a9-kube-api-access-szrlr\") pod \"certified-operators-8d8zv\" (UID: \"f00e97b3-98c3-41b6-9949-2961dd9524a9\") " pod="openshift-marketplace/certified-operators-8d8zv" Dec 05 07:22:50 crc kubenswrapper[4997]: I1205 07:22:50.477095 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f00e97b3-98c3-41b6-9949-2961dd9524a9-catalog-content\") pod \"certified-operators-8d8zv\" (UID: \"f00e97b3-98c3-41b6-9949-2961dd9524a9\") " pod="openshift-marketplace/certified-operators-8d8zv" Dec 05 07:22:50 crc kubenswrapper[4997]: I1205 07:22:50.579017 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f00e97b3-98c3-41b6-9949-2961dd9524a9-utilities\") pod \"certified-operators-8d8zv\" (UID: \"f00e97b3-98c3-41b6-9949-2961dd9524a9\") " pod="openshift-marketplace/certified-operators-8d8zv" Dec 05 07:22:50 crc kubenswrapper[4997]: I1205 07:22:50.579176 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szrlr\" (UniqueName: \"kubernetes.io/projected/f00e97b3-98c3-41b6-9949-2961dd9524a9-kube-api-access-szrlr\") pod \"certified-operators-8d8zv\" (UID: \"f00e97b3-98c3-41b6-9949-2961dd9524a9\") " pod="openshift-marketplace/certified-operators-8d8zv" Dec 05 07:22:50 crc kubenswrapper[4997]: I1205 07:22:50.579295 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f00e97b3-98c3-41b6-9949-2961dd9524a9-catalog-content\") pod \"certified-operators-8d8zv\" (UID: \"f00e97b3-98c3-41b6-9949-2961dd9524a9\") " pod="openshift-marketplace/certified-operators-8d8zv" Dec 05 07:22:50 crc kubenswrapper[4997]: I1205 07:22:50.579925 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f00e97b3-98c3-41b6-9949-2961dd9524a9-catalog-content\") pod \"certified-operators-8d8zv\" (UID: \"f00e97b3-98c3-41b6-9949-2961dd9524a9\") " pod="openshift-marketplace/certified-operators-8d8zv" Dec 05 07:22:50 crc kubenswrapper[4997]: I1205 07:22:50.580217 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f00e97b3-98c3-41b6-9949-2961dd9524a9-utilities\") pod \"certified-operators-8d8zv\" (UID: \"f00e97b3-98c3-41b6-9949-2961dd9524a9\") " pod="openshift-marketplace/certified-operators-8d8zv" Dec 05 07:22:50 crc kubenswrapper[4997]: I1205 07:22:50.602943 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szrlr\" (UniqueName: \"kubernetes.io/projected/f00e97b3-98c3-41b6-9949-2961dd9524a9-kube-api-access-szrlr\") pod \"certified-operators-8d8zv\" (UID: \"f00e97b3-98c3-41b6-9949-2961dd9524a9\") " pod="openshift-marketplace/certified-operators-8d8zv" Dec 05 07:22:50 crc kubenswrapper[4997]: I1205 07:22:50.684716 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8d8zv" Dec 05 07:22:51 crc kubenswrapper[4997]: I1205 07:22:51.263555 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8d8zv"] Dec 05 07:22:51 crc kubenswrapper[4997]: I1205 07:22:51.308025 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 07:22:51 crc kubenswrapper[4997]: I1205 07:22:51.308091 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 07:22:52 crc kubenswrapper[4997]: I1205 07:22:52.085444 4997 generic.go:334] "Generic (PLEG): container finished" podID="f00e97b3-98c3-41b6-9949-2961dd9524a9" containerID="e6cdc8b7d6baf3660b5da72cb1df31c0edb9ffe1cbe62631356d380c42d3e077" exitCode=0 Dec 05 07:22:52 crc kubenswrapper[4997]: I1205 07:22:52.085842 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8d8zv" event={"ID":"f00e97b3-98c3-41b6-9949-2961dd9524a9","Type":"ContainerDied","Data":"e6cdc8b7d6baf3660b5da72cb1df31c0edb9ffe1cbe62631356d380c42d3e077"} Dec 05 07:22:52 crc kubenswrapper[4997]: I1205 07:22:52.085879 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8d8zv" event={"ID":"f00e97b3-98c3-41b6-9949-2961dd9524a9","Type":"ContainerStarted","Data":"ce7a90b4187f38121ce33f75b1dd2b72343e829fe71e02dfc12b843d34c404c9"} Dec 05 07:22:52 crc kubenswrapper[4997]: I1205 07:22:52.318794 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5afeb230-4a35-4872-bffa-c882fe927643" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 07:22:52 crc kubenswrapper[4997]: I1205 07:22:52.318794 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5afeb230-4a35-4872-bffa-c882fe927643" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 07:22:53 crc kubenswrapper[4997]: I1205 07:22:53.100360 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8d8zv" event={"ID":"f00e97b3-98c3-41b6-9949-2961dd9524a9","Type":"ContainerStarted","Data":"beda7b0a7cd405b799e7f87e5d2fa4ad91916f784caf0e4ebffdebe24cf3fc16"} Dec 05 07:22:54 crc kubenswrapper[4997]: I1205 07:22:54.114528 4997 generic.go:334] "Generic (PLEG): container finished" podID="f00e97b3-98c3-41b6-9949-2961dd9524a9" containerID="beda7b0a7cd405b799e7f87e5d2fa4ad91916f784caf0e4ebffdebe24cf3fc16" exitCode=0 Dec 05 07:22:54 crc kubenswrapper[4997]: I1205 07:22:54.114578 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8d8zv" event={"ID":"f00e97b3-98c3-41b6-9949-2961dd9524a9","Type":"ContainerDied","Data":"beda7b0a7cd405b799e7f87e5d2fa4ad91916f784caf0e4ebffdebe24cf3fc16"} Dec 05 07:22:55 crc kubenswrapper[4997]: I1205 07:22:55.128549 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8d8zv" event={"ID":"f00e97b3-98c3-41b6-9949-2961dd9524a9","Type":"ContainerStarted","Data":"a85c29ddb4a8ae483a8bee5c910e71da300c3a157e62f1646c991dfb7f16c845"} Dec 05 07:22:55 crc kubenswrapper[4997]: I1205 07:22:55.150539 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8d8zv" podStartSLOduration=2.593812959 podStartE2EDuration="5.150511086s" podCreationTimestamp="2025-12-05 07:22:50 +0000 UTC" firstStartedPulling="2025-12-05 07:22:52.087859262 +0000 UTC m=+1672.616766523" lastFinishedPulling="2025-12-05 07:22:54.644557389 +0000 UTC m=+1675.173464650" observedRunningTime="2025-12-05 07:22:55.148807631 +0000 UTC m=+1675.677714902" watchObservedRunningTime="2025-12-05 07:22:55.150511086 +0000 UTC m=+1675.679418347" Dec 05 07:22:58 crc kubenswrapper[4997]: I1205 07:22:58.062353 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 07:22:58 crc kubenswrapper[4997]: I1205 07:22:58.316858 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 07:22:58 crc kubenswrapper[4997]: I1205 07:22:58.329279 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 07:22:58 crc kubenswrapper[4997]: I1205 07:22:58.330169 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 07:22:59 crc kubenswrapper[4997]: I1205 07:22:59.182245 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 07:23:00 crc kubenswrapper[4997]: I1205 07:23:00.685844 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8d8zv" Dec 05 07:23:00 crc kubenswrapper[4997]: I1205 07:23:00.686446 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8d8zv" Dec 05 07:23:00 crc kubenswrapper[4997]: I1205 07:23:00.757494 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8d8zv" Dec 05 07:23:01 crc kubenswrapper[4997]: I1205 07:23:01.251775 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8d8zv" Dec 05 07:23:01 crc kubenswrapper[4997]: I1205 07:23:01.322415 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8d8zv"] Dec 05 07:23:01 crc kubenswrapper[4997]: I1205 07:23:01.325724 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 07:23:01 crc kubenswrapper[4997]: I1205 07:23:01.326245 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 07:23:01 crc kubenswrapper[4997]: I1205 07:23:01.326680 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 07:23:01 crc kubenswrapper[4997]: I1205 07:23:01.336024 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 07:23:02 crc kubenswrapper[4997]: I1205 07:23:02.221282 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 07:23:02 crc kubenswrapper[4997]: I1205 07:23:02.228400 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 07:23:03 crc kubenswrapper[4997]: I1205 07:23:03.065097 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 07:23:03 crc kubenswrapper[4997]: I1205 07:23:03.065741 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="833f0bba-91d9-41ab-a412-9e8ae8c38ba2" containerName="kube-state-metrics" containerID="cri-o://a2793e5e578daef37735f5bea301361550f5c913442e8d7f862245cb9db435d5" gracePeriod=30 Dec 05 07:23:03 crc kubenswrapper[4997]: I1205 07:23:03.240236 4997 generic.go:334] "Generic (PLEG): container finished" podID="833f0bba-91d9-41ab-a412-9e8ae8c38ba2" containerID="a2793e5e578daef37735f5bea301361550f5c913442e8d7f862245cb9db435d5" exitCode=2 Dec 05 07:23:03 crc kubenswrapper[4997]: I1205 07:23:03.240601 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"833f0bba-91d9-41ab-a412-9e8ae8c38ba2","Type":"ContainerDied","Data":"a2793e5e578daef37735f5bea301361550f5c913442e8d7f862245cb9db435d5"} Dec 05 07:23:03 crc kubenswrapper[4997]: I1205 07:23:03.241198 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8d8zv" podUID="f00e97b3-98c3-41b6-9949-2961dd9524a9" containerName="registry-server" containerID="cri-o://a85c29ddb4a8ae483a8bee5c910e71da300c3a157e62f1646c991dfb7f16c845" gracePeriod=2 Dec 05 07:23:03 crc kubenswrapper[4997]: I1205 07:23:03.680465 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 07:23:03 crc kubenswrapper[4997]: I1205 07:23:03.753514 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:23:03 crc kubenswrapper[4997]: E1205 07:23:03.753897 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:23:03 crc kubenswrapper[4997]: I1205 07:23:03.760355 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8d8zv" Dec 05 07:23:03 crc kubenswrapper[4997]: I1205 07:23:03.821170 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nv24p\" (UniqueName: \"kubernetes.io/projected/833f0bba-91d9-41ab-a412-9e8ae8c38ba2-kube-api-access-nv24p\") pod \"833f0bba-91d9-41ab-a412-9e8ae8c38ba2\" (UID: \"833f0bba-91d9-41ab-a412-9e8ae8c38ba2\") " Dec 05 07:23:03 crc kubenswrapper[4997]: I1205 07:23:03.832162 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/833f0bba-91d9-41ab-a412-9e8ae8c38ba2-kube-api-access-nv24p" (OuterVolumeSpecName: "kube-api-access-nv24p") pod "833f0bba-91d9-41ab-a412-9e8ae8c38ba2" (UID: "833f0bba-91d9-41ab-a412-9e8ae8c38ba2"). InnerVolumeSpecName "kube-api-access-nv24p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:03 crc kubenswrapper[4997]: I1205 07:23:03.924727 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f00e97b3-98c3-41b6-9949-2961dd9524a9-catalog-content\") pod \"f00e97b3-98c3-41b6-9949-2961dd9524a9\" (UID: \"f00e97b3-98c3-41b6-9949-2961dd9524a9\") " Dec 05 07:23:03 crc kubenswrapper[4997]: I1205 07:23:03.924841 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szrlr\" (UniqueName: \"kubernetes.io/projected/f00e97b3-98c3-41b6-9949-2961dd9524a9-kube-api-access-szrlr\") pod \"f00e97b3-98c3-41b6-9949-2961dd9524a9\" (UID: \"f00e97b3-98c3-41b6-9949-2961dd9524a9\") " Dec 05 07:23:03 crc kubenswrapper[4997]: I1205 07:23:03.925252 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f00e97b3-98c3-41b6-9949-2961dd9524a9-utilities\") pod \"f00e97b3-98c3-41b6-9949-2961dd9524a9\" (UID: \"f00e97b3-98c3-41b6-9949-2961dd9524a9\") " Dec 05 07:23:03 crc kubenswrapper[4997]: I1205 07:23:03.926074 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nv24p\" (UniqueName: \"kubernetes.io/projected/833f0bba-91d9-41ab-a412-9e8ae8c38ba2-kube-api-access-nv24p\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:03 crc kubenswrapper[4997]: I1205 07:23:03.928488 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f00e97b3-98c3-41b6-9949-2961dd9524a9-utilities" (OuterVolumeSpecName: "utilities") pod "f00e97b3-98c3-41b6-9949-2961dd9524a9" (UID: "f00e97b3-98c3-41b6-9949-2961dd9524a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:03 crc kubenswrapper[4997]: I1205 07:23:03.932825 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f00e97b3-98c3-41b6-9949-2961dd9524a9-kube-api-access-szrlr" (OuterVolumeSpecName: "kube-api-access-szrlr") pod "f00e97b3-98c3-41b6-9949-2961dd9524a9" (UID: "f00e97b3-98c3-41b6-9949-2961dd9524a9"). InnerVolumeSpecName "kube-api-access-szrlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:03 crc kubenswrapper[4997]: I1205 07:23:03.980876 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f00e97b3-98c3-41b6-9949-2961dd9524a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f00e97b3-98c3-41b6-9949-2961dd9524a9" (UID: "f00e97b3-98c3-41b6-9949-2961dd9524a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.027892 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f00e97b3-98c3-41b6-9949-2961dd9524a9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.028218 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szrlr\" (UniqueName: \"kubernetes.io/projected/f00e97b3-98c3-41b6-9949-2961dd9524a9-kube-api-access-szrlr\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.028288 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f00e97b3-98c3-41b6-9949-2961dd9524a9-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.255941 4997 generic.go:334] "Generic (PLEG): container finished" podID="f00e97b3-98c3-41b6-9949-2961dd9524a9" containerID="a85c29ddb4a8ae483a8bee5c910e71da300c3a157e62f1646c991dfb7f16c845" exitCode=0 Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.256198 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8d8zv" event={"ID":"f00e97b3-98c3-41b6-9949-2961dd9524a9","Type":"ContainerDied","Data":"a85c29ddb4a8ae483a8bee5c910e71da300c3a157e62f1646c991dfb7f16c845"} Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.256856 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8d8zv" event={"ID":"f00e97b3-98c3-41b6-9949-2961dd9524a9","Type":"ContainerDied","Data":"ce7a90b4187f38121ce33f75b1dd2b72343e829fe71e02dfc12b843d34c404c9"} Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.256279 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8d8zv" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.256937 4997 scope.go:117] "RemoveContainer" containerID="a85c29ddb4a8ae483a8bee5c910e71da300c3a157e62f1646c991dfb7f16c845" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.272645 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.272787 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"833f0bba-91d9-41ab-a412-9e8ae8c38ba2","Type":"ContainerDied","Data":"60369b249264cabd66d0f28c6475b7de96dd2b3529ef70939c79e935e3edce00"} Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.297684 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8d8zv"] Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.313817 4997 scope.go:117] "RemoveContainer" containerID="beda7b0a7cd405b799e7f87e5d2fa4ad91916f784caf0e4ebffdebe24cf3fc16" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.318291 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8d8zv"] Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.353061 4997 scope.go:117] "RemoveContainer" containerID="e6cdc8b7d6baf3660b5da72cb1df31c0edb9ffe1cbe62631356d380c42d3e077" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.367717 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.378917 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.389454 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 07:23:04 crc kubenswrapper[4997]: E1205 07:23:04.390277 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f00e97b3-98c3-41b6-9949-2961dd9524a9" containerName="extract-utilities" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.390301 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f00e97b3-98c3-41b6-9949-2961dd9524a9" containerName="extract-utilities" Dec 05 07:23:04 crc kubenswrapper[4997]: E1205 07:23:04.390329 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f00e97b3-98c3-41b6-9949-2961dd9524a9" containerName="registry-server" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.390339 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f00e97b3-98c3-41b6-9949-2961dd9524a9" containerName="registry-server" Dec 05 07:23:04 crc kubenswrapper[4997]: E1205 07:23:04.390359 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="833f0bba-91d9-41ab-a412-9e8ae8c38ba2" containerName="kube-state-metrics" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.390366 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="833f0bba-91d9-41ab-a412-9e8ae8c38ba2" containerName="kube-state-metrics" Dec 05 07:23:04 crc kubenswrapper[4997]: E1205 07:23:04.390401 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f00e97b3-98c3-41b6-9949-2961dd9524a9" containerName="extract-content" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.390407 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f00e97b3-98c3-41b6-9949-2961dd9524a9" containerName="extract-content" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.390606 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="833f0bba-91d9-41ab-a412-9e8ae8c38ba2" containerName="kube-state-metrics" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.391255 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f00e97b3-98c3-41b6-9949-2961dd9524a9" containerName="registry-server" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.392206 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.396082 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.396308 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.401351 4997 scope.go:117] "RemoveContainer" containerID="a85c29ddb4a8ae483a8bee5c910e71da300c3a157e62f1646c991dfb7f16c845" Dec 05 07:23:04 crc kubenswrapper[4997]: E1205 07:23:04.402460 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a85c29ddb4a8ae483a8bee5c910e71da300c3a157e62f1646c991dfb7f16c845\": container with ID starting with a85c29ddb4a8ae483a8bee5c910e71da300c3a157e62f1646c991dfb7f16c845 not found: ID does not exist" containerID="a85c29ddb4a8ae483a8bee5c910e71da300c3a157e62f1646c991dfb7f16c845" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.402500 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a85c29ddb4a8ae483a8bee5c910e71da300c3a157e62f1646c991dfb7f16c845"} err="failed to get container status \"a85c29ddb4a8ae483a8bee5c910e71da300c3a157e62f1646c991dfb7f16c845\": rpc error: code = NotFound desc = could not find container \"a85c29ddb4a8ae483a8bee5c910e71da300c3a157e62f1646c991dfb7f16c845\": container with ID starting with a85c29ddb4a8ae483a8bee5c910e71da300c3a157e62f1646c991dfb7f16c845 not found: ID does not exist" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.402529 4997 scope.go:117] "RemoveContainer" containerID="beda7b0a7cd405b799e7f87e5d2fa4ad91916f784caf0e4ebffdebe24cf3fc16" Dec 05 07:23:04 crc kubenswrapper[4997]: E1205 07:23:04.402809 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"beda7b0a7cd405b799e7f87e5d2fa4ad91916f784caf0e4ebffdebe24cf3fc16\": container with ID starting with beda7b0a7cd405b799e7f87e5d2fa4ad91916f784caf0e4ebffdebe24cf3fc16 not found: ID does not exist" containerID="beda7b0a7cd405b799e7f87e5d2fa4ad91916f784caf0e4ebffdebe24cf3fc16" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.402834 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beda7b0a7cd405b799e7f87e5d2fa4ad91916f784caf0e4ebffdebe24cf3fc16"} err="failed to get container status \"beda7b0a7cd405b799e7f87e5d2fa4ad91916f784caf0e4ebffdebe24cf3fc16\": rpc error: code = NotFound desc = could not find container \"beda7b0a7cd405b799e7f87e5d2fa4ad91916f784caf0e4ebffdebe24cf3fc16\": container with ID starting with beda7b0a7cd405b799e7f87e5d2fa4ad91916f784caf0e4ebffdebe24cf3fc16 not found: ID does not exist" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.402871 4997 scope.go:117] "RemoveContainer" containerID="e6cdc8b7d6baf3660b5da72cb1df31c0edb9ffe1cbe62631356d380c42d3e077" Dec 05 07:23:04 crc kubenswrapper[4997]: E1205 07:23:04.403132 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6cdc8b7d6baf3660b5da72cb1df31c0edb9ffe1cbe62631356d380c42d3e077\": container with ID starting with e6cdc8b7d6baf3660b5da72cb1df31c0edb9ffe1cbe62631356d380c42d3e077 not found: ID does not exist" containerID="e6cdc8b7d6baf3660b5da72cb1df31c0edb9ffe1cbe62631356d380c42d3e077" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.403157 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6cdc8b7d6baf3660b5da72cb1df31c0edb9ffe1cbe62631356d380c42d3e077"} err="failed to get container status \"e6cdc8b7d6baf3660b5da72cb1df31c0edb9ffe1cbe62631356d380c42d3e077\": rpc error: code = NotFound desc = could not find container \"e6cdc8b7d6baf3660b5da72cb1df31c0edb9ffe1cbe62631356d380c42d3e077\": container with ID starting with e6cdc8b7d6baf3660b5da72cb1df31c0edb9ffe1cbe62631356d380c42d3e077 not found: ID does not exist" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.403204 4997 scope.go:117] "RemoveContainer" containerID="a2793e5e578daef37735f5bea301361550f5c913442e8d7f862245cb9db435d5" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.407647 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.540166 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\") " pod="openstack/kube-state-metrics-0" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.540241 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\") " pod="openstack/kube-state-metrics-0" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.540380 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ht6m\" (UniqueName: \"kubernetes.io/projected/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-api-access-7ht6m\") pod \"kube-state-metrics-0\" (UID: \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\") " pod="openstack/kube-state-metrics-0" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.540448 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\") " pod="openstack/kube-state-metrics-0" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.643914 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\") " pod="openstack/kube-state-metrics-0" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.643997 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\") " pod="openstack/kube-state-metrics-0" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.644080 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ht6m\" (UniqueName: \"kubernetes.io/projected/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-api-access-7ht6m\") pod \"kube-state-metrics-0\" (UID: \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\") " pod="openstack/kube-state-metrics-0" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.644214 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\") " pod="openstack/kube-state-metrics-0" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.650399 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\") " pod="openstack/kube-state-metrics-0" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.662891 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ht6m\" (UniqueName: \"kubernetes.io/projected/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-api-access-7ht6m\") pod \"kube-state-metrics-0\" (UID: \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\") " pod="openstack/kube-state-metrics-0" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.664039 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\") " pod="openstack/kube-state-metrics-0" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.664547 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\") " pod="openstack/kube-state-metrics-0" Dec 05 07:23:04 crc kubenswrapper[4997]: I1205 07:23:04.750527 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 07:23:05 crc kubenswrapper[4997]: I1205 07:23:05.031510 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 07:23:05 crc kubenswrapper[4997]: I1205 07:23:05.335280 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:23:05 crc kubenswrapper[4997]: I1205 07:23:05.336031 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerName="ceilometer-central-agent" containerID="cri-o://7c7e2b6ed58ca875f27f8c171f5355d4ba443fea9ebf25ae53a19570e4018b3d" gracePeriod=30 Dec 05 07:23:05 crc kubenswrapper[4997]: I1205 07:23:05.337874 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerName="proxy-httpd" containerID="cri-o://dc16e86a99e27d23b43ef7dd79991efeb1c1797f058c1872f6bb5014709bbfd6" gracePeriod=30 Dec 05 07:23:05 crc kubenswrapper[4997]: I1205 07:23:05.338429 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerName="ceilometer-notification-agent" containerID="cri-o://3494f8420cb24505d7f938e7c164c46b5a8f381b4f3970efe627fc0e233bf507" gracePeriod=30 Dec 05 07:23:05 crc kubenswrapper[4997]: I1205 07:23:05.338499 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerName="sg-core" containerID="cri-o://b4bed8cd43908d62f851b25ef5e348d4883411917a38b3826aa5063962705041" gracePeriod=30 Dec 05 07:23:05 crc kubenswrapper[4997]: I1205 07:23:05.423936 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946","Type":"ContainerStarted","Data":"f3a87fd0301e468e3f56ff45dc8706a15bba67091a215495595cb39473aa2e53"} Dec 05 07:23:05 crc kubenswrapper[4997]: I1205 07:23:05.762266 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="833f0bba-91d9-41ab-a412-9e8ae8c38ba2" path="/var/lib/kubelet/pods/833f0bba-91d9-41ab-a412-9e8ae8c38ba2/volumes" Dec 05 07:23:05 crc kubenswrapper[4997]: I1205 07:23:05.763297 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f00e97b3-98c3-41b6-9949-2961dd9524a9" path="/var/lib/kubelet/pods/f00e97b3-98c3-41b6-9949-2961dd9524a9/volumes" Dec 05 07:23:06 crc kubenswrapper[4997]: I1205 07:23:06.457576 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946","Type":"ContainerStarted","Data":"d9261932ff03a509a886dc6ed5c881f90c33d3fd0c490c206516f3523389c13c"} Dec 05 07:23:06 crc kubenswrapper[4997]: I1205 07:23:06.458304 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 05 07:23:06 crc kubenswrapper[4997]: I1205 07:23:06.462551 4997 generic.go:334] "Generic (PLEG): container finished" podID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerID="dc16e86a99e27d23b43ef7dd79991efeb1c1797f058c1872f6bb5014709bbfd6" exitCode=0 Dec 05 07:23:06 crc kubenswrapper[4997]: I1205 07:23:06.462595 4997 generic.go:334] "Generic (PLEG): container finished" podID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerID="b4bed8cd43908d62f851b25ef5e348d4883411917a38b3826aa5063962705041" exitCode=2 Dec 05 07:23:06 crc kubenswrapper[4997]: I1205 07:23:06.462642 4997 generic.go:334] "Generic (PLEG): container finished" podID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerID="7c7e2b6ed58ca875f27f8c171f5355d4ba443fea9ebf25ae53a19570e4018b3d" exitCode=0 Dec 05 07:23:06 crc kubenswrapper[4997]: I1205 07:23:06.462671 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ca00af9-cdab-474a-bc2d-8d57adaf6af3","Type":"ContainerDied","Data":"dc16e86a99e27d23b43ef7dd79991efeb1c1797f058c1872f6bb5014709bbfd6"} Dec 05 07:23:06 crc kubenswrapper[4997]: I1205 07:23:06.462705 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ca00af9-cdab-474a-bc2d-8d57adaf6af3","Type":"ContainerDied","Data":"b4bed8cd43908d62f851b25ef5e348d4883411917a38b3826aa5063962705041"} Dec 05 07:23:06 crc kubenswrapper[4997]: I1205 07:23:06.462724 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ca00af9-cdab-474a-bc2d-8d57adaf6af3","Type":"ContainerDied","Data":"7c7e2b6ed58ca875f27f8c171f5355d4ba443fea9ebf25ae53a19570e4018b3d"} Dec 05 07:23:06 crc kubenswrapper[4997]: I1205 07:23:06.490730 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.030407555 podStartE2EDuration="2.490678274s" podCreationTimestamp="2025-12-05 07:23:04 +0000 UTC" firstStartedPulling="2025-12-05 07:23:05.046517029 +0000 UTC m=+1685.575424290" lastFinishedPulling="2025-12-05 07:23:05.506787738 +0000 UTC m=+1686.035695009" observedRunningTime="2025-12-05 07:23:06.486972843 +0000 UTC m=+1687.015880114" watchObservedRunningTime="2025-12-05 07:23:06.490678274 +0000 UTC m=+1687.019585545" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.388094 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.484781 4997 generic.go:334] "Generic (PLEG): container finished" podID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerID="3494f8420cb24505d7f938e7c164c46b5a8f381b4f3970efe627fc0e233bf507" exitCode=0 Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.484835 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ca00af9-cdab-474a-bc2d-8d57adaf6af3","Type":"ContainerDied","Data":"3494f8420cb24505d7f938e7c164c46b5a8f381b4f3970efe627fc0e233bf507"} Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.484869 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ca00af9-cdab-474a-bc2d-8d57adaf6af3","Type":"ContainerDied","Data":"699c4552c983b77f8ef82fe2f83784c6634322d0ae65846e6f6427f5d19efb66"} Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.484890 4997 scope.go:117] "RemoveContainer" containerID="dc16e86a99e27d23b43ef7dd79991efeb1c1797f058c1872f6bb5014709bbfd6" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.485003 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.517231 4997 scope.go:117] "RemoveContainer" containerID="b4bed8cd43908d62f851b25ef5e348d4883411917a38b3826aa5063962705041" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.539954 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-config-data\") pod \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.540100 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-sg-core-conf-yaml\") pod \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.540219 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-scripts\") pod \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.540287 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtwl6\" (UniqueName: \"kubernetes.io/projected/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-kube-api-access-xtwl6\") pod \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.540441 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-run-httpd\") pod \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.540472 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-log-httpd\") pod \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.540502 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-combined-ca-bundle\") pod \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\" (UID: \"1ca00af9-cdab-474a-bc2d-8d57adaf6af3\") " Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.541059 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1ca00af9-cdab-474a-bc2d-8d57adaf6af3" (UID: "1ca00af9-cdab-474a-bc2d-8d57adaf6af3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.541233 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1ca00af9-cdab-474a-bc2d-8d57adaf6af3" (UID: "1ca00af9-cdab-474a-bc2d-8d57adaf6af3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.541959 4997 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.541997 4997 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.542235 4997 scope.go:117] "RemoveContainer" containerID="3494f8420cb24505d7f938e7c164c46b5a8f381b4f3970efe627fc0e233bf507" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.556063 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-kube-api-access-xtwl6" (OuterVolumeSpecName: "kube-api-access-xtwl6") pod "1ca00af9-cdab-474a-bc2d-8d57adaf6af3" (UID: "1ca00af9-cdab-474a-bc2d-8d57adaf6af3"). InnerVolumeSpecName "kube-api-access-xtwl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.556162 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-scripts" (OuterVolumeSpecName: "scripts") pod "1ca00af9-cdab-474a-bc2d-8d57adaf6af3" (UID: "1ca00af9-cdab-474a-bc2d-8d57adaf6af3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.587356 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1ca00af9-cdab-474a-bc2d-8d57adaf6af3" (UID: "1ca00af9-cdab-474a-bc2d-8d57adaf6af3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.644525 4997 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.644554 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.644564 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtwl6\" (UniqueName: \"kubernetes.io/projected/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-kube-api-access-xtwl6\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.653209 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ca00af9-cdab-474a-bc2d-8d57adaf6af3" (UID: "1ca00af9-cdab-474a-bc2d-8d57adaf6af3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.675733 4997 scope.go:117] "RemoveContainer" containerID="7c7e2b6ed58ca875f27f8c171f5355d4ba443fea9ebf25ae53a19570e4018b3d" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.683243 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-config-data" (OuterVolumeSpecName: "config-data") pod "1ca00af9-cdab-474a-bc2d-8d57adaf6af3" (UID: "1ca00af9-cdab-474a-bc2d-8d57adaf6af3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.705420 4997 scope.go:117] "RemoveContainer" containerID="dc16e86a99e27d23b43ef7dd79991efeb1c1797f058c1872f6bb5014709bbfd6" Dec 05 07:23:08 crc kubenswrapper[4997]: E1205 07:23:08.706248 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc16e86a99e27d23b43ef7dd79991efeb1c1797f058c1872f6bb5014709bbfd6\": container with ID starting with dc16e86a99e27d23b43ef7dd79991efeb1c1797f058c1872f6bb5014709bbfd6 not found: ID does not exist" containerID="dc16e86a99e27d23b43ef7dd79991efeb1c1797f058c1872f6bb5014709bbfd6" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.706328 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc16e86a99e27d23b43ef7dd79991efeb1c1797f058c1872f6bb5014709bbfd6"} err="failed to get container status \"dc16e86a99e27d23b43ef7dd79991efeb1c1797f058c1872f6bb5014709bbfd6\": rpc error: code = NotFound desc = could not find container \"dc16e86a99e27d23b43ef7dd79991efeb1c1797f058c1872f6bb5014709bbfd6\": container with ID starting with dc16e86a99e27d23b43ef7dd79991efeb1c1797f058c1872f6bb5014709bbfd6 not found: ID does not exist" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.706368 4997 scope.go:117] "RemoveContainer" containerID="b4bed8cd43908d62f851b25ef5e348d4883411917a38b3826aa5063962705041" Dec 05 07:23:08 crc kubenswrapper[4997]: E1205 07:23:08.706897 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4bed8cd43908d62f851b25ef5e348d4883411917a38b3826aa5063962705041\": container with ID starting with b4bed8cd43908d62f851b25ef5e348d4883411917a38b3826aa5063962705041 not found: ID does not exist" containerID="b4bed8cd43908d62f851b25ef5e348d4883411917a38b3826aa5063962705041" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.706939 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4bed8cd43908d62f851b25ef5e348d4883411917a38b3826aa5063962705041"} err="failed to get container status \"b4bed8cd43908d62f851b25ef5e348d4883411917a38b3826aa5063962705041\": rpc error: code = NotFound desc = could not find container \"b4bed8cd43908d62f851b25ef5e348d4883411917a38b3826aa5063962705041\": container with ID starting with b4bed8cd43908d62f851b25ef5e348d4883411917a38b3826aa5063962705041 not found: ID does not exist" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.706968 4997 scope.go:117] "RemoveContainer" containerID="3494f8420cb24505d7f938e7c164c46b5a8f381b4f3970efe627fc0e233bf507" Dec 05 07:23:08 crc kubenswrapper[4997]: E1205 07:23:08.707368 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3494f8420cb24505d7f938e7c164c46b5a8f381b4f3970efe627fc0e233bf507\": container with ID starting with 3494f8420cb24505d7f938e7c164c46b5a8f381b4f3970efe627fc0e233bf507 not found: ID does not exist" containerID="3494f8420cb24505d7f938e7c164c46b5a8f381b4f3970efe627fc0e233bf507" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.707393 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3494f8420cb24505d7f938e7c164c46b5a8f381b4f3970efe627fc0e233bf507"} err="failed to get container status \"3494f8420cb24505d7f938e7c164c46b5a8f381b4f3970efe627fc0e233bf507\": rpc error: code = NotFound desc = could not find container \"3494f8420cb24505d7f938e7c164c46b5a8f381b4f3970efe627fc0e233bf507\": container with ID starting with 3494f8420cb24505d7f938e7c164c46b5a8f381b4f3970efe627fc0e233bf507 not found: ID does not exist" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.707412 4997 scope.go:117] "RemoveContainer" containerID="7c7e2b6ed58ca875f27f8c171f5355d4ba443fea9ebf25ae53a19570e4018b3d" Dec 05 07:23:08 crc kubenswrapper[4997]: E1205 07:23:08.707794 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c7e2b6ed58ca875f27f8c171f5355d4ba443fea9ebf25ae53a19570e4018b3d\": container with ID starting with 7c7e2b6ed58ca875f27f8c171f5355d4ba443fea9ebf25ae53a19570e4018b3d not found: ID does not exist" containerID="7c7e2b6ed58ca875f27f8c171f5355d4ba443fea9ebf25ae53a19570e4018b3d" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.707818 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c7e2b6ed58ca875f27f8c171f5355d4ba443fea9ebf25ae53a19570e4018b3d"} err="failed to get container status \"7c7e2b6ed58ca875f27f8c171f5355d4ba443fea9ebf25ae53a19570e4018b3d\": rpc error: code = NotFound desc = could not find container \"7c7e2b6ed58ca875f27f8c171f5355d4ba443fea9ebf25ae53a19570e4018b3d\": container with ID starting with 7c7e2b6ed58ca875f27f8c171f5355d4ba443fea9ebf25ae53a19570e4018b3d not found: ID does not exist" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.747047 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.747105 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ca00af9-cdab-474a-bc2d-8d57adaf6af3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.829435 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.861362 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.885762 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:23:08 crc kubenswrapper[4997]: E1205 07:23:08.886332 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerName="sg-core" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.886351 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerName="sg-core" Dec 05 07:23:08 crc kubenswrapper[4997]: E1205 07:23:08.886368 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerName="ceilometer-notification-agent" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.886375 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerName="ceilometer-notification-agent" Dec 05 07:23:08 crc kubenswrapper[4997]: E1205 07:23:08.886390 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerName="ceilometer-central-agent" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.886396 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerName="ceilometer-central-agent" Dec 05 07:23:08 crc kubenswrapper[4997]: E1205 07:23:08.886420 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerName="proxy-httpd" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.886425 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerName="proxy-httpd" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.886628 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerName="ceilometer-notification-agent" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.886648 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerName="proxy-httpd" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.886656 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerName="ceilometer-central-agent" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.886671 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" containerName="sg-core" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.888622 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.893018 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.893290 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.893440 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 05 07:23:08 crc kubenswrapper[4997]: I1205 07:23:08.895217 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.055393 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-config-data\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.055443 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.055502 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-scripts\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.055538 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.055567 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db136058-17a8-4955-8e95-576da7b9e847-run-httpd\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.055679 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db136058-17a8-4955-8e95-576da7b9e847-log-httpd\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.055771 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmrcw\" (UniqueName: \"kubernetes.io/projected/db136058-17a8-4955-8e95-576da7b9e847-kube-api-access-dmrcw\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.055959 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.158842 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-scripts\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.158918 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.158951 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db136058-17a8-4955-8e95-576da7b9e847-run-httpd\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.158979 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db136058-17a8-4955-8e95-576da7b9e847-log-httpd\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.159059 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmrcw\" (UniqueName: \"kubernetes.io/projected/db136058-17a8-4955-8e95-576da7b9e847-kube-api-access-dmrcw\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.159105 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.159145 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-config-data\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.159167 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.159858 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db136058-17a8-4955-8e95-576da7b9e847-log-httpd\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.159904 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db136058-17a8-4955-8e95-576da7b9e847-run-httpd\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.164433 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.166323 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.167683 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-config-data\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.168549 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.171602 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-scripts\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.176878 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmrcw\" (UniqueName: \"kubernetes.io/projected/db136058-17a8-4955-8e95-576da7b9e847-kube-api-access-dmrcw\") pod \"ceilometer-0\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.261543 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.585423 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:23:09 crc kubenswrapper[4997]: W1205 07:23:09.597953 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb136058_17a8_4955_8e95_576da7b9e847.slice/crio-0d961eed77dab6e083f388b65d5739d3ebd5af4ac39cadc5d8e1338bcaa03957 WatchSource:0}: Error finding container 0d961eed77dab6e083f388b65d5739d3ebd5af4ac39cadc5d8e1338bcaa03957: Status 404 returned error can't find the container with id 0d961eed77dab6e083f388b65d5739d3ebd5af4ac39cadc5d8e1338bcaa03957 Dec 05 07:23:09 crc kubenswrapper[4997]: I1205 07:23:09.764471 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ca00af9-cdab-474a-bc2d-8d57adaf6af3" path="/var/lib/kubelet/pods/1ca00af9-cdab-474a-bc2d-8d57adaf6af3/volumes" Dec 05 07:23:10 crc kubenswrapper[4997]: I1205 07:23:10.518105 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db136058-17a8-4955-8e95-576da7b9e847","Type":"ContainerStarted","Data":"07f1631824a73521a5df4d80b3bf4feb8dde04c7a5b2c5e5688b95c1efabb84c"} Dec 05 07:23:10 crc kubenswrapper[4997]: I1205 07:23:10.518459 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db136058-17a8-4955-8e95-576da7b9e847","Type":"ContainerStarted","Data":"0d961eed77dab6e083f388b65d5739d3ebd5af4ac39cadc5d8e1338bcaa03957"} Dec 05 07:23:11 crc kubenswrapper[4997]: I1205 07:23:11.551216 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db136058-17a8-4955-8e95-576da7b9e847","Type":"ContainerStarted","Data":"b3cfedfae8a42f278d8fff88f2e1586ca243b6222c804510c5b0435a9a5e1e73"} Dec 05 07:23:12 crc kubenswrapper[4997]: I1205 07:23:12.570139 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db136058-17a8-4955-8e95-576da7b9e847","Type":"ContainerStarted","Data":"3fc11187932225dbd83eb0de61fd45f5e409020f11e6e1a8a12baa4ea259218b"} Dec 05 07:23:13 crc kubenswrapper[4997]: I1205 07:23:13.588073 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db136058-17a8-4955-8e95-576da7b9e847","Type":"ContainerStarted","Data":"fdb72dc3adf6464397dca00e5b9745e3a6b89a71985325529d858ab69ce19e3c"} Dec 05 07:23:13 crc kubenswrapper[4997]: I1205 07:23:13.588972 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 07:23:13 crc kubenswrapper[4997]: I1205 07:23:13.620575 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.619457854 podStartE2EDuration="5.62054796s" podCreationTimestamp="2025-12-05 07:23:08 +0000 UTC" firstStartedPulling="2025-12-05 07:23:09.601817124 +0000 UTC m=+1690.130724385" lastFinishedPulling="2025-12-05 07:23:12.60290723 +0000 UTC m=+1693.131814491" observedRunningTime="2025-12-05 07:23:13.617721344 +0000 UTC m=+1694.146628615" watchObservedRunningTime="2025-12-05 07:23:13.62054796 +0000 UTC m=+1694.149455221" Dec 05 07:23:14 crc kubenswrapper[4997]: I1205 07:23:14.769646 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 05 07:23:18 crc kubenswrapper[4997]: I1205 07:23:18.750854 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:23:18 crc kubenswrapper[4997]: E1205 07:23:18.752680 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.394243 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.412066 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="fe32fb42-61e6-4cd4-a75b-b684a4171224" containerName="cinder-scheduler" containerID="cri-o://b762656e8963968dd62e28b778f1374e72ed4bfb67fbcee7edeacd3920071001" gracePeriod=30 Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.412681 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="fe32fb42-61e6-4cd4-a75b-b684a4171224" containerName="probe" containerID="cri-o://921875cb2f26d740baa583b8c4418587dfe774e4ea61c9ad78bb1723d653c892" gracePeriod=30 Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.600925 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.737722 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.738044 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="126f5f4c-a3f5-4744-99cd-691bdd33444b" containerName="openstackclient" containerID="cri-o://6ea597f614ba1df9ed5420084187fc9eec112551e2d7b84b30b72cdf64053515" gracePeriod=2 Dec 05 07:23:21 crc kubenswrapper[4997]: E1205 07:23:21.762210 4997 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 05 07:23:21 crc kubenswrapper[4997]: E1205 07:23:21.762299 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-config-data podName:1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:22.262277241 +0000 UTC m=+1702.791184502 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-config-data") pod "rabbitmq-server-0" (UID: "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79") : configmap "rabbitmq-config-data" not found Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.796036 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.796089 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.796105 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.796344 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0d3915c5-30d2-43be-abd4-438cd9d8ebf7" containerName="cinder-api-log" containerID="cri-o://057a82bd61c9e67d2a9d3ed1d85e6973ab9ffca83e3bd81b232a535bb4ad1866" gracePeriod=30 Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.796904 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0d3915c5-30d2-43be-abd4-438cd9d8ebf7" containerName="cinder-api" containerID="cri-o://8f36673252ccfc4fc6b7d58b342de4054e621d8bc6722f917b4cf09a6dd9bcaf" gracePeriod=30 Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.796980 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="7c6db4c3-5869-4a82-900c-5718abfe3a3b" containerName="openstack-network-exporter" containerID="cri-o://5d79e06e59e6f9cd89fbe9c40174b9a6f2059177958ca28bb43f3b31b4fdf3bd" gracePeriod=300 Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.829560 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron84af-account-delete-pgl29"] Dec 05 07:23:21 crc kubenswrapper[4997]: E1205 07:23:21.834034 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="126f5f4c-a3f5-4744-99cd-691bdd33444b" containerName="openstackclient" Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.834075 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="126f5f4c-a3f5-4744-99cd-691bdd33444b" containerName="openstackclient" Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.834328 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="126f5f4c-a3f5-4744-99cd-691bdd33444b" containerName="openstackclient" Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.845004 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron84af-account-delete-pgl29" Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.897676 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cindera458-account-delete-vpgnx"] Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.899513 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cindera458-account-delete-vpgnx" Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.924704 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron84af-account-delete-pgl29"] Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.941322 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cindera458-account-delete-vpgnx"] Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.970777 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68490a68-dc32-448f-90cc-c146ba49c1a8-operator-scripts\") pod \"neutron84af-account-delete-pgl29\" (UID: \"68490a68-dc32-448f-90cc-c146ba49c1a8\") " pod="openstack/neutron84af-account-delete-pgl29" Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.971002 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2lxh\" (UniqueName: \"kubernetes.io/projected/68490a68-dc32-448f-90cc-c146ba49c1a8-kube-api-access-w2lxh\") pod \"neutron84af-account-delete-pgl29\" (UID: \"68490a68-dc32-448f-90cc-c146ba49c1a8\") " pod="openstack/neutron84af-account-delete-pgl29" Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.971178 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm6mx\" (UniqueName: \"kubernetes.io/projected/b8c5aa81-a666-4ae6-a41c-fd31df260223-kube-api-access-xm6mx\") pod \"cindera458-account-delete-vpgnx\" (UID: \"b8c5aa81-a666-4ae6-a41c-fd31df260223\") " pod="openstack/cindera458-account-delete-vpgnx" Dec 05 07:23:21 crc kubenswrapper[4997]: I1205 07:23:21.971285 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8c5aa81-a666-4ae6-a41c-fd31df260223-operator-scripts\") pod \"cindera458-account-delete-vpgnx\" (UID: \"b8c5aa81-a666-4ae6-a41c-fd31df260223\") " pod="openstack/cindera458-account-delete-vpgnx" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.073522 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68490a68-dc32-448f-90cc-c146ba49c1a8-operator-scripts\") pod \"neutron84af-account-delete-pgl29\" (UID: \"68490a68-dc32-448f-90cc-c146ba49c1a8\") " pod="openstack/neutron84af-account-delete-pgl29" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.073921 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.075062 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2lxh\" (UniqueName: \"kubernetes.io/projected/68490a68-dc32-448f-90cc-c146ba49c1a8-kube-api-access-w2lxh\") pod \"neutron84af-account-delete-pgl29\" (UID: \"68490a68-dc32-448f-90cc-c146ba49c1a8\") " pod="openstack/neutron84af-account-delete-pgl29" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.075157 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm6mx\" (UniqueName: \"kubernetes.io/projected/b8c5aa81-a666-4ae6-a41c-fd31df260223-kube-api-access-xm6mx\") pod \"cindera458-account-delete-vpgnx\" (UID: \"b8c5aa81-a666-4ae6-a41c-fd31df260223\") " pod="openstack/cindera458-account-delete-vpgnx" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.075209 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8c5aa81-a666-4ae6-a41c-fd31df260223-operator-scripts\") pod \"cindera458-account-delete-vpgnx\" (UID: \"b8c5aa81-a666-4ae6-a41c-fd31df260223\") " pod="openstack/cindera458-account-delete-vpgnx" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.076203 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8c5aa81-a666-4ae6-a41c-fd31df260223-operator-scripts\") pod \"cindera458-account-delete-vpgnx\" (UID: \"b8c5aa81-a666-4ae6-a41c-fd31df260223\") " pod="openstack/cindera458-account-delete-vpgnx" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.076790 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68490a68-dc32-448f-90cc-c146ba49c1a8-operator-scripts\") pod \"neutron84af-account-delete-pgl29\" (UID: \"68490a68-dc32-448f-90cc-c146ba49c1a8\") " pod="openstack/neutron84af-account-delete-pgl29" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.138595 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican69df-account-delete-tqlsz"] Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.140498 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican69df-account-delete-tqlsz" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.157400 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm6mx\" (UniqueName: \"kubernetes.io/projected/b8c5aa81-a666-4ae6-a41c-fd31df260223-kube-api-access-xm6mx\") pod \"cindera458-account-delete-vpgnx\" (UID: \"b8c5aa81-a666-4ae6-a41c-fd31df260223\") " pod="openstack/cindera458-account-delete-vpgnx" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.157582 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2lxh\" (UniqueName: \"kubernetes.io/projected/68490a68-dc32-448f-90cc-c146ba49c1a8-kube-api-access-w2lxh\") pod \"neutron84af-account-delete-pgl29\" (UID: \"68490a68-dc32-448f-90cc-c146ba49c1a8\") " pod="openstack/neutron84af-account-delete-pgl29" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.177370 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron84af-account-delete-pgl29" Dec 05 07:23:22 crc kubenswrapper[4997]: E1205 07:23:22.185303 4997 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 05 07:23:22 crc kubenswrapper[4997]: E1205 07:23:22.185371 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-config-data podName:0f158874-152b-46ce-ac42-d202684853ca nodeName:}" failed. No retries permitted until 2025-12-05 07:23:22.685351621 +0000 UTC m=+1703.214258882 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-config-data") pod "rabbitmq-cell1-server-0" (UID: "0f158874-152b-46ce-ac42-d202684853ca") : configmap "rabbitmq-cell1-config-data" not found Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.193406 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican69df-account-delete-tqlsz"] Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.289472 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cindera458-account-delete-vpgnx" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.365100 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nb7s\" (UniqueName: \"kubernetes.io/projected/d65cd492-d4bd-4489-9d28-a6c42407101a-kube-api-access-9nb7s\") pod \"barbican69df-account-delete-tqlsz\" (UID: \"d65cd492-d4bd-4489-9d28-a6c42407101a\") " pod="openstack/barbican69df-account-delete-tqlsz" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.365247 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d65cd492-d4bd-4489-9d28-a6c42407101a-operator-scripts\") pod \"barbican69df-account-delete-tqlsz\" (UID: \"d65cd492-d4bd-4489-9d28-a6c42407101a\") " pod="openstack/barbican69df-account-delete-tqlsz" Dec 05 07:23:22 crc kubenswrapper[4997]: E1205 07:23:22.365464 4997 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 05 07:23:22 crc kubenswrapper[4997]: E1205 07:23:22.365542 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-config-data podName:1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:23.365515126 +0000 UTC m=+1703.894422387 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-config-data") pod "rabbitmq-server-0" (UID: "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79") : configmap "rabbitmq-config-data" not found Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.560954 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nb7s\" (UniqueName: \"kubernetes.io/projected/d65cd492-d4bd-4489-9d28-a6c42407101a-kube-api-access-9nb7s\") pod \"barbican69df-account-delete-tqlsz\" (UID: \"d65cd492-d4bd-4489-9d28-a6c42407101a\") " pod="openstack/barbican69df-account-delete-tqlsz" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.561260 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d65cd492-d4bd-4489-9d28-a6c42407101a-operator-scripts\") pod \"barbican69df-account-delete-tqlsz\" (UID: \"d65cd492-d4bd-4489-9d28-a6c42407101a\") " pod="openstack/barbican69df-account-delete-tqlsz" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.565099 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d65cd492-d4bd-4489-9d28-a6c42407101a-operator-scripts\") pod \"barbican69df-account-delete-tqlsz\" (UID: \"d65cd492-d4bd-4489-9d28-a6c42407101a\") " pod="openstack/barbican69df-account-delete-tqlsz" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.612903 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.613192 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="2f0e39f0-ca27-4656-879e-c7a20fcd904c" containerName="ovn-northd" containerID="cri-o://3caac945976009c44a6d871779a90a30a1333a9614f3564c29ec3e0c616fdbe7" gracePeriod=30 Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.615741 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="2f0e39f0-ca27-4656-879e-c7a20fcd904c" containerName="openstack-network-exporter" containerID="cri-o://b7780da612c7177ca63143d557d61798aed63fe6008628f0f9c5e7fba16024f1" gracePeriod=30 Dec 05 07:23:22 crc kubenswrapper[4997]: E1205 07:23:22.658288 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3caac945976009c44a6d871779a90a30a1333a9614f3564c29ec3e0c616fdbe7" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.684224 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="7c6db4c3-5869-4a82-900c-5718abfe3a3b" containerName="ovsdbserver-sb" containerID="cri-o://b92bd2595a4bc2bb590760b577f8f2341220a48a3841768940be448c3738ff62" gracePeriod=300 Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.714705 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement94df-account-delete-4qxc6"] Dec 05 07:23:22 crc kubenswrapper[4997]: E1205 07:23:22.714829 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3caac945976009c44a6d871779a90a30a1333a9614f3564c29ec3e0c616fdbe7" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.716415 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement94df-account-delete-4qxc6" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.719159 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nb7s\" (UniqueName: \"kubernetes.io/projected/d65cd492-d4bd-4489-9d28-a6c42407101a-kube-api-access-9nb7s\") pod \"barbican69df-account-delete-tqlsz\" (UID: \"d65cd492-d4bd-4489-9d28-a6c42407101a\") " pod="openstack/barbican69df-account-delete-tqlsz" Dec 05 07:23:22 crc kubenswrapper[4997]: E1205 07:23:22.730201 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3caac945976009c44a6d871779a90a30a1333a9614f3564c29ec3e0c616fdbe7" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 05 07:23:22 crc kubenswrapper[4997]: E1205 07:23:22.730263 4997 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Liveness" pod="openstack/ovn-northd-0" podUID="2f0e39f0-ca27-4656-879e-c7a20fcd904c" containerName="ovn-northd" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.751475 4997 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/ovn-controller-vnncj" secret="" err="secret \"ovncontroller-ovncontroller-dockercfg-xghdc\" not found" Dec 05 07:23:22 crc kubenswrapper[4997]: E1205 07:23:22.789932 4997 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 05 07:23:22 crc kubenswrapper[4997]: E1205 07:23:22.790012 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-config-data podName:0f158874-152b-46ce-ac42-d202684853ca nodeName:}" failed. No retries permitted until 2025-12-05 07:23:23.789992714 +0000 UTC m=+1704.318899975 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-config-data") pod "rabbitmq-cell1-server-0" (UID: "0f158874-152b-46ce-ac42-d202684853ca") : configmap "rabbitmq-cell1-config-data" not found Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.791777 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement94df-account-delete-4qxc6"] Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.828525 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell1418a-account-delete-76wq7"] Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.832362 4997 generic.go:334] "Generic (PLEG): container finished" podID="0d3915c5-30d2-43be-abd4-438cd9d8ebf7" containerID="057a82bd61c9e67d2a9d3ed1d85e6973ab9ffca83e3bd81b232a535bb4ad1866" exitCode=143 Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.845057 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d3915c5-30d2-43be-abd4-438cd9d8ebf7","Type":"ContainerDied","Data":"057a82bd61c9e67d2a9d3ed1d85e6973ab9ffca83e3bd81b232a535bb4ad1866"} Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.845217 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell1418a-account-delete-76wq7" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.858368 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican69df-account-delete-tqlsz" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.893246 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell1418a-account-delete-76wq7"] Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.915536 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c51acefb-ed8a-480c-9ab6-0345f2ff0cd3-operator-scripts\") pod \"placement94df-account-delete-4qxc6\" (UID: \"c51acefb-ed8a-480c-9ab6-0345f2ff0cd3\") " pod="openstack/placement94df-account-delete-4qxc6" Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.915730 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6s6q\" (UniqueName: \"kubernetes.io/projected/c51acefb-ed8a-480c-9ab6-0345f2ff0cd3-kube-api-access-q6s6q\") pod \"placement94df-account-delete-4qxc6\" (UID: \"c51acefb-ed8a-480c-9ab6-0345f2ff0cd3\") " pod="openstack/placement94df-account-delete-4qxc6" Dec 05 07:23:22 crc kubenswrapper[4997]: E1205 07:23:22.916058 4997 configmap.go:193] Couldn't get configMap openstack/ovncontroller-scripts: configmap "ovncontroller-scripts" not found Dec 05 07:23:22 crc kubenswrapper[4997]: E1205 07:23:22.916120 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-scripts podName:5e4237cc-8581-4ddb-b472-e1c6a4a41cb5 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:23.416103883 +0000 UTC m=+1703.945011144 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/configmap/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-scripts") pod "ovn-controller-vnncj" (UID: "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5") : configmap "ovncontroller-scripts" not found Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.950196 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-92zcp"] Dec 05 07:23:22 crc kubenswrapper[4997]: I1205 07:23:22.980373 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-92zcp"] Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.009176 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-k8cjm"] Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.018070 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz6hx\" (UniqueName: \"kubernetes.io/projected/77e3250e-2f69-4047-8ea9-cce3b817cdb8-kube-api-access-qz6hx\") pod \"novacell1418a-account-delete-76wq7\" (UID: \"77e3250e-2f69-4047-8ea9-cce3b817cdb8\") " pod="openstack/novacell1418a-account-delete-76wq7" Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.018171 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e3250e-2f69-4047-8ea9-cce3b817cdb8-operator-scripts\") pod \"novacell1418a-account-delete-76wq7\" (UID: \"77e3250e-2f69-4047-8ea9-cce3b817cdb8\") " pod="openstack/novacell1418a-account-delete-76wq7" Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.018262 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c51acefb-ed8a-480c-9ab6-0345f2ff0cd3-operator-scripts\") pod \"placement94df-account-delete-4qxc6\" (UID: \"c51acefb-ed8a-480c-9ab6-0345f2ff0cd3\") " pod="openstack/placement94df-account-delete-4qxc6" Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.018380 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6s6q\" (UniqueName: \"kubernetes.io/projected/c51acefb-ed8a-480c-9ab6-0345f2ff0cd3-kube-api-access-q6s6q\") pod \"placement94df-account-delete-4qxc6\" (UID: \"c51acefb-ed8a-480c-9ab6-0345f2ff0cd3\") " pod="openstack/placement94df-account-delete-4qxc6" Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.019070 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c51acefb-ed8a-480c-9ab6-0345f2ff0cd3-operator-scripts\") pod \"placement94df-account-delete-4qxc6\" (UID: \"c51acefb-ed8a-480c-9ab6-0345f2ff0cd3\") " pod="openstack/placement94df-account-delete-4qxc6" Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.074553 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-h79p5"] Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.104117 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-h79p5"] Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.104765 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6s6q\" (UniqueName: \"kubernetes.io/projected/c51acefb-ed8a-480c-9ab6-0345f2ff0cd3-kube-api-access-q6s6q\") pod \"placement94df-account-delete-4qxc6\" (UID: \"c51acefb-ed8a-480c-9ab6-0345f2ff0cd3\") " pod="openstack/placement94df-account-delete-4qxc6" Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.118294 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement94df-account-delete-4qxc6" Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.119963 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz6hx\" (UniqueName: \"kubernetes.io/projected/77e3250e-2f69-4047-8ea9-cce3b817cdb8-kube-api-access-qz6hx\") pod \"novacell1418a-account-delete-76wq7\" (UID: \"77e3250e-2f69-4047-8ea9-cce3b817cdb8\") " pod="openstack/novacell1418a-account-delete-76wq7" Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.120030 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e3250e-2f69-4047-8ea9-cce3b817cdb8-operator-scripts\") pod \"novacell1418a-account-delete-76wq7\" (UID: \"77e3250e-2f69-4047-8ea9-cce3b817cdb8\") " pod="openstack/novacell1418a-account-delete-76wq7" Dec 05 07:23:23 crc kubenswrapper[4997]: E1205 07:23:23.120163 4997 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Dec 05 07:23:23 crc kubenswrapper[4997]: E1205 07:23:23.120224 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/77e3250e-2f69-4047-8ea9-cce3b817cdb8-operator-scripts podName:77e3250e-2f69-4047-8ea9-cce3b817cdb8 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:23.620200767 +0000 UTC m=+1704.149108028 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/77e3250e-2f69-4047-8ea9-cce3b817cdb8-operator-scripts") pod "novacell1418a-account-delete-76wq7" (UID: "77e3250e-2f69-4047-8ea9-cce3b817cdb8") : configmap "openstack-cell1-scripts" not found Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.138521 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-k8cjm"] Dec 05 07:23:23 crc kubenswrapper[4997]: E1205 07:23:23.154823 4997 projected.go:194] Error preparing data for projected volume kube-api-access-qz6hx for pod openstack/novacell1418a-account-delete-76wq7: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 05 07:23:23 crc kubenswrapper[4997]: E1205 07:23:23.154920 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77e3250e-2f69-4047-8ea9-cce3b817cdb8-kube-api-access-qz6hx podName:77e3250e-2f69-4047-8ea9-cce3b817cdb8 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:23.654894497 +0000 UTC m=+1704.183801758 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-qz6hx" (UniqueName: "kubernetes.io/projected/77e3250e-2f69-4047-8ea9-cce3b817cdb8-kube-api-access-qz6hx") pod "novacell1418a-account-delete-76wq7" (UID: "77e3250e-2f69-4047-8ea9-cce3b817cdb8") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.176950 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell02ec3-account-delete-8p9cc"] Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.178555 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell02ec3-account-delete-8p9cc" Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.195380 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-6lm2t"] Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.218424 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-6lm2t"] Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.269190 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell02ec3-account-delete-8p9cc"] Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.291347 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-89tk2"] Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.291651 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-89tk2" podUID="cb2d44ce-8c93-493e-9393-569c208f076f" containerName="openstack-network-exporter" containerID="cri-o://c77fadb14da56ec3ffb9023a9bbf4752c6ab4d4d2977ec528e7ff29efb0be39c" gracePeriod=30 Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.318767 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-sctbp"] Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.324431 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf72e65a-bae4-4ce4-b65a-d5a587ef5f48-operator-scripts\") pod \"novacell02ec3-account-delete-8p9cc\" (UID: \"cf72e65a-bae4-4ce4-b65a-d5a587ef5f48\") " pod="openstack/novacell02ec3-account-delete-8p9cc" Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.324557 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tklqq\" (UniqueName: \"kubernetes.io/projected/cf72e65a-bae4-4ce4-b65a-d5a587ef5f48-kube-api-access-tklqq\") pod \"novacell02ec3-account-delete-8p9cc\" (UID: \"cf72e65a-bae4-4ce4-b65a-d5a587ef5f48\") " pod="openstack/novacell02ec3-account-delete-8p9cc" Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.348158 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="2f0e39f0-ca27-4656-879e-c7a20fcd904c" containerName="ovn-northd" probeResult="failure" output=< Dec 05 07:23:23 crc kubenswrapper[4997]: 2025-12-05T07:23:22Z|00001|unixctl|WARN|failed to connect to /tmp/ovn-northd.1.ctl Dec 05 07:23:23 crc kubenswrapper[4997]: ovn-appctl: cannot connect to "/tmp/ovn-northd.1.ctl" (No such file or directory) Dec 05 07:23:23 crc kubenswrapper[4997]: 2025-12-05T07:23:23Z|00001|unixctl|WARN|failed to connect to /tmp/ovn-northd.1.ctl Dec 05 07:23:23 crc kubenswrapper[4997]: ovn-appctl: cannot connect to "/tmp/ovn-northd.1.ctl" (No such file or directory) Dec 05 07:23:23 crc kubenswrapper[4997]: > Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.381749 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-vnncj"] Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.406698 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-dv6nr"] Dec 05 07:23:23 crc kubenswrapper[4997]: E1205 07:23:23.419833 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b92bd2595a4bc2bb590760b577f8f2341220a48a3841768940be448c3738ff62 is running failed: container process not found" containerID="b92bd2595a4bc2bb590760b577f8f2341220a48a3841768940be448c3738ff62" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 05 07:23:23 crc kubenswrapper[4997]: E1205 07:23:23.424913 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b92bd2595a4bc2bb590760b577f8f2341220a48a3841768940be448c3738ff62 is running failed: container process not found" containerID="b92bd2595a4bc2bb590760b577f8f2341220a48a3841768940be448c3738ff62" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.428472 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf72e65a-bae4-4ce4-b65a-d5a587ef5f48-operator-scripts\") pod \"novacell02ec3-account-delete-8p9cc\" (UID: \"cf72e65a-bae4-4ce4-b65a-d5a587ef5f48\") " pod="openstack/novacell02ec3-account-delete-8p9cc" Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.428607 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tklqq\" (UniqueName: \"kubernetes.io/projected/cf72e65a-bae4-4ce4-b65a-d5a587ef5f48-kube-api-access-tklqq\") pod \"novacell02ec3-account-delete-8p9cc\" (UID: \"cf72e65a-bae4-4ce4-b65a-d5a587ef5f48\") " pod="openstack/novacell02ec3-account-delete-8p9cc" Dec 05 07:23:23 crc kubenswrapper[4997]: E1205 07:23:23.430850 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b92bd2595a4bc2bb590760b577f8f2341220a48a3841768940be448c3738ff62 is running failed: container process not found" containerID="b92bd2595a4bc2bb590760b577f8f2341220a48a3841768940be448c3738ff62" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 05 07:23:23 crc kubenswrapper[4997]: E1205 07:23:23.430949 4997 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b92bd2595a4bc2bb590760b577f8f2341220a48a3841768940be448c3738ff62 is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="7c6db4c3-5869-4a82-900c-5718abfe3a3b" containerName="ovsdbserver-sb" Dec 05 07:23:23 crc kubenswrapper[4997]: E1205 07:23:23.431326 4997 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 05 07:23:23 crc kubenswrapper[4997]: E1205 07:23:23.431389 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-config-data podName:1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:25.431367873 +0000 UTC m=+1705.960275134 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-config-data") pod "rabbitmq-server-0" (UID: "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79") : configmap "rabbitmq-config-data" not found Dec 05 07:23:23 crc kubenswrapper[4997]: E1205 07:23:23.432061 4997 configmap.go:193] Couldn't get configMap openstack/ovncontroller-scripts: configmap "ovncontroller-scripts" not found Dec 05 07:23:23 crc kubenswrapper[4997]: E1205 07:23:23.432132 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-scripts podName:5e4237cc-8581-4ddb-b472-e1c6a4a41cb5 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:24.432113153 +0000 UTC m=+1704.961020414 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/configmap/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-scripts") pod "ovn-controller-vnncj" (UID: "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5") : configmap "ovncontroller-scripts" not found Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.432190 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf72e65a-bae4-4ce4-b65a-d5a587ef5f48-operator-scripts\") pod \"novacell02ec3-account-delete-8p9cc\" (UID: \"cf72e65a-bae4-4ce4-b65a-d5a587ef5f48\") " pod="openstack/novacell02ec3-account-delete-8p9cc" Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.477840 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-w6b8v"] Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.492462 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tklqq\" (UniqueName: \"kubernetes.io/projected/cf72e65a-bae4-4ce4-b65a-d5a587ef5f48-kube-api-access-tklqq\") pod \"novacell02ec3-account-delete-8p9cc\" (UID: \"cf72e65a-bae4-4ce4-b65a-d5a587ef5f48\") " pod="openstack/novacell02ec3-account-delete-8p9cc" Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.503413 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-dv6nr"] Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.540445 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-w6b8v"] Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.557113 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapi8ba6-account-delete-t8gw7"] Dec 05 07:23:23 crc kubenswrapper[4997]: I1205 07:23:23.558522 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi8ba6-account-delete-t8gw7" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.608897 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi8ba6-account-delete-t8gw7"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.625589 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell02ec3-account-delete-8p9cc" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.632573 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e3250e-2f69-4047-8ea9-cce3b817cdb8-operator-scripts\") pod \"novacell1418a-account-delete-76wq7\" (UID: \"77e3250e-2f69-4047-8ea9-cce3b817cdb8\") " pod="openstack/novacell1418a-account-delete-76wq7" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.632633 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlspm\" (UniqueName: \"kubernetes.io/projected/a5edac3e-6bee-494f-97ba-c4047064d02f-kube-api-access-rlspm\") pod \"novaapi8ba6-account-delete-t8gw7\" (UID: \"a5edac3e-6bee-494f-97ba-c4047064d02f\") " pod="openstack/novaapi8ba6-account-delete-t8gw7" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.633008 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5edac3e-6bee-494f-97ba-c4047064d02f-operator-scripts\") pod \"novaapi8ba6-account-delete-t8gw7\" (UID: \"a5edac3e-6bee-494f-97ba-c4047064d02f\") " pod="openstack/novaapi8ba6-account-delete-t8gw7" Dec 05 07:23:24 crc kubenswrapper[4997]: E1205 07:23:23.633186 4997 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Dec 05 07:23:24 crc kubenswrapper[4997]: E1205 07:23:23.633251 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/77e3250e-2f69-4047-8ea9-cce3b817cdb8-operator-scripts podName:77e3250e-2f69-4047-8ea9-cce3b817cdb8 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:24.633222986 +0000 UTC m=+1705.162130247 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/77e3250e-2f69-4047-8ea9-cce3b817cdb8-operator-scripts") pod "novacell1418a-account-delete-76wq7" (UID: "77e3250e-2f69-4047-8ea9-cce3b817cdb8") : configmap "openstack-cell1-scripts" not found Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.734970 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5edac3e-6bee-494f-97ba-c4047064d02f-operator-scripts\") pod \"novaapi8ba6-account-delete-t8gw7\" (UID: \"a5edac3e-6bee-494f-97ba-c4047064d02f\") " pod="openstack/novaapi8ba6-account-delete-t8gw7" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.735029 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz6hx\" (UniqueName: \"kubernetes.io/projected/77e3250e-2f69-4047-8ea9-cce3b817cdb8-kube-api-access-qz6hx\") pod \"novacell1418a-account-delete-76wq7\" (UID: \"77e3250e-2f69-4047-8ea9-cce3b817cdb8\") " pod="openstack/novacell1418a-account-delete-76wq7" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.735121 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlspm\" (UniqueName: \"kubernetes.io/projected/a5edac3e-6bee-494f-97ba-c4047064d02f-kube-api-access-rlspm\") pod \"novaapi8ba6-account-delete-t8gw7\" (UID: \"a5edac3e-6bee-494f-97ba-c4047064d02f\") " pod="openstack/novaapi8ba6-account-delete-t8gw7" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.736526 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5edac3e-6bee-494f-97ba-c4047064d02f-operator-scripts\") pod \"novaapi8ba6-account-delete-t8gw7\" (UID: \"a5edac3e-6bee-494f-97ba-c4047064d02f\") " pod="openstack/novaapi8ba6-account-delete-t8gw7" Dec 05 07:23:24 crc kubenswrapper[4997]: E1205 07:23:23.743280 4997 projected.go:194] Error preparing data for projected volume kube-api-access-qz6hx for pod openstack/novacell1418a-account-delete-76wq7: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 05 07:23:24 crc kubenswrapper[4997]: E1205 07:23:23.743377 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77e3250e-2f69-4047-8ea9-cce3b817cdb8-kube-api-access-qz6hx podName:77e3250e-2f69-4047-8ea9-cce3b817cdb8 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:24.743350492 +0000 UTC m=+1705.272257753 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-qz6hx" (UniqueName: "kubernetes.io/projected/77e3250e-2f69-4047-8ea9-cce3b817cdb8-kube-api-access-qz6hx") pod "novacell1418a-account-delete-76wq7" (UID: "77e3250e-2f69-4047-8ea9-cce3b817cdb8") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.785546 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlspm\" (UniqueName: \"kubernetes.io/projected/a5edac3e-6bee-494f-97ba-c4047064d02f-kube-api-access-rlspm\") pod \"novaapi8ba6-account-delete-t8gw7\" (UID: \"a5edac3e-6bee-494f-97ba-c4047064d02f\") " pod="openstack/novaapi8ba6-account-delete-t8gw7" Dec 05 07:23:24 crc kubenswrapper[4997]: E1205 07:23:23.840115 4997 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 05 07:23:24 crc kubenswrapper[4997]: E1205 07:23:23.840210 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-config-data podName:0f158874-152b-46ce-ac42-d202684853ca nodeName:}" failed. No retries permitted until 2025-12-05 07:23:25.840182367 +0000 UTC m=+1706.369089618 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-config-data") pod "rabbitmq-cell1-server-0" (UID: "0f158874-152b-46ce-ac42-d202684853ca") : configmap "rabbitmq-cell1-config-data" not found Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.907387 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7c6db4c3-5869-4a82-900c-5718abfe3a3b/ovsdbserver-sb/0.log" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.907856 4997 generic.go:334] "Generic (PLEG): container finished" podID="7c6db4c3-5869-4a82-900c-5718abfe3a3b" containerID="5d79e06e59e6f9cd89fbe9c40174b9a6f2059177958ca28bb43f3b31b4fdf3bd" exitCode=2 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.907885 4997 generic.go:334] "Generic (PLEG): container finished" podID="7c6db4c3-5869-4a82-900c-5718abfe3a3b" containerID="b92bd2595a4bc2bb590760b577f8f2341220a48a3841768940be448c3738ff62" exitCode=143 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.957237 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2f0e39f0-ca27-4656-879e-c7a20fcd904c/ovn-northd/0.log" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.957292 4997 generic.go:334] "Generic (PLEG): container finished" podID="2f0e39f0-ca27-4656-879e-c7a20fcd904c" containerID="b7780da612c7177ca63143d557d61798aed63fe6008628f0f9c5e7fba16024f1" exitCode=2 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.957312 4997 generic.go:334] "Generic (PLEG): container finished" podID="2f0e39f0-ca27-4656-879e-c7a20fcd904c" containerID="3caac945976009c44a6d871779a90a30a1333a9614f3564c29ec3e0c616fdbe7" exitCode=143 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.997084 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-89tk2_cb2d44ce-8c93-493e-9393-569c208f076f/openstack-network-exporter/0.log" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:23.997141 4997 generic.go:334] "Generic (PLEG): container finished" podID="cb2d44ce-8c93-493e-9393-569c208f076f" containerID="c77fadb14da56ec3ffb9023a9bbf4752c6ab4d4d2977ec528e7ff29efb0be39c" exitCode=2 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.173772 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e522d4c-e094-4518-bdb5-8ad8e3eccc97" path="/var/lib/kubelet/pods/1e522d4c-e094-4518-bdb5-8ad8e3eccc97/volumes" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.174536 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cf40633-ef49-416c-afcc-88c33fa76750" path="/var/lib/kubelet/pods/3cf40633-ef49-416c-afcc-88c33fa76750/volumes" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.175294 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78c3f4c1-158a-4d14-96c2-f9b6a663fd45" path="/var/lib/kubelet/pods/78c3f4c1-158a-4d14-96c2-f9b6a663fd45/volumes" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.176318 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2325c5c-6d81-45b1-90ae-d3570c255bea" path="/var/lib/kubelet/pods/b2325c5c-6d81-45b1-90ae-d3570c255bea/volumes" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.181408 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3a4d04e-bcc8-450c-830f-154f9637d513" path="/var/lib/kubelet/pods/d3a4d04e-bcc8-450c-830f-154f9637d513/volumes" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.191351 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5" path="/var/lib/kubelet/pods/ff349aa8-88f5-4c3d-bc2a-f4a9d89ba5a5/volumes" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.193462 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.193508 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-d6ftx"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.193527 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7c6db4c3-5869-4a82-900c-5718abfe3a3b","Type":"ContainerDied","Data":"5d79e06e59e6f9cd89fbe9c40174b9a6f2059177958ca28bb43f3b31b4fdf3bd"} Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.193556 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7c6db4c3-5869-4a82-900c-5718abfe3a3b","Type":"ContainerDied","Data":"b92bd2595a4bc2bb590760b577f8f2341220a48a3841768940be448c3738ff62"} Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.193574 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-6bdvj"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.193591 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-6bdvj"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.193613 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2f0e39f0-ca27-4656-879e-c7a20fcd904c","Type":"ContainerDied","Data":"b7780da612c7177ca63143d557d61798aed63fe6008628f0f9c5e7fba16024f1"} Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.193630 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2f0e39f0-ca27-4656-879e-c7a20fcd904c","Type":"ContainerDied","Data":"3caac945976009c44a6d871779a90a30a1333a9614f3564c29ec3e0c616fdbe7"} Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.193949 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-89tk2" event={"ID":"cb2d44ce-8c93-493e-9393-569c208f076f","Type":"ContainerDied","Data":"c77fadb14da56ec3ffb9023a9bbf4752c6ab4d4d2977ec528e7ff29efb0be39c"} Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.193964 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-74f5b8d45c-jbkcz"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.193983 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance7904-account-delete-v5xm8"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.194155 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="dd426a72-68fd-46b9-b427-e193a70aad3a" containerName="openstack-network-exporter" containerID="cri-o://8058fdc67614c4fcf41bcd46098bf3584ceead6983f82e156bf569ef8aa5918f" gracePeriod=300 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.194594 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-74f5b8d45c-jbkcz" podUID="c835ac5d-bd54-46d4-aab3-3ef7682f50b7" containerName="neutron-api" containerID="cri-o://8ffd311c9d4ec727b6a0cad0412afdc5d1ea879f7894edd49a446d48d71d135f" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.195151 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" podUID="def8f706-0e1b-43ab-8d4e-2cab420b795d" containerName="dnsmasq-dns" containerID="cri-o://2ac514569d94e75e167d03cc51b57c157ef81c1e5e086d9812c749f5a5524ef7" gracePeriod=10 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.195417 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-74f5b8d45c-jbkcz" podUID="c835ac5d-bd54-46d4-aab3-3ef7682f50b7" containerName="neutron-httpd" containerID="cri-o://7f4731817ad8f583546eee8efc40ce792a1f83dad3127d8c88136943c5c893be" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.196231 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-jqnrc"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.196255 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance7904-account-delete-v5xm8"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.196268 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-jqnrc"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.196306 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.196328 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-54ff4c8496-8z64m"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.196343 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.196355 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.196390 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.196403 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.196415 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.196427 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.196440 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.196862 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance7904-account-delete-v5xm8" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.199958 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="account-server" containerID="cri-o://71f731b2d431508653bb17ac618c7642061a78120cb47e7911dfe3ee31b05713" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.200463 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-54ff4c8496-8z64m" podUID="2a733d85-ad49-41b8-a75b-842dce56e85c" containerName="placement-log" containerID="cri-o://ba7342d2d5bfb967137e1fd9c51e48f142b7847eec802846b07240462ab579d3" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.203592 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-replicator" containerID="cri-o://f483369c789ec0f4e5bb24b517217914add2901c3b684a8154f738f83e6fa37b" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.203952 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="swift-recon-cron" containerID="cri-o://4fb495cc6962ca168a43dc4d03761c144cd6fa4532a0aa11050db4ee869e4f5e" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.204050 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="rsync" containerID="cri-o://d09e7dc200cd7427ec542e14f38b16b200be9b00f5c8bcc72aa1988da898ee42" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.204127 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-expirer" containerID="cri-o://82fa232b03602f1c8aad22b6c099f8b323fcdf24fc81e0fba8dc78d929f37d29" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.204218 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-updater" containerID="cri-o://6916fdc8a79724bc4f616151198996d4533bcdfd1af49ce395a01c0843637cf3" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.204286 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-auditor" containerID="cri-o://7f5cbf566dd19f6211cf0d1dae49e71d2d0c93c0e3cac1ccb8cc390c300b5e33" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.204373 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="account-reaper" containerID="cri-o://e62736a4a4741c846c90c4aeb88e01bdf1db4768265d9eb4a765c3853650e11e" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.204434 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-server" containerID="cri-o://b87c752a856aa32bd9d97eb8af0b3d65d2a9e9d6f3d911ba2d99662c5e3deef3" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.204493 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="container-updater" containerID="cri-o://732523b307a12a50490d6b3563dce9bc973897fbfa53f0e26ffcb66918c5cef4" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.204550 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="container-auditor" containerID="cri-o://2c3fe897d3a3071e0e6d73643c8249d5b3a509b04f3d1cf2c20850f1110f9dc0" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.204601 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="container-replicator" containerID="cri-o://ade129587cb10e9f61150cba8b6492aafd796f81df59fc5a8e2e4f2b826b4574" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.204677 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="container-server" containerID="cri-o://e982dc1e734d65f84320cc1a7e6b801c7989b49043d99b850c0c838b530b2638" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.204787 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="account-auditor" containerID="cri-o://7ce96074e3871701d5fda41b91a5e0c7fc16c1a429f4da49f7b22d0a58f5bcfa" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.204841 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="account-replicator" containerID="cri-o://5adcaef44afa1157fae6d04d911a954e618faa7d796126d8888521add2da41db" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.206605 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-54ff4c8496-8z64m" podUID="2a733d85-ad49-41b8-a75b-842dce56e85c" containerName="placement-api" containerID="cri-o://08946f5deeba339b1b854496b842f982af5b2b54721ebee20c813990c70fbe66" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.238277 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-66698d9c4f-5rmpv"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.244746 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-66698d9c4f-5rmpv" podUID="27ff3957-ee08-40be-a41d-02979f192fda" containerName="proxy-httpd" containerID="cri-o://6f749a47e57e47c7006263ba4386dbf5b34d5aa77a620b464309e9c5d820056b" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.244924 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bcbe26f4-fcd9-46c7-aa3a-5038b801741d" containerName="nova-metadata-log" containerID="cri-o://db7397da7d823a2ab29a061ca9792ba46bd6e1c49ac1ae528d2649ad3365c16a" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.245002 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bcbe26f4-fcd9-46c7-aa3a-5038b801741d" containerName="nova-metadata-metadata" containerID="cri-o://96ac136074ed0bcb85702f5bfe22b13a7895276ce80e0413118cb761bfafcad0" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.245034 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-66698d9c4f-5rmpv" podUID="27ff3957-ee08-40be-a41d-02979f192fda" containerName="proxy-server" containerID="cri-o://025cfd66536602dc692bb16ce499c3e7d95d59c6864c476cec229bbace2c59d3" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.245127 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e08a39e3-f267-496d-80b3-b12a9eef14c1" containerName="glance-log" containerID="cri-o://5624c5a4633297b9a0b3880002a7fb4866e561e1bf970698aec5d0b949bbfcc6" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.245193 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e08a39e3-f267-496d-80b3-b12a9eef14c1" containerName="glance-httpd" containerID="cri-o://d09f8f75d78ec8de394fe2d3878fbdc0a819558c405bcad43068ea535ea0a421" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.245377 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5afeb230-4a35-4872-bffa-c882fe927643" containerName="nova-api-log" containerID="cri-o://e1a5fbc1a91ff3e0d6d191e2796f352f99e1dd7df658b4c7a2370f2df1211989" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.245824 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="67d9e01d-5189-4aac-8f1a-b1f09fe138b8" containerName="glance-log" containerID="cri-o://d4b6f46bb0d9eab2ada6b1349324da20be55e9bbe3ce62bb67888afe1b1a469c" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.245970 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5afeb230-4a35-4872-bffa-c882fe927643" containerName="nova-api-api" containerID="cri-o://27d910165cbca8f6eb45e1a947fe3daf0e9d27a117dff110f3fc7aeef2a06961" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.246062 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="67d9e01d-5189-4aac-8f1a-b1f09fe138b8" containerName="glance-httpd" containerID="cri-o://793a4abf63252321ba4fb0dafb87c02e599c7ee609d534abf01e11eb3eff2e86" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.276498 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts\") pod \"glance7904-account-delete-v5xm8\" (UID: \"8f2677aa-a856-47de-8417-5bc3d08d6ff1\") " pod="openstack/glance7904-account-delete-v5xm8" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.276619 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-566jt\" (UniqueName: \"kubernetes.io/projected/8f2677aa-a856-47de-8417-5bc3d08d6ff1-kube-api-access-566jt\") pod \"glance7904-account-delete-v5xm8\" (UID: \"8f2677aa-a856-47de-8417-5bc3d08d6ff1\") " pod="openstack/glance7904-account-delete-v5xm8" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.298447 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell1418a-account-delete-76wq7"] Dec 05 07:23:24 crc kubenswrapper[4997]: E1205 07:23:24.299747 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-qz6hx operator-scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/novacell1418a-account-delete-76wq7" podUID="77e3250e-2f69-4047-8ea9-cce3b817cdb8" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.329791 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-db8868644-mchk9"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.330131 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-db8868644-mchk9" podUID="13fc4b26-4375-4ea2-8eaa-2417d64b3cd1" containerName="barbican-keystone-listener-log" containerID="cri-o://32e855f8095dea6a97466b7d110e63c5065aa57883dd146dc7d5cd8bf980f1a0" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.330283 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-db8868644-mchk9" podUID="13fc4b26-4375-4ea2-8eaa-2417d64b3cd1" containerName="barbican-keystone-listener" containerID="cri-o://63372a7e68a9d93e249283016625478d1f7e92640812a3890997060ed1267282" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.334942 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-66698d9c4f-5rmpv" podUID="27ff3957-ee08-40be-a41d-02979f192fda" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.166:8080/healthcheck\": EOF" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.382131 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-418a-account-create-update-mbkgn"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.383540 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts\") pod \"glance7904-account-delete-v5xm8\" (UID: \"8f2677aa-a856-47de-8417-5bc3d08d6ff1\") " pod="openstack/glance7904-account-delete-v5xm8" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.391713 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts\") pod \"glance7904-account-delete-v5xm8\" (UID: \"8f2677aa-a856-47de-8417-5bc3d08d6ff1\") " pod="openstack/glance7904-account-delete-v5xm8" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.395382 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-566jt\" (UniqueName: \"kubernetes.io/projected/8f2677aa-a856-47de-8417-5bc3d08d6ff1-kube-api-access-566jt\") pod \"glance7904-account-delete-v5xm8\" (UID: \"8f2677aa-a856-47de-8417-5bc3d08d6ff1\") " pod="openstack/glance7904-account-delete-v5xm8" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.433466 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-566jt\" (UniqueName: \"kubernetes.io/projected/8f2677aa-a856-47de-8417-5bc3d08d6ff1-kube-api-access-566jt\") pod \"glance7904-account-delete-v5xm8\" (UID: \"8f2677aa-a856-47de-8417-5bc3d08d6ff1\") " pod="openstack/glance7904-account-delete-v5xm8" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.436398 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.436737 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="0abba971-b935-4cb0-865e-b364f8521f81" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://d4ab663388d68e77285b0ecaf1e1e3e7993510fd57825fc77051658b8edf094e" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.451820 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-555c8b7b77-44qw4"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.452171 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-555c8b7b77-44qw4" podUID="a6bfb8b2-df28-4992-8831-2a5060323ddf" containerName="barbican-worker-log" containerID="cri-o://af3bd6a46ce774d4927e32dd1f7727318d24638c24860778360780d82ef22b47" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.452330 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-555c8b7b77-44qw4" podUID="a6bfb8b2-df28-4992-8831-2a5060323ddf" containerName="barbican-worker" containerID="cri-o://b77018d13ada4b2350299017976f4a2b5e3d76cc39ecbc054f294f67cbfa4f8f" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.463522 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-bmcxb"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.497296 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5cb79f8cdd-zkk89"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.497571 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5cb79f8cdd-zkk89" podUID="2179d060-8f4c-413d-a202-e31ebc242dfc" containerName="barbican-api-log" containerID="cri-o://5c7c977eca43b417984db669a02fedc758e9ee1c5a3cc1077d207c4c29b0e85d" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.498099 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5cb79f8cdd-zkk89" podUID="2179d060-8f4c-413d-a202-e31ebc242dfc" containerName="barbican-api" containerID="cri-o://d14f08fb852081fdc4815d2ce697e3418dfd98d7534ab2afb3c35a5bfea2457a" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.510630 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-418a-account-create-update-mbkgn"] Dec 05 07:23:24 crc kubenswrapper[4997]: E1205 07:23:24.511438 4997 configmap.go:193] Couldn't get configMap openstack/ovncontroller-scripts: configmap "ovncontroller-scripts" not found Dec 05 07:23:24 crc kubenswrapper[4997]: E1205 07:23:24.511528 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-scripts podName:5e4237cc-8581-4ddb-b472-e1c6a4a41cb5 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:26.511511718 +0000 UTC m=+1707.040418979 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/configmap/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-scripts") pod "ovn-controller-vnncj" (UID: "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5") : configmap "ovncontroller-scripts" not found Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.574904 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-bmcxb"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.640181 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e3250e-2f69-4047-8ea9-cce3b817cdb8-operator-scripts\") pod \"novacell1418a-account-delete-76wq7\" (UID: \"77e3250e-2f69-4047-8ea9-cce3b817cdb8\") " pod="openstack/novacell1418a-account-delete-76wq7" Dec 05 07:23:24 crc kubenswrapper[4997]: E1205 07:23:24.640646 4997 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Dec 05 07:23:24 crc kubenswrapper[4997]: E1205 07:23:24.643810 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/77e3250e-2f69-4047-8ea9-cce3b817cdb8-operator-scripts podName:77e3250e-2f69-4047-8ea9-cce3b817cdb8 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:26.643783034 +0000 UTC m=+1707.172690295 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/77e3250e-2f69-4047-8ea9-cce3b817cdb8-operator-scripts") pod "novacell1418a-account-delete-76wq7" (UID: "77e3250e-2f69-4047-8ea9-cce3b817cdb8") : configmap "openstack-cell1-scripts" not found Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.648794 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron84af-account-delete-pgl29"] Dec 05 07:23:24 crc kubenswrapper[4997]: E1205 07:23:24.672311 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f0e39f0_ca27_4656_879e_c7a20fcd904c.slice/crio-3caac945976009c44a6d871779a90a30a1333a9614f3564c29ec3e0c616fdbe7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c6db4c3_5869_4a82_900c_5718abfe3a3b.slice/crio-b92bd2595a4bc2bb590760b577f8f2341220a48a3841768940be448c3738ff62.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c6db4c3_5869_4a82_900c_5718abfe3a3b.slice/crio-conmon-b92bd2595a4bc2bb590760b577f8f2341220a48a3841768940be448c3738ff62.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod126f5f4c_a3f5_4744_99cd_691bdd33444b.slice/crio-6ea597f614ba1df9ed5420084187fc9eec112551e2d7b84b30b72cdf64053515.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f0e39f0_ca27_4656_879e_c7a20fcd904c.slice/crio-conmon-3caac945976009c44a6d871779a90a30a1333a9614f3564c29ec3e0c616fdbe7.scope\": RecentStats: unable to find data in memory cache]" Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.675718 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.676139 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="8374cc67-e735-47f0-b310-6bebd608cece" containerName="nova-cell1-conductor-conductor" containerID="cri-o://175b0ae7934dc4fc48a3e2c7faa4ea18aa39089d2851c19371ca3e1b197fb37a" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.681245 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="dd426a72-68fd-46b9-b427-e193a70aad3a" containerName="ovsdbserver-nb" containerID="cri-o://a5b9db6868281fc770b42b2466291fa6a87195e877608ca9e09e648d60355857" gracePeriod=300 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.681435 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="0f158874-152b-46ce-ac42-d202684853ca" containerName="rabbitmq" containerID="cri-o://ad9f0fd9866254fcb3d271814a479f3ce3ebdaecf8def2eb7a15d7c6654387e1" gracePeriod=604800 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.699978 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" containerName="rabbitmq" containerID="cri-o://bd15ae01a5d3986a7e4425b35e0fa45dbfda1aea5cdf404c87ede67fab61b69c" gracePeriod=604800 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.700267 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xg2jl"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.708050 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xg2jl"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.742790 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.743155 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="c706bba7-965c-497f-ae7d-b7087d37d70e" containerName="nova-cell0-conductor-conductor" containerID="cri-o://b81d606450f4ec2753efb3ba9b01b31144b20d257dc2ed0c1a6d4b204cf562b4" gracePeriod=30 Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.752717 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz6hx\" (UniqueName: \"kubernetes.io/projected/77e3250e-2f69-4047-8ea9-cce3b817cdb8-kube-api-access-qz6hx\") pod \"novacell1418a-account-delete-76wq7\" (UID: \"77e3250e-2f69-4047-8ea9-cce3b817cdb8\") " pod="openstack/novacell1418a-account-delete-76wq7" Dec 05 07:23:24 crc kubenswrapper[4997]: E1205 07:23:24.763225 4997 projected.go:194] Error preparing data for projected volume kube-api-access-qz6hx for pod openstack/novacell1418a-account-delete-76wq7: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 05 07:23:24 crc kubenswrapper[4997]: E1205 07:23:24.763305 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77e3250e-2f69-4047-8ea9-cce3b817cdb8-kube-api-access-qz6hx podName:77e3250e-2f69-4047-8ea9-cce3b817cdb8 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:26.763285124 +0000 UTC m=+1707.292192385 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-qz6hx" (UniqueName: "kubernetes.io/projected/77e3250e-2f69-4047-8ea9-cce3b817cdb8-kube-api-access-qz6hx") pod "novacell1418a-account-delete-76wq7" (UID: "77e3250e-2f69-4047-8ea9-cce3b817cdb8") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.800792 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-qzrvc"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.854756 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-qzrvc"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.941153 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:23:24 crc kubenswrapper[4997]: I1205 07:23:24.941427 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed" containerName="nova-scheduler-scheduler" containerID="cri-o://96f6c22ff0cd4cb36b8fadc88840e34b4298c41452961e008b5dd7fc23ef4889" gracePeriod=30 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.085441 4997 generic.go:334] "Generic (PLEG): container finished" podID="c835ac5d-bd54-46d4-aab3-3ef7682f50b7" containerID="7f4731817ad8f583546eee8efc40ce792a1f83dad3127d8c88136943c5c893be" exitCode=0 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.085532 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74f5b8d45c-jbkcz" event={"ID":"c835ac5d-bd54-46d4-aab3-3ef7682f50b7","Type":"ContainerDied","Data":"7f4731817ad8f583546eee8efc40ce792a1f83dad3127d8c88136943c5c893be"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.096433 4997 generic.go:334] "Generic (PLEG): container finished" podID="13fc4b26-4375-4ea2-8eaa-2417d64b3cd1" containerID="32e855f8095dea6a97466b7d110e63c5065aa57883dd146dc7d5cd8bf980f1a0" exitCode=143 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.096587 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-db8868644-mchk9" event={"ID":"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1","Type":"ContainerDied","Data":"32e855f8095dea6a97466b7d110e63c5065aa57883dd146dc7d5cd8bf980f1a0"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.104123 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-89tk2_cb2d44ce-8c93-493e-9393-569c208f076f/openstack-network-exporter/0.log" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.104237 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-89tk2" event={"ID":"cb2d44ce-8c93-493e-9393-569c208f076f","Type":"ContainerDied","Data":"ffc8b8171d24cc0361e4ef7a4a54bb55664d374163a75fc1a7e0ed31f219a224"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.104289 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffc8b8171d24cc0361e4ef7a4a54bb55664d374163a75fc1a7e0ed31f219a224" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.118347 4997 generic.go:334] "Generic (PLEG): container finished" podID="e08a39e3-f267-496d-80b3-b12a9eef14c1" containerID="5624c5a4633297b9a0b3880002a7fb4866e561e1bf970698aec5d0b949bbfcc6" exitCode=143 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.118465 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e08a39e3-f267-496d-80b3-b12a9eef14c1","Type":"ContainerDied","Data":"5624c5a4633297b9a0b3880002a7fb4866e561e1bf970698aec5d0b949bbfcc6"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.147595 4997 generic.go:334] "Generic (PLEG): container finished" podID="27ff3957-ee08-40be-a41d-02979f192fda" containerID="6f749a47e57e47c7006263ba4386dbf5b34d5aa77a620b464309e9c5d820056b" exitCode=0 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.147715 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-66698d9c4f-5rmpv" event={"ID":"27ff3957-ee08-40be-a41d-02979f192fda","Type":"ContainerDied","Data":"6f749a47e57e47c7006263ba4386dbf5b34d5aa77a620b464309e9c5d820056b"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.149962 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" containerName="galera" containerID="cri-o://5f394487f180524c3058b7242cb6bd0e8d1659b04d10a82da617c25ca61bffc0" gracePeriod=30 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.152109 4997 generic.go:334] "Generic (PLEG): container finished" podID="126f5f4c-a3f5-4744-99cd-691bdd33444b" containerID="6ea597f614ba1df9ed5420084187fc9eec112551e2d7b84b30b72cdf64053515" exitCode=137 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.159876 4997 generic.go:334] "Generic (PLEG): container finished" podID="bcbe26f4-fcd9-46c7-aa3a-5038b801741d" containerID="db7397da7d823a2ab29a061ca9792ba46bd6e1c49ac1ae528d2649ad3365c16a" exitCode=143 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.159956 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bcbe26f4-fcd9-46c7-aa3a-5038b801741d","Type":"ContainerDied","Data":"db7397da7d823a2ab29a061ca9792ba46bd6e1c49ac1ae528d2649ad3365c16a"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.168210 4997 generic.go:334] "Generic (PLEG): container finished" podID="def8f706-0e1b-43ab-8d4e-2cab420b795d" containerID="2ac514569d94e75e167d03cc51b57c157ef81c1e5e086d9812c749f5a5524ef7" exitCode=0 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.168288 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" event={"ID":"def8f706-0e1b-43ab-8d4e-2cab420b795d","Type":"ContainerDied","Data":"2ac514569d94e75e167d03cc51b57c157ef81c1e5e086d9812c749f5a5524ef7"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.373744 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-sctbp" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovs-vswitchd" containerID="cri-o://68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" gracePeriod=28 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374173 4997 generic.go:334] "Generic (PLEG): container finished" podID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerID="d09e7dc200cd7427ec542e14f38b16b200be9b00f5c8bcc72aa1988da898ee42" exitCode=0 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374204 4997 generic.go:334] "Generic (PLEG): container finished" podID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerID="82fa232b03602f1c8aad22b6c099f8b323fcdf24fc81e0fba8dc78d929f37d29" exitCode=0 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374212 4997 generic.go:334] "Generic (PLEG): container finished" podID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerID="6916fdc8a79724bc4f616151198996d4533bcdfd1af49ce395a01c0843637cf3" exitCode=0 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374219 4997 generic.go:334] "Generic (PLEG): container finished" podID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerID="7f5cbf566dd19f6211cf0d1dae49e71d2d0c93c0e3cac1ccb8cc390c300b5e33" exitCode=0 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374229 4997 generic.go:334] "Generic (PLEG): container finished" podID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerID="f483369c789ec0f4e5bb24b517217914add2901c3b684a8154f738f83e6fa37b" exitCode=0 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374237 4997 generic.go:334] "Generic (PLEG): container finished" podID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerID="732523b307a12a50490d6b3563dce9bc973897fbfa53f0e26ffcb66918c5cef4" exitCode=0 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374246 4997 generic.go:334] "Generic (PLEG): container finished" podID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerID="2c3fe897d3a3071e0e6d73643c8249d5b3a509b04f3d1cf2c20850f1110f9dc0" exitCode=0 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374257 4997 generic.go:334] "Generic (PLEG): container finished" podID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerID="ade129587cb10e9f61150cba8b6492aafd796f81df59fc5a8e2e4f2b826b4574" exitCode=0 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374265 4997 generic.go:334] "Generic (PLEG): container finished" podID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerID="e62736a4a4741c846c90c4aeb88e01bdf1db4768265d9eb4a765c3853650e11e" exitCode=0 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374273 4997 generic.go:334] "Generic (PLEG): container finished" podID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerID="7ce96074e3871701d5fda41b91a5e0c7fc16c1a429f4da49f7b22d0a58f5bcfa" exitCode=0 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374281 4997 generic.go:334] "Generic (PLEG): container finished" podID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerID="5adcaef44afa1157fae6d04d911a954e618faa7d796126d8888521add2da41db" exitCode=0 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374337 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerDied","Data":"d09e7dc200cd7427ec542e14f38b16b200be9b00f5c8bcc72aa1988da898ee42"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374369 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerDied","Data":"82fa232b03602f1c8aad22b6c099f8b323fcdf24fc81e0fba8dc78d929f37d29"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374380 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerDied","Data":"6916fdc8a79724bc4f616151198996d4533bcdfd1af49ce395a01c0843637cf3"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374391 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerDied","Data":"7f5cbf566dd19f6211cf0d1dae49e71d2d0c93c0e3cac1ccb8cc390c300b5e33"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374402 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerDied","Data":"f483369c789ec0f4e5bb24b517217914add2901c3b684a8154f738f83e6fa37b"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374411 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerDied","Data":"732523b307a12a50490d6b3563dce9bc973897fbfa53f0e26ffcb66918c5cef4"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374424 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerDied","Data":"2c3fe897d3a3071e0e6d73643c8249d5b3a509b04f3d1cf2c20850f1110f9dc0"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374433 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerDied","Data":"ade129587cb10e9f61150cba8b6492aafd796f81df59fc5a8e2e4f2b826b4574"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374442 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerDied","Data":"e62736a4a4741c846c90c4aeb88e01bdf1db4768265d9eb4a765c3853650e11e"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374451 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerDied","Data":"7ce96074e3871701d5fda41b91a5e0c7fc16c1a429f4da49f7b22d0a58f5bcfa"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.374459 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerDied","Data":"5adcaef44afa1157fae6d04d911a954e618faa7d796126d8888521add2da41db"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.383864 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2f0e39f0-ca27-4656-879e-c7a20fcd904c/ovn-northd/0.log" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.383967 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2f0e39f0-ca27-4656-879e-c7a20fcd904c","Type":"ContainerDied","Data":"ec9ff4ae268b09d7939f787bf96284c22711b8f29e8e136858254ff329c18a43"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.384007 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec9ff4ae268b09d7939f787bf96284c22711b8f29e8e136858254ff329c18a43" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.387684 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron84af-account-delete-pgl29" event={"ID":"68490a68-dc32-448f-90cc-c146ba49c1a8","Type":"ContainerStarted","Data":"fb2b75957eaf704e76b01bf7cb86971ae443f6f4cb671f899c836d9e2f573577"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.419699 4997 generic.go:334] "Generic (PLEG): container finished" podID="2179d060-8f4c-413d-a202-e31ebc242dfc" containerID="5c7c977eca43b417984db669a02fedc758e9ee1c5a3cc1077d207c4c29b0e85d" exitCode=143 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.419786 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cb79f8cdd-zkk89" event={"ID":"2179d060-8f4c-413d-a202-e31ebc242dfc","Type":"ContainerDied","Data":"5c7c977eca43b417984db669a02fedc758e9ee1c5a3cc1077d207c4c29b0e85d"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.430753 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi8ba6-account-delete-t8gw7" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.435845 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7c6db4c3-5869-4a82-900c-5718abfe3a3b/ovsdbserver-sb/0.log" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.435927 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7c6db4c3-5869-4a82-900c-5718abfe3a3b","Type":"ContainerDied","Data":"cec8912e079eb100a0bcc221020ceff6158597b2969316c396c5a9aea58e1df6"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.435965 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cec8912e079eb100a0bcc221020ceff6158597b2969316c396c5a9aea58e1df6" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.455581 4997 generic.go:334] "Generic (PLEG): container finished" podID="a6bfb8b2-df28-4992-8831-2a5060323ddf" containerID="af3bd6a46ce774d4927e32dd1f7727318d24638c24860778360780d82ef22b47" exitCode=143 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.455993 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-555c8b7b77-44qw4" event={"ID":"a6bfb8b2-df28-4992-8831-2a5060323ddf","Type":"ContainerDied","Data":"af3bd6a46ce774d4927e32dd1f7727318d24638c24860778360780d82ef22b47"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.467184 4997 generic.go:334] "Generic (PLEG): container finished" podID="67d9e01d-5189-4aac-8f1a-b1f09fe138b8" containerID="d4b6f46bb0d9eab2ada6b1349324da20be55e9bbe3ce62bb67888afe1b1a469c" exitCode=143 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.467258 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67d9e01d-5189-4aac-8f1a-b1f09fe138b8","Type":"ContainerDied","Data":"d4b6f46bb0d9eab2ada6b1349324da20be55e9bbe3ce62bb67888afe1b1a469c"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.468073 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2f0e39f0-ca27-4656-879e-c7a20fcd904c/ovn-northd/0.log" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.468149 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.469808 4997 generic.go:334] "Generic (PLEG): container finished" podID="5afeb230-4a35-4872-bffa-c882fe927643" containerID="e1a5fbc1a91ff3e0d6d191e2796f352f99e1dd7df658b4c7a2370f2df1211989" exitCode=143 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.469862 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5afeb230-4a35-4872-bffa-c882fe927643","Type":"ContainerDied","Data":"e1a5fbc1a91ff3e0d6d191e2796f352f99e1dd7df658b4c7a2370f2df1211989"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.471197 4997 generic.go:334] "Generic (PLEG): container finished" podID="2a733d85-ad49-41b8-a75b-842dce56e85c" containerID="ba7342d2d5bfb967137e1fd9c51e48f142b7847eec802846b07240462ab579d3" exitCode=143 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.471243 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-54ff4c8496-8z64m" event={"ID":"2a733d85-ad49-41b8-a75b-842dce56e85c","Type":"ContainerDied","Data":"ba7342d2d5bfb967137e1fd9c51e48f142b7847eec802846b07240462ab579d3"} Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.483326 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-sctbp" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovsdb-server" containerID="cri-o://85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" gracePeriod=28 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.493423 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_dd426a72-68fd-46b9-b427-e193a70aad3a/ovsdbserver-nb/0.log" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.493483 4997 generic.go:334] "Generic (PLEG): container finished" podID="dd426a72-68fd-46b9-b427-e193a70aad3a" containerID="8058fdc67614c4fcf41bcd46098bf3584ceead6983f82e156bf569ef8aa5918f" exitCode=2 Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.493595 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell1418a-account-delete-76wq7" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.493770 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"dd426a72-68fd-46b9-b427-e193a70aad3a","Type":"ContainerDied","Data":"8058fdc67614c4fcf41bcd46098bf3584ceead6983f82e156bf569ef8aa5918f"} Dec 05 07:23:25 crc kubenswrapper[4997]: E1205 07:23:25.507962 4997 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 05 07:23:25 crc kubenswrapper[4997]: E1205 07:23:25.508061 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-config-data podName:1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:29.508041586 +0000 UTC m=+1710.036948857 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-config-data") pod "rabbitmq-server-0" (UID: "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79") : configmap "rabbitmq-config-data" not found Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.581872 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance7904-account-delete-v5xm8" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.611468 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7c6db4c3-5869-4a82-900c-5718abfe3a3b/ovsdbserver-sb/0.log" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.611803 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.611995 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f0e39f0-ca27-4656-879e-c7a20fcd904c-config\") pod \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.612088 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-metrics-certs-tls-certs\") pod \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.612158 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2f0e39f0-ca27-4656-879e-c7a20fcd904c-ovn-rundir\") pod \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.613456 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-89tk2_cb2d44ce-8c93-493e-9393-569c208f076f/openstack-network-exporter/0.log" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.613499 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.614680 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f0e39f0-ca27-4656-879e-c7a20fcd904c-config" (OuterVolumeSpecName: "config") pod "2f0e39f0-ca27-4656-879e-c7a20fcd904c" (UID: "2f0e39f0-ca27-4656-879e-c7a20fcd904c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.615720 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-combined-ca-bundle\") pod \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.615763 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f0e39f0-ca27-4656-879e-c7a20fcd904c-scripts\") pod \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.615826 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdbk9\" (UniqueName: \"kubernetes.io/projected/2f0e39f0-ca27-4656-879e-c7a20fcd904c-kube-api-access-qdbk9\") pod \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.615858 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-ovn-northd-tls-certs\") pod \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\" (UID: \"2f0e39f0-ca27-4656-879e-c7a20fcd904c\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.616039 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f0e39f0-ca27-4656-879e-c7a20fcd904c-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "2f0e39f0-ca27-4656-879e-c7a20fcd904c" (UID: "2f0e39f0-ca27-4656-879e-c7a20fcd904c"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.616815 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f0e39f0-ca27-4656-879e-c7a20fcd904c-scripts" (OuterVolumeSpecName: "scripts") pod "2f0e39f0-ca27-4656-879e-c7a20fcd904c" (UID: "2f0e39f0-ca27-4656-879e-c7a20fcd904c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.616943 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f0e39f0-ca27-4656-879e-c7a20fcd904c-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.616961 4997 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2f0e39f0-ca27-4656-879e-c7a20fcd904c-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.616972 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f0e39f0-ca27-4656-879e-c7a20fcd904c-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.632719 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f0e39f0-ca27-4656-879e-c7a20fcd904c-kube-api-access-qdbk9" (OuterVolumeSpecName: "kube-api-access-qdbk9") pod "2f0e39f0-ca27-4656-879e-c7a20fcd904c" (UID: "2f0e39f0-ca27-4656-879e-c7a20fcd904c"). InnerVolumeSpecName "kube-api-access-qdbk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.698662 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f0e39f0-ca27-4656-879e-c7a20fcd904c" (UID: "2f0e39f0-ca27-4656-879e-c7a20fcd904c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.719594 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb2d44ce-8c93-493e-9393-569c208f076f-combined-ca-bundle\") pod \"cb2d44ce-8c93-493e-9393-569c208f076f\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.719712 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-ovsdbserver-sb-tls-certs\") pod \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.719745 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c6db4c3-5869-4a82-900c-5718abfe3a3b-config\") pod \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.719912 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-metrics-certs-tls-certs\") pod \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.719975 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c6db4c3-5869-4a82-900c-5718abfe3a3b-ovsdb-rundir\") pod \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.720031 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cb2d44ce-8c93-493e-9393-569c208f076f-ovs-rundir\") pod \"cb2d44ce-8c93-493e-9393-569c208f076f\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.720063 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb2d44ce-8c93-493e-9393-569c208f076f-config\") pod \"cb2d44ce-8c93-493e-9393-569c208f076f\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.720176 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.720205 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4bkg\" (UniqueName: \"kubernetes.io/projected/7c6db4c3-5869-4a82-900c-5718abfe3a3b-kube-api-access-q4bkg\") pod \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.720232 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cb2d44ce-8c93-493e-9393-569c208f076f-ovn-rundir\") pod \"cb2d44ce-8c93-493e-9393-569c208f076f\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.720284 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-combined-ca-bundle\") pod \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.720314 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c6db4c3-5869-4a82-900c-5718abfe3a3b-scripts\") pod \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\" (UID: \"7c6db4c3-5869-4a82-900c-5718abfe3a3b\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.720387 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lf6bt\" (UniqueName: \"kubernetes.io/projected/cb2d44ce-8c93-493e-9393-569c208f076f-kube-api-access-lf6bt\") pod \"cb2d44ce-8c93-493e-9393-569c208f076f\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.720537 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb2d44ce-8c93-493e-9393-569c208f076f-metrics-certs-tls-certs\") pod \"cb2d44ce-8c93-493e-9393-569c208f076f\" (UID: \"cb2d44ce-8c93-493e-9393-569c208f076f\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.721008 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb2d44ce-8c93-493e-9393-569c208f076f-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "cb2d44ce-8c93-493e-9393-569c208f076f" (UID: "cb2d44ce-8c93-493e-9393-569c208f076f"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.721143 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c6db4c3-5869-4a82-900c-5718abfe3a3b-config" (OuterVolumeSpecName: "config") pod "7c6db4c3-5869-4a82-900c-5718abfe3a3b" (UID: "7c6db4c3-5869-4a82-900c-5718abfe3a3b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.721573 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c6db4c3-5869-4a82-900c-5718abfe3a3b-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.721595 4997 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cb2d44ce-8c93-493e-9393-569c208f076f-ovs-rundir\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.721613 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.721647 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdbk9\" (UniqueName: \"kubernetes.io/projected/2f0e39f0-ca27-4656-879e-c7a20fcd904c-kube-api-access-qdbk9\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.721732 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c6db4c3-5869-4a82-900c-5718abfe3a3b-scripts" (OuterVolumeSpecName: "scripts") pod "7c6db4c3-5869-4a82-900c-5718abfe3a3b" (UID: "7c6db4c3-5869-4a82-900c-5718abfe3a3b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.721809 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb2d44ce-8c93-493e-9393-569c208f076f-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "cb2d44ce-8c93-493e-9393-569c208f076f" (UID: "cb2d44ce-8c93-493e-9393-569c208f076f"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.724720 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb2d44ce-8c93-493e-9393-569c208f076f-config" (OuterVolumeSpecName: "config") pod "cb2d44ce-8c93-493e-9393-569c208f076f" (UID: "cb2d44ce-8c93-493e-9393-569c208f076f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.731056 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c6db4c3-5869-4a82-900c-5718abfe3a3b-kube-api-access-q4bkg" (OuterVolumeSpecName: "kube-api-access-q4bkg") pod "7c6db4c3-5869-4a82-900c-5718abfe3a3b" (UID: "7c6db4c3-5869-4a82-900c-5718abfe3a3b"). InnerVolumeSpecName "kube-api-access-q4bkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.736932 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "7c6db4c3-5869-4a82-900c-5718abfe3a3b" (UID: "7c6db4c3-5869-4a82-900c-5718abfe3a3b"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.737417 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c6db4c3-5869-4a82-900c-5718abfe3a3b-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "7c6db4c3-5869-4a82-900c-5718abfe3a3b" (UID: "7c6db4c3-5869-4a82-900c-5718abfe3a3b"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.737876 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.738566 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell1418a-account-delete-76wq7" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.753936 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb2d44ce-8c93-493e-9393-569c208f076f-kube-api-access-lf6bt" (OuterVolumeSpecName: "kube-api-access-lf6bt") pod "cb2d44ce-8c93-493e-9393-569c208f076f" (UID: "cb2d44ce-8c93-493e-9393-569c208f076f"). InnerVolumeSpecName "kube-api-access-lf6bt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.772546 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.797244 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="508d4dec-7bf0-4053-a665-b1bf87897981" path="/var/lib/kubelet/pods/508d4dec-7bf0-4053-a665-b1bf87897981/volumes" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.797748 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb2d44ce-8c93-493e-9393-569c208f076f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb2d44ce-8c93-493e-9393-569c208f076f" (UID: "cb2d44ce-8c93-493e-9393-569c208f076f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.797974 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="778a1f81-32bc-4c34-b77d-3675d24d5467" path="/var/lib/kubelet/pods/778a1f81-32bc-4c34-b77d-3675d24d5467/volumes" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.798571 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdf7e311-569b-43a2-bba8-729226e28415" path="/var/lib/kubelet/pods/cdf7e311-569b-43a2-bba8-729226e28415/volumes" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.803723 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0d26c9f-3fc4-4626-9212-abac9cd65264" path="/var/lib/kubelet/pods/d0d26c9f-3fc4-4626-9212-abac9cd65264/volumes" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.804369 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfe50d2f-e9dc-4222-a00d-f879885600bd" path="/var/lib/kubelet/pods/dfe50d2f-e9dc-4222-a00d-f879885600bd/volumes" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.805070 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6066f8c-fc22-4592-b88e-603063a5a06a" path="/var/lib/kubelet/pods/e6066f8c-fc22-4592-b88e-603063a5a06a/volumes" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.816829 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "2f0e39f0-ca27-4656-879e-c7a20fcd904c" (UID: "2f0e39f0-ca27-4656-879e-c7a20fcd904c"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.824334 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lf6bt\" (UniqueName: \"kubernetes.io/projected/cb2d44ce-8c93-493e-9393-569c208f076f-kube-api-access-lf6bt\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.824363 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb2d44ce-8c93-493e-9393-569c208f076f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.824373 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c6db4c3-5869-4a82-900c-5718abfe3a3b-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.824383 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb2d44ce-8c93-493e-9393-569c208f076f-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.824393 4997 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.824401 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4bkg\" (UniqueName: \"kubernetes.io/projected/7c6db4c3-5869-4a82-900c-5718abfe3a3b-kube-api-access-q4bkg\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.824425 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.824438 4997 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cb2d44ce-8c93-493e-9393-569c208f076f-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.824447 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c6db4c3-5869-4a82-900c-5718abfe3a3b-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.829995 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "2f0e39f0-ca27-4656-879e-c7a20fcd904c" (UID: "2f0e39f0-ca27-4656-879e-c7a20fcd904c"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.857841 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c6db4c3-5869-4a82-900c-5718abfe3a3b" (UID: "7c6db4c3-5869-4a82-900c-5718abfe3a3b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.935725 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-dns-svc\") pod \"def8f706-0e1b-43ab-8d4e-2cab420b795d\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.935798 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/126f5f4c-a3f5-4744-99cd-691bdd33444b-openstack-config\") pod \"126f5f4c-a3f5-4744-99cd-691bdd33444b\" (UID: \"126f5f4c-a3f5-4744-99cd-691bdd33444b\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.935883 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkgmc\" (UniqueName: \"kubernetes.io/projected/def8f706-0e1b-43ab-8d4e-2cab420b795d-kube-api-access-qkgmc\") pod \"def8f706-0e1b-43ab-8d4e-2cab420b795d\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.935968 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/126f5f4c-a3f5-4744-99cd-691bdd33444b-combined-ca-bundle\") pod \"126f5f4c-a3f5-4744-99cd-691bdd33444b\" (UID: \"126f5f4c-a3f5-4744-99cd-691bdd33444b\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.936010 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-config\") pod \"def8f706-0e1b-43ab-8d4e-2cab420b795d\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.936041 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-ovsdbserver-nb\") pod \"def8f706-0e1b-43ab-8d4e-2cab420b795d\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.936192 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-dns-swift-storage-0\") pod \"def8f706-0e1b-43ab-8d4e-2cab420b795d\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.936240 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-ovsdbserver-sb\") pod \"def8f706-0e1b-43ab-8d4e-2cab420b795d\" (UID: \"def8f706-0e1b-43ab-8d4e-2cab420b795d\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.936287 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvzz5\" (UniqueName: \"kubernetes.io/projected/126f5f4c-a3f5-4744-99cd-691bdd33444b-kube-api-access-bvzz5\") pod \"126f5f4c-a3f5-4744-99cd-691bdd33444b\" (UID: \"126f5f4c-a3f5-4744-99cd-691bdd33444b\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.936397 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/126f5f4c-a3f5-4744-99cd-691bdd33444b-openstack-config-secret\") pod \"126f5f4c-a3f5-4744-99cd-691bdd33444b\" (UID: \"126f5f4c-a3f5-4744-99cd-691bdd33444b\") " Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.937050 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.937086 4997 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f0e39f0-ca27-4656-879e-c7a20fcd904c-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:25 crc kubenswrapper[4997]: E1205 07:23:25.937230 4997 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 05 07:23:25 crc kubenswrapper[4997]: E1205 07:23:25.937314 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-config-data podName:0f158874-152b-46ce-ac42-d202684853ca nodeName:}" failed. No retries permitted until 2025-12-05 07:23:29.937286694 +0000 UTC m=+1710.466193955 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-config-data") pod "rabbitmq-cell1-server-0" (UID: "0f158874-152b-46ce-ac42-d202684853ca") : configmap "rabbitmq-cell1-config-data" not found Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.953771 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/126f5f4c-a3f5-4744-99cd-691bdd33444b-kube-api-access-bvzz5" (OuterVolumeSpecName: "kube-api-access-bvzz5") pod "126f5f4c-a3f5-4744-99cd-691bdd33444b" (UID: "126f5f4c-a3f5-4744-99cd-691bdd33444b"). InnerVolumeSpecName "kube-api-access-bvzz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.961595 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/def8f706-0e1b-43ab-8d4e-2cab420b795d-kube-api-access-qkgmc" (OuterVolumeSpecName: "kube-api-access-qkgmc") pod "def8f706-0e1b-43ab-8d4e-2cab420b795d" (UID: "def8f706-0e1b-43ab-8d4e-2cab420b795d"). InnerVolumeSpecName "kube-api-access-qkgmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:25 crc kubenswrapper[4997]: I1205 07:23:25.978737 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "7c6db4c3-5869-4a82-900c-5718abfe3a3b" (UID: "7c6db4c3-5869-4a82-900c-5718abfe3a3b"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.004607 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "7c6db4c3-5869-4a82-900c-5718abfe3a3b" (UID: "7c6db4c3-5869-4a82-900c-5718abfe3a3b"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.005029 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb2d44ce-8c93-493e-9393-569c208f076f-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "cb2d44ce-8c93-493e-9393-569c208f076f" (UID: "cb2d44ce-8c93-493e-9393-569c208f076f"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.019371 4997 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.062814 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="0d3915c5-30d2-43be-abd4-438cd9d8ebf7" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.174:8776/healthcheck\": read tcp 10.217.0.2:36684->10.217.0.174:8776: read: connection reset by peer" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.069327 4997 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.069361 4997 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb2d44ce-8c93-493e-9393-569c208f076f-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.069381 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvzz5\" (UniqueName: \"kubernetes.io/projected/126f5f4c-a3f5-4744-99cd-691bdd33444b-kube-api-access-bvzz5\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.069395 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.069406 4997 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c6db4c3-5869-4a82-900c-5718abfe3a3b-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.069416 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkgmc\" (UniqueName: \"kubernetes.io/projected/def8f706-0e1b-43ab-8d4e-2cab420b795d-kube-api-access-qkgmc\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.069901 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_dd426a72-68fd-46b9-b427-e193a70aad3a/ovsdbserver-nb/0.log" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.070018 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.089843 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican69df-account-delete-tqlsz"] Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.142452 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "def8f706-0e1b-43ab-8d4e-2cab420b795d" (UID: "def8f706-0e1b-43ab-8d4e-2cab420b795d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.149423 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell02ec3-account-delete-8p9cc"] Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.171074 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-ovsdbserver-nb-tls-certs\") pod \"dd426a72-68fd-46b9-b427-e193a70aad3a\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.171151 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lrqm\" (UniqueName: \"kubernetes.io/projected/dd426a72-68fd-46b9-b427-e193a70aad3a-kube-api-access-5lrqm\") pod \"dd426a72-68fd-46b9-b427-e193a70aad3a\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.171180 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dd426a72-68fd-46b9-b427-e193a70aad3a-ovsdb-rundir\") pod \"dd426a72-68fd-46b9-b427-e193a70aad3a\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.171260 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd426a72-68fd-46b9-b427-e193a70aad3a-scripts\") pod \"dd426a72-68fd-46b9-b427-e193a70aad3a\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.171320 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-combined-ca-bundle\") pod \"dd426a72-68fd-46b9-b427-e193a70aad3a\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.171355 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"dd426a72-68fd-46b9-b427-e193a70aad3a\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.171507 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-metrics-certs-tls-certs\") pod \"dd426a72-68fd-46b9-b427-e193a70aad3a\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.171628 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd426a72-68fd-46b9-b427-e193a70aad3a-config\") pod \"dd426a72-68fd-46b9-b427-e193a70aad3a\" (UID: \"dd426a72-68fd-46b9-b427-e193a70aad3a\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.172202 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.179199 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd426a72-68fd-46b9-b427-e193a70aad3a-scripts" (OuterVolumeSpecName: "scripts") pod "dd426a72-68fd-46b9-b427-e193a70aad3a" (UID: "dd426a72-68fd-46b9-b427-e193a70aad3a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.179793 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd426a72-68fd-46b9-b427-e193a70aad3a-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "dd426a72-68fd-46b9-b427-e193a70aad3a" (UID: "dd426a72-68fd-46b9-b427-e193a70aad3a"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.179870 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/126f5f4c-a3f5-4744-99cd-691bdd33444b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "126f5f4c-a3f5-4744-99cd-691bdd33444b" (UID: "126f5f4c-a3f5-4744-99cd-691bdd33444b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.189071 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement94df-account-delete-4qxc6"] Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.190393 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd426a72-68fd-46b9-b427-e193a70aad3a-config" (OuterVolumeSpecName: "config") pod "dd426a72-68fd-46b9-b427-e193a70aad3a" (UID: "dd426a72-68fd-46b9-b427-e193a70aad3a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.205338 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd426a72-68fd-46b9-b427-e193a70aad3a-kube-api-access-5lrqm" (OuterVolumeSpecName: "kube-api-access-5lrqm") pod "dd426a72-68fd-46b9-b427-e193a70aad3a" (UID: "dd426a72-68fd-46b9-b427-e193a70aad3a"). InnerVolumeSpecName "kube-api-access-5lrqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.205499 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "dd426a72-68fd-46b9-b427-e193a70aad3a" (UID: "dd426a72-68fd-46b9-b427-e193a70aad3a"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.205731 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-config" (OuterVolumeSpecName: "config") pod "def8f706-0e1b-43ab-8d4e-2cab420b795d" (UID: "def8f706-0e1b-43ab-8d4e-2cab420b795d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: E1205 07:23:26.210995 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96f6c22ff0cd4cb36b8fadc88840e34b4298c41452961e008b5dd7fc23ef4889" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 07:23:26 crc kubenswrapper[4997]: E1205 07:23:26.216426 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96f6c22ff0cd4cb36b8fadc88840e34b4298c41452961e008b5dd7fc23ef4889" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.218385 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cindera458-account-delete-vpgnx"] Dec 05 07:23:26 crc kubenswrapper[4997]: E1205 07:23:26.220694 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96f6c22ff0cd4cb36b8fadc88840e34b4298c41452961e008b5dd7fc23ef4889" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 07:23:26 crc kubenswrapper[4997]: E1205 07:23:26.220805 4997 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed" containerName="nova-scheduler-scheduler" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.222172 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/126f5f4c-a3f5-4744-99cd-691bdd33444b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "126f5f4c-a3f5-4744-99cd-691bdd33444b" (UID: "126f5f4c-a3f5-4744-99cd-691bdd33444b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.276331 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lrqm\" (UniqueName: \"kubernetes.io/projected/dd426a72-68fd-46b9-b427-e193a70aad3a-kube-api-access-5lrqm\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.276362 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dd426a72-68fd-46b9-b427-e193a70aad3a-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.276397 4997 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/126f5f4c-a3f5-4744-99cd-691bdd33444b-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.276412 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd426a72-68fd-46b9-b427-e193a70aad3a-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.276448 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.276488 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/126f5f4c-a3f5-4744-99cd-691bdd33444b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.276505 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.276515 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd426a72-68fd-46b9-b427-e193a70aad3a-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.311072 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/126f5f4c-a3f5-4744-99cd-691bdd33444b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "126f5f4c-a3f5-4744-99cd-691bdd33444b" (UID: "126f5f4c-a3f5-4744-99cd-691bdd33444b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.322980 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "def8f706-0e1b-43ab-8d4e-2cab420b795d" (UID: "def8f706-0e1b-43ab-8d4e-2cab420b795d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.342758 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd426a72-68fd-46b9-b427-e193a70aad3a" (UID: "dd426a72-68fd-46b9-b427-e193a70aad3a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.349236 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "def8f706-0e1b-43ab-8d4e-2cab420b795d" (UID: "def8f706-0e1b-43ab-8d4e-2cab420b795d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.374063 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "def8f706-0e1b-43ab-8d4e-2cab420b795d" (UID: "def8f706-0e1b-43ab-8d4e-2cab420b795d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.378701 4997 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.378744 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.378754 4997 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/126f5f4c-a3f5-4744-99cd-691bdd33444b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.378765 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.378775 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/def8f706-0e1b-43ab-8d4e-2cab420b795d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.393676 4997 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.484292 4997 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.492858 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "dd426a72-68fd-46b9-b427-e193a70aad3a" (UID: "dd426a72-68fd-46b9-b427-e193a70aad3a"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.520078 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "dd426a72-68fd-46b9-b427-e193a70aad3a" (UID: "dd426a72-68fd-46b9-b427-e193a70aad3a"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.529272 4997 generic.go:334] "Generic (PLEG): container finished" podID="68490a68-dc32-448f-90cc-c146ba49c1a8" containerID="769b1779b305cc0d585e4dba78fb1ee7190f831ce52ad7c2bc764a5532dc31e6" exitCode=0 Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.529371 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron84af-account-delete-pgl29" event={"ID":"68490a68-dc32-448f-90cc-c146ba49c1a8","Type":"ContainerDied","Data":"769b1779b305cc0d585e4dba78fb1ee7190f831ce52ad7c2bc764a5532dc31e6"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.541459 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell02ec3-account-delete-8p9cc" event={"ID":"cf72e65a-bae4-4ce4-b65a-d5a587ef5f48","Type":"ContainerStarted","Data":"7deda0533d064e3d0aa9a66967efbd9e7313bf31bbe6f07556e6ed3b65bf764e"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.562966 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_dd426a72-68fd-46b9-b427-e193a70aad3a/ovsdbserver-nb/0.log" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.563029 4997 generic.go:334] "Generic (PLEG): container finished" podID="dd426a72-68fd-46b9-b427-e193a70aad3a" containerID="a5b9db6868281fc770b42b2466291fa6a87195e877608ca9e09e648d60355857" exitCode=143 Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.563234 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.563343 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"dd426a72-68fd-46b9-b427-e193a70aad3a","Type":"ContainerDied","Data":"a5b9db6868281fc770b42b2466291fa6a87195e877608ca9e09e648d60355857"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.563404 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"dd426a72-68fd-46b9-b427-e193a70aad3a","Type":"ContainerDied","Data":"eef2665cee84fa9c47233a9b8e042e05cba63a5f3ba4909bf0df0fd50eb9127e"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.563425 4997 scope.go:117] "RemoveContainer" containerID="8058fdc67614c4fcf41bcd46098bf3584ceead6983f82e156bf569ef8aa5918f" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.574236 4997 generic.go:334] "Generic (PLEG): container finished" podID="fe32fb42-61e6-4cd4-a75b-b684a4171224" containerID="921875cb2f26d740baa583b8c4418587dfe774e4ea61c9ad78bb1723d653c892" exitCode=0 Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.574280 4997 generic.go:334] "Generic (PLEG): container finished" podID="fe32fb42-61e6-4cd4-a75b-b684a4171224" containerID="b762656e8963968dd62e28b778f1374e72ed4bfb67fbcee7edeacd3920071001" exitCode=0 Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.574414 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fe32fb42-61e6-4cd4-a75b-b684a4171224","Type":"ContainerDied","Data":"921875cb2f26d740baa583b8c4418587dfe774e4ea61c9ad78bb1723d653c892"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.574467 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fe32fb42-61e6-4cd4-a75b-b684a4171224","Type":"ContainerDied","Data":"b762656e8963968dd62e28b778f1374e72ed4bfb67fbcee7edeacd3920071001"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.584230 4997 generic.go:334] "Generic (PLEG): container finished" podID="7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" containerID="5f394487f180524c3058b7242cb6bd0e8d1659b04d10a82da617c25ca61bffc0" exitCode=0 Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.584359 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b","Type":"ContainerDied","Data":"5f394487f180524c3058b7242cb6bd0e8d1659b04d10a82da617c25ca61bffc0"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.585548 4997 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: E1205 07:23:26.585560 4997 configmap.go:193] Couldn't get configMap openstack/ovncontroller-scripts: configmap "ovncontroller-scripts" not found Dec 05 07:23:26 crc kubenswrapper[4997]: E1205 07:23:26.585654 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-scripts podName:5e4237cc-8581-4ddb-b472-e1c6a4a41cb5 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:30.585630432 +0000 UTC m=+1711.114537843 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/configmap/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-scripts") pod "ovn-controller-vnncj" (UID: "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5") : configmap "ovncontroller-scripts" not found Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.585567 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd426a72-68fd-46b9-b427-e193a70aad3a-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.590074 4997 generic.go:334] "Generic (PLEG): container finished" podID="27ff3957-ee08-40be-a41d-02979f192fda" containerID="025cfd66536602dc692bb16ce499c3e7d95d59c6864c476cec229bbace2c59d3" exitCode=0 Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.590160 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-66698d9c4f-5rmpv" event={"ID":"27ff3957-ee08-40be-a41d-02979f192fda","Type":"ContainerDied","Data":"025cfd66536602dc692bb16ce499c3e7d95d59c6864c476cec229bbace2c59d3"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.590186 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-66698d9c4f-5rmpv" event={"ID":"27ff3957-ee08-40be-a41d-02979f192fda","Type":"ContainerDied","Data":"8dc02dafe00b2c60dc0fe31d05475bf78494b2f5bdf3443d1d12c8e10e0bcb87"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.590205 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8dc02dafe00b2c60dc0fe31d05475bf78494b2f5bdf3443d1d12c8e10e0bcb87" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.593172 4997 generic.go:334] "Generic (PLEG): container finished" podID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" exitCode=0 Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.593410 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sctbp" event={"ID":"5388a6f5-ecde-42db-9c94-a52c17e8a302","Type":"ContainerDied","Data":"85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.599338 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cindera458-account-delete-vpgnx" event={"ID":"b8c5aa81-a666-4ae6-a41c-fd31df260223","Type":"ContainerStarted","Data":"0afdb95173171611a53b5d9cfa50ee71774a0d899adb7dedf3b9e5e5b5a7657a"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.604109 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0abba971-b935-4cb0-865e-b364f8521f81","Type":"ContainerDied","Data":"d4ab663388d68e77285b0ecaf1e1e3e7993510fd57825fc77051658b8edf094e"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.603093 4997 generic.go:334] "Generic (PLEG): container finished" podID="0abba971-b935-4cb0-865e-b364f8521f81" containerID="d4ab663388d68e77285b0ecaf1e1e3e7993510fd57825fc77051658b8edf094e" exitCode=0 Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.621632 4997 generic.go:334] "Generic (PLEG): container finished" podID="8374cc67-e735-47f0-b310-6bebd608cece" containerID="175b0ae7934dc4fc48a3e2c7faa4ea18aa39089d2851c19371ca3e1b197fb37a" exitCode=0 Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.621717 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8374cc67-e735-47f0-b310-6bebd608cece","Type":"ContainerDied","Data":"175b0ae7934dc4fc48a3e2c7faa4ea18aa39089d2851c19371ca3e1b197fb37a"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.632994 4997 generic.go:334] "Generic (PLEG): container finished" podID="0d3915c5-30d2-43be-abd4-438cd9d8ebf7" containerID="8f36673252ccfc4fc6b7d58b342de4054e621d8bc6722f917b4cf09a6dd9bcaf" exitCode=0 Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.633081 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d3915c5-30d2-43be-abd4-438cd9d8ebf7","Type":"ContainerDied","Data":"8f36673252ccfc4fc6b7d58b342de4054e621d8bc6722f917b4cf09a6dd9bcaf"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.639305 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" event={"ID":"def8f706-0e1b-43ab-8d4e-2cab420b795d","Type":"ContainerDied","Data":"3d905d5b288a9f67b4075ad27d33434d8375b6a96e22194b534bb2fcbc98e05e"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.639465 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf4ff87b5-d6ftx" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.655934 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.660298 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi8ba6-account-delete-t8gw7"] Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.668002 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican69df-account-delete-tqlsz" event={"ID":"d65cd492-d4bd-4489-9d28-a6c42407101a","Type":"ContainerStarted","Data":"9d06463b05ff535e9822e6bdc25fcff443a779823bd9ed62a7c592c795eacbc9"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.688548 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e3250e-2f69-4047-8ea9-cce3b817cdb8-operator-scripts\") pod \"novacell1418a-account-delete-76wq7\" (UID: \"77e3250e-2f69-4047-8ea9-cce3b817cdb8\") " pod="openstack/novacell1418a-account-delete-76wq7" Dec 05 07:23:26 crc kubenswrapper[4997]: E1205 07:23:26.688825 4997 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Dec 05 07:23:26 crc kubenswrapper[4997]: E1205 07:23:26.688896 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/77e3250e-2f69-4047-8ea9-cce3b817cdb8-operator-scripts podName:77e3250e-2f69-4047-8ea9-cce3b817cdb8 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:30.688873771 +0000 UTC m=+1711.217781032 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/77e3250e-2f69-4047-8ea9-cce3b817cdb8-operator-scripts") pod "novacell1418a-account-delete-76wq7" (UID: "77e3250e-2f69-4047-8ea9-cce3b817cdb8") : configmap "openstack-cell1-scripts" not found Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.693899 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement94df-account-delete-4qxc6" event={"ID":"c51acefb-ed8a-480c-9ab6-0345f2ff0cd3","Type":"ContainerStarted","Data":"a1221a18388371ad38dd286f6a30f31ac2a036d9c6b7fc40448d3191ce825a9a"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.705197 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.712396 4997 scope.go:117] "RemoveContainer" containerID="a5b9db6868281fc770b42b2466291fa6a87195e877608ca9e09e648d60355857" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.714338 4997 generic.go:334] "Generic (PLEG): container finished" podID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerID="b87c752a856aa32bd9d97eb8af0b3d65d2a9e9d6f3d911ba2d99662c5e3deef3" exitCode=0 Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.714365 4997 generic.go:334] "Generic (PLEG): container finished" podID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerID="e982dc1e734d65f84320cc1a7e6b801c7989b49043d99b850c0c838b530b2638" exitCode=0 Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.714375 4997 generic.go:334] "Generic (PLEG): container finished" podID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerID="71f731b2d431508653bb17ac618c7642061a78120cb47e7911dfe3ee31b05713" exitCode=0 Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.714452 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-89tk2" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.715719 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.715751 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerDied","Data":"b87c752a856aa32bd9d97eb8af0b3d65d2a9e9d6f3d911ba2d99662c5e3deef3"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.715842 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerDied","Data":"e982dc1e734d65f84320cc1a7e6b801c7989b49043d99b850c0c838b530b2638"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.715865 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerDied","Data":"71f731b2d431508653bb17ac618c7642061a78120cb47e7911dfe3ee31b05713"} Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.715966 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.717920 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell1418a-account-delete-76wq7" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.778914 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.816858 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz6hx\" (UniqueName: \"kubernetes.io/projected/77e3250e-2f69-4047-8ea9-cce3b817cdb8-kube-api-access-qz6hx\") pod \"novacell1418a-account-delete-76wq7\" (UID: \"77e3250e-2f69-4047-8ea9-cce3b817cdb8\") " pod="openstack/novacell1418a-account-delete-76wq7" Dec 05 07:23:26 crc kubenswrapper[4997]: E1205 07:23:26.827461 4997 projected.go:194] Error preparing data for projected volume kube-api-access-qz6hx for pod openstack/novacell1418a-account-delete-76wq7: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 05 07:23:26 crc kubenswrapper[4997]: E1205 07:23:26.827551 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77e3250e-2f69-4047-8ea9-cce3b817cdb8-kube-api-access-qz6hx podName:77e3250e-2f69-4047-8ea9-cce3b817cdb8 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:30.82752732 +0000 UTC m=+1711.356434581 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-qz6hx" (UniqueName: "kubernetes.io/projected/77e3250e-2f69-4047-8ea9-cce3b817cdb8-kube-api-access-qz6hx") pod "novacell1418a-account-delete-76wq7" (UID: "77e3250e-2f69-4047-8ea9-cce3b817cdb8") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.844110 4997 scope.go:117] "RemoveContainer" containerID="8058fdc67614c4fcf41bcd46098bf3584ceead6983f82e156bf569ef8aa5918f" Dec 05 07:23:26 crc kubenswrapper[4997]: E1205 07:23:26.845796 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8058fdc67614c4fcf41bcd46098bf3584ceead6983f82e156bf569ef8aa5918f\": container with ID starting with 8058fdc67614c4fcf41bcd46098bf3584ceead6983f82e156bf569ef8aa5918f not found: ID does not exist" containerID="8058fdc67614c4fcf41bcd46098bf3584ceead6983f82e156bf569ef8aa5918f" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.845861 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8058fdc67614c4fcf41bcd46098bf3584ceead6983f82e156bf569ef8aa5918f"} err="failed to get container status \"8058fdc67614c4fcf41bcd46098bf3584ceead6983f82e156bf569ef8aa5918f\": rpc error: code = NotFound desc = could not find container \"8058fdc67614c4fcf41bcd46098bf3584ceead6983f82e156bf569ef8aa5918f\": container with ID starting with 8058fdc67614c4fcf41bcd46098bf3584ceead6983f82e156bf569ef8aa5918f not found: ID does not exist" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.845901 4997 scope.go:117] "RemoveContainer" containerID="a5b9db6868281fc770b42b2466291fa6a87195e877608ca9e09e648d60355857" Dec 05 07:23:26 crc kubenswrapper[4997]: E1205 07:23:26.849546 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5b9db6868281fc770b42b2466291fa6a87195e877608ca9e09e648d60355857\": container with ID starting with a5b9db6868281fc770b42b2466291fa6a87195e877608ca9e09e648d60355857 not found: ID does not exist" containerID="a5b9db6868281fc770b42b2466291fa6a87195e877608ca9e09e648d60355857" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.849622 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5b9db6868281fc770b42b2466291fa6a87195e877608ca9e09e648d60355857"} err="failed to get container status \"a5b9db6868281fc770b42b2466291fa6a87195e877608ca9e09e648d60355857\": rpc error: code = NotFound desc = could not find container \"a5b9db6868281fc770b42b2466291fa6a87195e877608ca9e09e648d60355857\": container with ID starting with a5b9db6868281fc770b42b2466291fa6a87195e877608ca9e09e648d60355857 not found: ID does not exist" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.851467 4997 scope.go:117] "RemoveContainer" containerID="2ac514569d94e75e167d03cc51b57c157ef81c1e5e086d9812c749f5a5524ef7" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.870454 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.899433 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-d6ftx"] Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.917838 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-vencrypt-tls-certs\") pod \"0abba971-b935-4cb0-865e-b364f8521f81\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.917901 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gs6jl\" (UniqueName: \"kubernetes.io/projected/0abba971-b935-4cb0-865e-b364f8521f81-kube-api-access-gs6jl\") pod \"0abba971-b935-4cb0-865e-b364f8521f81\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.917941 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-combined-ca-bundle\") pod \"0abba971-b935-4cb0-865e-b364f8521f81\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.917982 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-combined-ca-bundle\") pod \"27ff3957-ee08-40be-a41d-02979f192fda\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.918000 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-internal-tls-certs\") pod \"27ff3957-ee08-40be-a41d-02979f192fda\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.918041 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmh68\" (UniqueName: \"kubernetes.io/projected/8374cc67-e735-47f0-b310-6bebd608cece-kube-api-access-xmh68\") pod \"8374cc67-e735-47f0-b310-6bebd608cece\" (UID: \"8374cc67-e735-47f0-b310-6bebd608cece\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.918063 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8374cc67-e735-47f0-b310-6bebd608cece-combined-ca-bundle\") pod \"8374cc67-e735-47f0-b310-6bebd608cece\" (UID: \"8374cc67-e735-47f0-b310-6bebd608cece\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.918088 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27ff3957-ee08-40be-a41d-02979f192fda-run-httpd\") pod \"27ff3957-ee08-40be-a41d-02979f192fda\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.918134 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/27ff3957-ee08-40be-a41d-02979f192fda-etc-swift\") pod \"27ff3957-ee08-40be-a41d-02979f192fda\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.918153 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nlwv\" (UniqueName: \"kubernetes.io/projected/27ff3957-ee08-40be-a41d-02979f192fda-kube-api-access-2nlwv\") pod \"27ff3957-ee08-40be-a41d-02979f192fda\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.918179 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-public-tls-certs\") pod \"27ff3957-ee08-40be-a41d-02979f192fda\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.918216 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-config-data\") pod \"27ff3957-ee08-40be-a41d-02979f192fda\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.918241 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8374cc67-e735-47f0-b310-6bebd608cece-config-data\") pod \"8374cc67-e735-47f0-b310-6bebd608cece\" (UID: \"8374cc67-e735-47f0-b310-6bebd608cece\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.918267 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-nova-novncproxy-tls-certs\") pod \"0abba971-b935-4cb0-865e-b364f8521f81\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.918317 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-config-data\") pod \"0abba971-b935-4cb0-865e-b364f8521f81\" (UID: \"0abba971-b935-4cb0-865e-b364f8521f81\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.918336 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27ff3957-ee08-40be-a41d-02979f192fda-log-httpd\") pod \"27ff3957-ee08-40be-a41d-02979f192fda\" (UID: \"27ff3957-ee08-40be-a41d-02979f192fda\") " Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.919039 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27ff3957-ee08-40be-a41d-02979f192fda-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "27ff3957-ee08-40be-a41d-02979f192fda" (UID: "27ff3957-ee08-40be-a41d-02979f192fda"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.919365 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27ff3957-ee08-40be-a41d-02979f192fda-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "27ff3957-ee08-40be-a41d-02979f192fda" (UID: "27ff3957-ee08-40be-a41d-02979f192fda"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.921907 4997 scope.go:117] "RemoveContainer" containerID="10c39b4960112188aefd1d6e4212dc87504830944ea889c17f95beb408918e4c" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.957355 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-d6ftx"] Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.980607 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0abba971-b935-4cb0-865e-b364f8521f81-kube-api-access-gs6jl" (OuterVolumeSpecName: "kube-api-access-gs6jl") pod "0abba971-b935-4cb0-865e-b364f8521f81" (UID: "0abba971-b935-4cb0-865e-b364f8521f81"). InnerVolumeSpecName "kube-api-access-gs6jl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.991593 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 07:23:26 crc kubenswrapper[4997]: I1205 07:23:26.994684 4997 scope.go:117] "RemoveContainer" containerID="6ea597f614ba1df9ed5420084187fc9eec112551e2d7b84b30b72cdf64053515" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.000735 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.009416 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.029967 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27ff3957-ee08-40be-a41d-02979f192fda-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "27ff3957-ee08-40be-a41d-02979f192fda" (UID: "27ff3957-ee08-40be-a41d-02979f192fda"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.031812 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.046103 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-89tk2"] Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.049151 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-89tk2"] Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.058734 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8374cc67-e735-47f0-b310-6bebd608cece-kube-api-access-xmh68" (OuterVolumeSpecName: "kube-api-access-xmh68") pod "8374cc67-e735-47f0-b310-6bebd608cece" (UID: "8374cc67-e735-47f0-b310-6bebd608cece"). InnerVolumeSpecName "kube-api-access-xmh68". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.068530 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmh68\" (UniqueName: \"kubernetes.io/projected/8374cc67-e735-47f0-b310-6bebd608cece-kube-api-access-xmh68\") pod \"8374cc67-e735-47f0-b310-6bebd608cece\" (UID: \"8374cc67-e735-47f0-b310-6bebd608cece\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.068891 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell1418a-account-delete-76wq7"] Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.069062 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 07:23:27 crc kubenswrapper[4997]: W1205 07:23:27.069514 4997 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/8374cc67-e735-47f0-b310-6bebd608cece/volumes/kubernetes.io~projected/kube-api-access-xmh68 Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.069608 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8374cc67-e735-47f0-b310-6bebd608cece-kube-api-access-xmh68" (OuterVolumeSpecName: "kube-api-access-xmh68") pod "8374cc67-e735-47f0-b310-6bebd608cece" (UID: "8374cc67-e735-47f0-b310-6bebd608cece"). InnerVolumeSpecName "kube-api-access-xmh68". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.070458 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmh68\" (UniqueName: \"kubernetes.io/projected/8374cc67-e735-47f0-b310-6bebd608cece-kube-api-access-xmh68\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.070513 4997 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27ff3957-ee08-40be-a41d-02979f192fda-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.070525 4997 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/27ff3957-ee08-40be-a41d-02979f192fda-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.070536 4997 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27ff3957-ee08-40be-a41d-02979f192fda-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.070545 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gs6jl\" (UniqueName: \"kubernetes.io/projected/0abba971-b935-4cb0-865e-b364f8521f81-kube-api-access-gs6jl\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.071201 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27ff3957-ee08-40be-a41d-02979f192fda-kube-api-access-2nlwv" (OuterVolumeSpecName: "kube-api-access-2nlwv") pod "27ff3957-ee08-40be-a41d-02979f192fda" (UID: "27ff3957-ee08-40be-a41d-02979f192fda"). InnerVolumeSpecName "kube-api-access-2nlwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.076882 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell1418a-account-delete-76wq7"] Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.107423 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.116951 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 07:23:27 crc kubenswrapper[4997]: W1205 07:23:27.143802 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f2677aa_a856_47de_8417_5bc3d08d6ff1.slice/crio-705ac0ba630b3336a44765fcb97172ba22ee01af0565e74f3793afff5a05fc9a WatchSource:0}: Error finding container 705ac0ba630b3336a44765fcb97172ba22ee01af0565e74f3793afff5a05fc9a: Status 404 returned error can't find the container with id 705ac0ba630b3336a44765fcb97172ba22ee01af0565e74f3793afff5a05fc9a Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.164055 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.166144 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.175332 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-config-data\") pod \"fe32fb42-61e6-4cd4-a75b-b684a4171224\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.175377 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe32fb42-61e6-4cd4-a75b-b684a4171224-etc-machine-id\") pod \"fe32fb42-61e6-4cd4-a75b-b684a4171224\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.175455 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-combined-ca-bundle\") pod \"fe32fb42-61e6-4cd4-a75b-b684a4171224\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.175510 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-config-data-custom\") pod \"fe32fb42-61e6-4cd4-a75b-b684a4171224\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.175835 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vblvp\" (UniqueName: \"kubernetes.io/projected/fe32fb42-61e6-4cd4-a75b-b684a4171224-kube-api-access-vblvp\") pod \"fe32fb42-61e6-4cd4-a75b-b684a4171224\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.176093 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-scripts\") pod \"fe32fb42-61e6-4cd4-a75b-b684a4171224\" (UID: \"fe32fb42-61e6-4cd4-a75b-b684a4171224\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.176858 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e3250e-2f69-4047-8ea9-cce3b817cdb8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.176879 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nlwv\" (UniqueName: \"kubernetes.io/projected/27ff3957-ee08-40be-a41d-02979f192fda-kube-api-access-2nlwv\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.176892 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qz6hx\" (UniqueName: \"kubernetes.io/projected/77e3250e-2f69-4047-8ea9-cce3b817cdb8-kube-api-access-qz6hx\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.188793 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fe32fb42-61e6-4cd4-a75b-b684a4171224-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "fe32fb42-61e6-4cd4-a75b-b684a4171224" (UID: "fe32fb42-61e6-4cd4-a75b-b684a4171224"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.193519 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance7904-account-delete-v5xm8"] Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.278532 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-scripts\") pod \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.278589 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-combined-ca-bundle\") pod \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.278651 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-config-data\") pod \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.278719 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-operator-scripts\") pod \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.278751 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffwz9\" (UniqueName: \"kubernetes.io/projected/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-kube-api-access-ffwz9\") pod \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.278794 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-combined-ca-bundle\") pod \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.278810 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-internal-tls-certs\") pod \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.278861 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-logs\") pod \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.278926 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-config-data-default\") pod \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.278943 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mn64\" (UniqueName: \"kubernetes.io/projected/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-kube-api-access-6mn64\") pod \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.278989 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-kolla-config\") pod \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.279025 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-config-data-custom\") pod \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.279086 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-etc-machine-id\") pod \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.279120 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.279285 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-config-data-generated\") pod \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.279307 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-galera-tls-certs\") pod \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\" (UID: \"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.279338 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-public-tls-certs\") pod \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\" (UID: \"0d3915c5-30d2-43be-abd4-438cd9d8ebf7\") " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.279753 4997 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe32fb42-61e6-4cd4-a75b-b684a4171224-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.289031 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" (UID: "7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.293779 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0d3915c5-30d2-43be-abd4-438cd9d8ebf7" (UID: "0d3915c5-30d2-43be-abd4-438cd9d8ebf7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.294910 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-logs" (OuterVolumeSpecName: "logs") pod "0d3915c5-30d2-43be-abd4-438cd9d8ebf7" (UID: "0d3915c5-30d2-43be-abd4-438cd9d8ebf7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.303936 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" (UID: "7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.304358 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" (UID: "7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.304872 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" (UID: "7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.384925 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-scripts" (OuterVolumeSpecName: "scripts") pod "fe32fb42-61e6-4cd4-a75b-b684a4171224" (UID: "fe32fb42-61e6-4cd4-a75b-b684a4171224"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.387148 4997 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.387165 4997 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.387177 4997 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.387191 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.387200 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.387208 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.387217 4997 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.387998 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fe32fb42-61e6-4cd4-a75b-b684a4171224" (UID: "fe32fb42-61e6-4cd4-a75b-b684a4171224"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.459975 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe32fb42-61e6-4cd4-a75b-b684a4171224-kube-api-access-vblvp" (OuterVolumeSpecName: "kube-api-access-vblvp") pod "fe32fb42-61e6-4cd4-a75b-b684a4171224" (UID: "fe32fb42-61e6-4cd4-a75b-b684a4171224"). InnerVolumeSpecName "kube-api-access-vblvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.490918 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vblvp\" (UniqueName: \"kubernetes.io/projected/fe32fb42-61e6-4cd4-a75b-b684a4171224-kube-api-access-vblvp\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.490952 4997 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: E1205 07:23:27.528790 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b81d606450f4ec2753efb3ba9b01b31144b20d257dc2ed0c1a6d4b204cf562b4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 07:23:27 crc kubenswrapper[4997]: E1205 07:23:27.544810 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b81d606450f4ec2753efb3ba9b01b31144b20d257dc2ed0c1a6d4b204cf562b4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 07:23:27 crc kubenswrapper[4997]: E1205 07:23:27.551396 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b81d606450f4ec2753efb3ba9b01b31144b20d257dc2ed0c1a6d4b204cf562b4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 07:23:27 crc kubenswrapper[4997]: E1205 07:23:27.551473 4997 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="c706bba7-965c-497f-ae7d-b7087d37d70e" containerName="nova-cell0-conductor-conductor" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.552255 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-scripts" (OuterVolumeSpecName: "scripts") pod "0d3915c5-30d2-43be-abd4-438cd9d8ebf7" (UID: "0d3915c5-30d2-43be-abd4-438cd9d8ebf7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.564392 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-kube-api-access-6mn64" (OuterVolumeSpecName: "kube-api-access-6mn64") pod "0d3915c5-30d2-43be-abd4-438cd9d8ebf7" (UID: "0d3915c5-30d2-43be-abd4-438cd9d8ebf7"). InnerVolumeSpecName "kube-api-access-6mn64". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.565174 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-kube-api-access-ffwz9" (OuterVolumeSpecName: "kube-api-access-ffwz9") pod "7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" (UID: "7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b"). InnerVolumeSpecName "kube-api-access-ffwz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.572898 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0d3915c5-30d2-43be-abd4-438cd9d8ebf7" (UID: "0d3915c5-30d2-43be-abd4-438cd9d8ebf7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.594443 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffwz9\" (UniqueName: \"kubernetes.io/projected/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-kube-api-access-ffwz9\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.594483 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mn64\" (UniqueName: \"kubernetes.io/projected/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-kube-api-access-6mn64\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.594494 4997 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.594503 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.615944 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "mysql-db") pod "7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" (UID: "7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.696239 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.736092 4997 generic.go:334] "Generic (PLEG): container finished" podID="9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed" containerID="96f6c22ff0cd4cb36b8fadc88840e34b4298c41452961e008b5dd7fc23ef4889" exitCode=0 Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.736173 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed","Type":"ContainerDied","Data":"96f6c22ff0cd4cb36b8fadc88840e34b4298c41452961e008b5dd7fc23ef4889"} Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.745735 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell02ec3-account-delete-8p9cc" event={"ID":"cf72e65a-bae4-4ce4-b65a-d5a587ef5f48","Type":"ContainerStarted","Data":"e0d6bff0be26dd0adebe52af15c98c304087e7208454b30644f8e9193c0a2c1f"} Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.807225 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.810431 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="126f5f4c-a3f5-4744-99cd-691bdd33444b" path="/var/lib/kubelet/pods/126f5f4c-a3f5-4744-99cd-691bdd33444b/volumes" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.812452 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f0e39f0-ca27-4656-879e-c7a20fcd904c" path="/var/lib/kubelet/pods/2f0e39f0-ca27-4656-879e-c7a20fcd904c/volumes" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.813138 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77e3250e-2f69-4047-8ea9-cce3b817cdb8" path="/var/lib/kubelet/pods/77e3250e-2f69-4047-8ea9-cce3b817cdb8/volumes" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.814327 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c6db4c3-5869-4a82-900c-5718abfe3a3b" path="/var/lib/kubelet/pods/7c6db4c3-5869-4a82-900c-5718abfe3a3b/volumes" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.815157 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb2d44ce-8c93-493e-9393-569c208f076f" path="/var/lib/kubelet/pods/cb2d44ce-8c93-493e-9393-569c208f076f/volumes" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.817230 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd426a72-68fd-46b9-b427-e193a70aad3a" path="/var/lib/kubelet/pods/dd426a72-68fd-46b9-b427-e193a70aad3a/volumes" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.818150 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="def8f706-0e1b-43ab-8d4e-2cab420b795d" path="/var/lib/kubelet/pods/def8f706-0e1b-43ab-8d4e-2cab420b795d/volumes" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.838537 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.856136 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 07:23:27 crc kubenswrapper[4997]: I1205 07:23:27.874548 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.005129 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.013618 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-66698d9c4f-5rmpv" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.067294 4997 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.113161 4997 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.198788 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-config-data" (OuterVolumeSpecName: "config-data") pod "0abba971-b935-4cb0-865e-b364f8521f81" (UID: "0abba971-b935-4cb0-865e-b364f8521f81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.216563 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.255526 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0d3915c5-30d2-43be-abd4-438cd9d8ebf7" (UID: "0d3915c5-30d2-43be-abd4-438cd9d8ebf7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: E1205 07:23:28.307096 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:28 crc kubenswrapper[4997]: E1205 07:23:28.307301 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:28 crc kubenswrapper[4997]: E1205 07:23:28.310078 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:28 crc kubenswrapper[4997]: E1205 07:23:28.310895 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:28 crc kubenswrapper[4997]: E1205 07:23:28.310933 4997 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sctbp" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovsdb-server" Dec 05 07:23:28 crc kubenswrapper[4997]: E1205 07:23:28.315103 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:28 crc kubenswrapper[4997]: E1205 07:23:28.319484 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:28 crc kubenswrapper[4997]: E1205 07:23:28.319591 4997 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sctbp" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovs-vswitchd" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.322962 4997 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.339076 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" (UID: "7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.350676 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe32fb42-61e6-4cd4-a75b-b684a4171224" (UID: "fe32fb42-61e6-4cd4-a75b-b684a4171224"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.431604 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.431651 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.447900 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0d3915c5-30d2-43be-abd4-438cd9d8ebf7" (UID: "0d3915c5-30d2-43be-abd4-438cd9d8ebf7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.469888 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8374cc67-e735-47f0-b310-6bebd608cece-config-data" (OuterVolumeSpecName: "config-data") pod "8374cc67-e735-47f0-b310-6bebd608cece" (UID: "8374cc67-e735-47f0-b310-6bebd608cece"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.487823 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0abba971-b935-4cb0-865e-b364f8521f81" (UID: "0abba971-b935-4cb0-865e-b364f8521f81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.513041 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8374cc67-e735-47f0-b310-6bebd608cece-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8374cc67-e735-47f0-b310-6bebd608cece" (UID: "8374cc67-e735-47f0-b310-6bebd608cece"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.526970 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vnncj" podUID="5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" containerName="ovn-controller" probeResult="failure" output=< Dec 05 07:23:28 crc kubenswrapper[4997]: ERROR - Failed to get connection status from ovn-controller, ovn-appctl exit status: 0 Dec 05 07:23:28 crc kubenswrapper[4997]: > Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.537554 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8374cc67-e735-47f0-b310-6bebd608cece-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.537825 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8374cc67-e735-47f0-b310-6bebd608cece-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.537911 4997 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.537996 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.543211 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d3915c5-30d2-43be-abd4-438cd9d8ebf7" (UID: "0d3915c5-30d2-43be-abd4-438cd9d8ebf7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.543547 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" (UID: "7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.561393 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "0abba971-b935-4cb0-865e-b364f8521f81" (UID: "0abba971-b935-4cb0-865e-b364f8521f81"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.581643 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "27ff3957-ee08-40be-a41d-02979f192fda" (UID: "27ff3957-ee08-40be-a41d-02979f192fda"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.593962 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-config-data" (OuterVolumeSpecName: "config-data") pod "27ff3957-ee08-40be-a41d-02979f192fda" (UID: "27ff3957-ee08-40be-a41d-02979f192fda"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.600818 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "0abba971-b935-4cb0-865e-b364f8521f81" (UID: "0abba971-b935-4cb0-865e-b364f8521f81"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.601671 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-config-data" (OuterVolumeSpecName: "config-data") pod "0d3915c5-30d2-43be-abd4-438cd9d8ebf7" (UID: "0d3915c5-30d2-43be-abd4-438cd9d8ebf7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.610758 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "27ff3957-ee08-40be-a41d-02979f192fda" (UID: "27ff3957-ee08-40be-a41d-02979f192fda"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.626265 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27ff3957-ee08-40be-a41d-02979f192fda" (UID: "27ff3957-ee08-40be-a41d-02979f192fda"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.641522 4997 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.641643 4997 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.641658 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.641674 4997 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.641689 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.641700 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d3915c5-30d2-43be-abd4-438cd9d8ebf7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.641713 4997 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/0abba971-b935-4cb0-865e-b364f8521f81-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.641726 4997 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.641737 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27ff3957-ee08-40be-a41d-02979f192fda-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.719889 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-config-data" (OuterVolumeSpecName: "config-data") pod "fe32fb42-61e6-4cd4-a75b-b684a4171224" (UID: "fe32fb42-61e6-4cd4-a75b-b684a4171224"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:28 crc kubenswrapper[4997]: I1205 07:23:28.744359 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe32fb42-61e6-4cd4-a75b-b684a4171224-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.073413 4997 generic.go:334] "Generic (PLEG): container finished" podID="13fc4b26-4375-4ea2-8eaa-2417d64b3cd1" containerID="63372a7e68a9d93e249283016625478d1f7e92640812a3890997060ed1267282" exitCode=0 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.077196 4997 generic.go:334] "Generic (PLEG): container finished" podID="67d9e01d-5189-4aac-8f1a-b1f09fe138b8" containerID="793a4abf63252321ba4fb0dafb87c02e599c7ee609d534abf01e11eb3eff2e86" exitCode=0 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.080078 4997 generic.go:334] "Generic (PLEG): container finished" podID="c51acefb-ed8a-480c-9ab6-0345f2ff0cd3" containerID="84c9f79e02c9e635085a99dbc7faaaf90bfa04701ecbeee66bafb3d0bf7cf96b" exitCode=0 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.097138 4997 generic.go:334] "Generic (PLEG): container finished" podID="e08a39e3-f267-496d-80b3-b12a9eef14c1" containerID="d09f8f75d78ec8de394fe2d3878fbdc0a819558c405bcad43068ea535ea0a421" exitCode=0 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.124455 4997 generic.go:334] "Generic (PLEG): container finished" podID="5afeb230-4a35-4872-bffa-c882fe927643" containerID="27d910165cbca8f6eb45e1a947fe3daf0e9d27a117dff110f3fc7aeef2a06961" exitCode=0 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.126443 4997 generic.go:334] "Generic (PLEG): container finished" podID="a6bfb8b2-df28-4992-8831-2a5060323ddf" containerID="b77018d13ada4b2350299017976f4a2b5e3d76cc39ecbc054f294f67cbfa4f8f" exitCode=0 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.129257 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novacell02ec3-account-delete-8p9cc" podStartSLOduration=7.129224044 podStartE2EDuration="7.129224044s" podCreationTimestamp="2025-12-05 07:23:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:23:27.77358697 +0000 UTC m=+1708.302494231" watchObservedRunningTime="2025-12-05 07:23:29.129224044 +0000 UTC m=+1709.658131305" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.130797 4997 generic.go:334] "Generic (PLEG): container finished" podID="cf72e65a-bae4-4ce4-b65a-d5a587ef5f48" containerID="e0d6bff0be26dd0adebe52af15c98c304087e7208454b30644f8e9193c0a2c1f" exitCode=0 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.143633 4997 generic.go:334] "Generic (PLEG): container finished" podID="b8c5aa81-a666-4ae6-a41c-fd31df260223" containerID="331b3a19e5c1b6a99d65e10dc3dc2975027b515da4a33c5e24c9c58c4f701c55" exitCode=0 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.197399 4997 generic.go:334] "Generic (PLEG): container finished" podID="d65cd492-d4bd-4489-9d28-a6c42407101a" containerID="8a2e07a2742d3964e8caee339bfc34622988235ccfcc24681f0a0082129f5702" exitCode=0 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.256121 4997 generic.go:334] "Generic (PLEG): container finished" podID="bcbe26f4-fcd9-46c7-aa3a-5038b801741d" containerID="96ac136074ed0bcb85702f5bfe22b13a7895276ce80e0413118cb761bfafcad0" exitCode=0 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.304423 4997 generic.go:334] "Generic (PLEG): container finished" podID="2a733d85-ad49-41b8-a75b-842dce56e85c" containerID="08946f5deeba339b1b854496b842f982af5b2b54721ebee20c813990c70fbe66" exitCode=0 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.370566 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5cb79f8cdd-zkk89" podUID="2179d060-8f4c-413d-a202-e31ebc242dfc" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": dial tcp 10.217.0.159:9311: connect: connection refused" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.370887 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5cb79f8cdd-zkk89" podUID="2179d060-8f4c-413d-a202-e31ebc242dfc" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": dial tcp 10.217.0.159:9311: connect: connection refused" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.598861 4997 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.599324 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-config-data podName:1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:37.59930264 +0000 UTC m=+1718.128209901 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-config-data") pod "rabbitmq-server-0" (UID: "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79") : configmap "rabbitmq-config-data" not found Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.717078 4997 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.969s" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.717123 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fe32fb42-61e6-4cd4-a75b-b684a4171224","Type":"ContainerDied","Data":"b35c9b6435ee745827f49e345e658de1e9600ff778a331dca340af2f70164c49"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.717260 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0abba971-b935-4cb0-865e-b364f8521f81","Type":"ContainerDied","Data":"40db7048fdd40f1c31e882668f3781cab1f076430d89705b142af270080d6e41"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.717280 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b","Type":"ContainerDied","Data":"7a399ac2c7f99edcffa9919bc82a175c87e153671bf6886319821c1a6e648b0b"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.717299 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8374cc67-e735-47f0-b310-6bebd608cece","Type":"ContainerDied","Data":"18cf304d9afa92bcfc6cfed42f01f81ce744f5e8753b217fbfb36d68b769eed7"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.717315 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.717342 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.717370 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.717387 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance7904-account-delete-v5xm8" event={"ID":"8f2677aa-a856-47de-8417-5bc3d08d6ff1","Type":"ContainerStarted","Data":"705ac0ba630b3336a44765fcb97172ba22ee01af0565e74f3793afff5a05fc9a"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.717400 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi8ba6-account-delete-t8gw7" event={"ID":"a5edac3e-6bee-494f-97ba-c4047064d02f","Type":"ContainerStarted","Data":"4c10b2bcf8b062435e447eb41d03529fe258d8e27842fe2f7a56df03885dbf7e"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.717413 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-f69dn"] Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.717430 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-f69dn"] Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.717451 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone3eec-account-delete-v5k7n"] Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.717861 4997 scope.go:117] "RemoveContainer" containerID="921875cb2f26d740baa583b8c4418587dfe774e4ea61c9ad78bb1723d653c892" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718122 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" containerName="galera" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718172 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" containerName="galera" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718189 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="def8f706-0e1b-43ab-8d4e-2cab420b795d" containerName="dnsmasq-dns" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718196 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="def8f706-0e1b-43ab-8d4e-2cab420b795d" containerName="dnsmasq-dns" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718203 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb2d44ce-8c93-493e-9393-569c208f076f" containerName="openstack-network-exporter" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718210 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb2d44ce-8c93-493e-9393-569c208f076f" containerName="openstack-network-exporter" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718221 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f0e39f0-ca27-4656-879e-c7a20fcd904c" containerName="openstack-network-exporter" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718248 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f0e39f0-ca27-4656-879e-c7a20fcd904c" containerName="openstack-network-exporter" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718265 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c6db4c3-5869-4a82-900c-5718abfe3a3b" containerName="openstack-network-exporter" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718270 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c6db4c3-5869-4a82-900c-5718abfe3a3b" containerName="openstack-network-exporter" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718283 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd426a72-68fd-46b9-b427-e193a70aad3a" containerName="ovsdbserver-nb" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718289 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd426a72-68fd-46b9-b427-e193a70aad3a" containerName="ovsdbserver-nb" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718299 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd426a72-68fd-46b9-b427-e193a70aad3a" containerName="openstack-network-exporter" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718306 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd426a72-68fd-46b9-b427-e193a70aad3a" containerName="openstack-network-exporter" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718337 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27ff3957-ee08-40be-a41d-02979f192fda" containerName="proxy-httpd" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718344 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="27ff3957-ee08-40be-a41d-02979f192fda" containerName="proxy-httpd" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718352 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27ff3957-ee08-40be-a41d-02979f192fda" containerName="proxy-server" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718358 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="27ff3957-ee08-40be-a41d-02979f192fda" containerName="proxy-server" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718368 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe32fb42-61e6-4cd4-a75b-b684a4171224" containerName="cinder-scheduler" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718374 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe32fb42-61e6-4cd4-a75b-b684a4171224" containerName="cinder-scheduler" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718384 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe32fb42-61e6-4cd4-a75b-b684a4171224" containerName="probe" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718407 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe32fb42-61e6-4cd4-a75b-b684a4171224" containerName="probe" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718417 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d3915c5-30d2-43be-abd4-438cd9d8ebf7" containerName="cinder-api-log" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718424 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d3915c5-30d2-43be-abd4-438cd9d8ebf7" containerName="cinder-api-log" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718435 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8374cc67-e735-47f0-b310-6bebd608cece" containerName="nova-cell1-conductor-conductor" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718443 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8374cc67-e735-47f0-b310-6bebd608cece" containerName="nova-cell1-conductor-conductor" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718499 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f0e39f0-ca27-4656-879e-c7a20fcd904c" containerName="ovn-northd" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718508 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f0e39f0-ca27-4656-879e-c7a20fcd904c" containerName="ovn-northd" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718520 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" containerName="mysql-bootstrap" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718526 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" containerName="mysql-bootstrap" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718536 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c6db4c3-5869-4a82-900c-5718abfe3a3b" containerName="ovsdbserver-sb" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718561 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c6db4c3-5869-4a82-900c-5718abfe3a3b" containerName="ovsdbserver-sb" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718570 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="def8f706-0e1b-43ab-8d4e-2cab420b795d" containerName="init" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718577 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="def8f706-0e1b-43ab-8d4e-2cab420b795d" containerName="init" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718592 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d3915c5-30d2-43be-abd4-438cd9d8ebf7" containerName="cinder-api" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718598 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d3915c5-30d2-43be-abd4-438cd9d8ebf7" containerName="cinder-api" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.718651 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0abba971-b935-4cb0-865e-b364f8521f81" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718660 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0abba971-b935-4cb0-865e-b364f8521f81" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718980 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb2d44ce-8c93-493e-9393-569c208f076f" containerName="openstack-network-exporter" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.718992 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d3915c5-30d2-43be-abd4-438cd9d8ebf7" containerName="cinder-api" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.719005 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd426a72-68fd-46b9-b427-e193a70aad3a" containerName="openstack-network-exporter" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.719036 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f0e39f0-ca27-4656-879e-c7a20fcd904c" containerName="openstack-network-exporter" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.719047 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f0e39f0-ca27-4656-879e-c7a20fcd904c" containerName="ovn-northd" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.719056 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="def8f706-0e1b-43ab-8d4e-2cab420b795d" containerName="dnsmasq-dns" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.719069 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c6db4c3-5869-4a82-900c-5718abfe3a3b" containerName="ovsdbserver-sb" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.719083 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd426a72-68fd-46b9-b427-e193a70aad3a" containerName="ovsdbserver-nb" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.719113 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0abba971-b935-4cb0-865e-b364f8521f81" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.719124 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" containerName="galera" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.719132 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="27ff3957-ee08-40be-a41d-02979f192fda" containerName="proxy-server" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.719140 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d3915c5-30d2-43be-abd4-438cd9d8ebf7" containerName="cinder-api-log" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.719149 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="8374cc67-e735-47f0-b310-6bebd608cece" containerName="nova-cell1-conductor-conductor" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.719160 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="27ff3957-ee08-40be-a41d-02979f192fda" containerName="proxy-httpd" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.719169 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe32fb42-61e6-4cd4-a75b-b684a4171224" containerName="cinder-scheduler" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.719203 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe32fb42-61e6-4cd4-a75b-b684a4171224" containerName="probe" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.719215 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c6db4c3-5869-4a82-900c-5718abfe3a3b" containerName="openstack-network-exporter" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.722227 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db136058-17a8-4955-8e95-576da7b9e847" containerName="ceilometer-central-agent" containerID="cri-o://07f1631824a73521a5df4d80b3bf4feb8dde04c7a5b2c5e5688b95c1efabb84c" gracePeriod=30 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.723061 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db136058-17a8-4955-8e95-576da7b9e847" containerName="sg-core" containerID="cri-o://3fc11187932225dbd83eb0de61fd45f5e409020f11e6e1a8a12baa4ea259218b" gracePeriod=30 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.722984 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db136058-17a8-4955-8e95-576da7b9e847" containerName="proxy-httpd" containerID="cri-o://fdb72dc3adf6464397dca00e5b9745e3a6b89a71985325529d858ab69ce19e3c" gracePeriod=30 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.723145 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db136058-17a8-4955-8e95-576da7b9e847" containerName="ceilometer-notification-agent" containerID="cri-o://b3cfedfae8a42f278d8fff88f2e1586ca243b6222c804510c5b0435a9a5e1e73" gracePeriod=30 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.723465 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946" containerName="kube-state-metrics" containerID="cri-o://d9261932ff03a509a886dc6ed5c881f90c33d3fd0c490c206516f3523389c13c" gracePeriod=30 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.723884 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="ef43a612-a3df-4b50-9bf3-e5d8098cd0fe" containerName="memcached" containerID="cri-o://175b217107beb960b8632016072bdd9ad602c7c079fe319f224a856dd6e14c82" gracePeriod=30 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726262 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone3eec-account-delete-v5k7n"] Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726306 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-7dc7z"] Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726323 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-7dc7z"] Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726339 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5477474c5d-zs25t"] Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726358 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0d3915c5-30d2-43be-abd4-438cd9d8ebf7","Type":"ContainerDied","Data":"5a78b3f3d78b1711cb5549e03a9ce76384ae3d42e9846ce209eed741535e916f"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726390 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726410 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-db8868644-mchk9" event={"ID":"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1","Type":"ContainerDied","Data":"63372a7e68a9d93e249283016625478d1f7e92640812a3890997060ed1267282"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726425 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67d9e01d-5189-4aac-8f1a-b1f09fe138b8","Type":"ContainerDied","Data":"793a4abf63252321ba4fb0dafb87c02e599c7ee609d534abf01e11eb3eff2e86"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726441 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67d9e01d-5189-4aac-8f1a-b1f09fe138b8","Type":"ContainerDied","Data":"784b13fb1238ccd16ae49390528e4338365e3eaa7c310f19fc69bb52dc47a6b7"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726453 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="784b13fb1238ccd16ae49390528e4338365e3eaa7c310f19fc69bb52dc47a6b7" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726467 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement94df-account-delete-4qxc6" event={"ID":"c51acefb-ed8a-480c-9ab6-0345f2ff0cd3","Type":"ContainerDied","Data":"84c9f79e02c9e635085a99dbc7faaaf90bfa04701ecbeee66bafb3d0bf7cf96b"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726488 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e08a39e3-f267-496d-80b3-b12a9eef14c1","Type":"ContainerDied","Data":"d09f8f75d78ec8de394fe2d3878fbdc0a819558c405bcad43068ea535ea0a421"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726505 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron84af-account-delete-pgl29" event={"ID":"68490a68-dc32-448f-90cc-c146ba49c1a8","Type":"ContainerDied","Data":"fb2b75957eaf704e76b01bf7cb86971ae443f6f4cb671f899c836d9e2f573577"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726518 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb2b75957eaf704e76b01bf7cb86971ae443f6f4cb671f899c836d9e2f573577" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726535 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5afeb230-4a35-4872-bffa-c882fe927643","Type":"ContainerDied","Data":"27d910165cbca8f6eb45e1a947fe3daf0e9d27a117dff110f3fc7aeef2a06961"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726560 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5afeb230-4a35-4872-bffa-c882fe927643","Type":"ContainerDied","Data":"5082ac2136f164cf222572a1bf4fe1a02e45628512baa05a6250afdfaff826ff"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726570 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5082ac2136f164cf222572a1bf4fe1a02e45628512baa05a6250afdfaff826ff" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726582 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-555c8b7b77-44qw4" event={"ID":"a6bfb8b2-df28-4992-8831-2a5060323ddf","Type":"ContainerDied","Data":"b77018d13ada4b2350299017976f4a2b5e3d76cc39ecbc054f294f67cbfa4f8f"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726614 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed","Type":"ContainerDied","Data":"8fcc5ac3680b52427d3eabdfd38a2e59e4209c08c59501e314ae1a992f0a4113"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726646 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fcc5ac3680b52427d3eabdfd38a2e59e4209c08c59501e314ae1a992f0a4113" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726652 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone3eec-account-delete-v5k7n" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726658 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell02ec3-account-delete-8p9cc" event={"ID":"cf72e65a-bae4-4ce4-b65a-d5a587ef5f48","Type":"ContainerDied","Data":"e0d6bff0be26dd0adebe52af15c98c304087e7208454b30644f8e9193c0a2c1f"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726676 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cindera458-account-delete-vpgnx" event={"ID":"b8c5aa81-a666-4ae6-a41c-fd31df260223","Type":"ContainerDied","Data":"331b3a19e5c1b6a99d65e10dc3dc2975027b515da4a33c5e24c9c58c4f701c55"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726696 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican69df-account-delete-tqlsz" event={"ID":"d65cd492-d4bd-4489-9d28-a6c42407101a","Type":"ContainerDied","Data":"8a2e07a2742d3964e8caee339bfc34622988235ccfcc24681f0a0082129f5702"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726714 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bcbe26f4-fcd9-46c7-aa3a-5038b801741d","Type":"ContainerDied","Data":"96ac136074ed0bcb85702f5bfe22b13a7895276ce80e0413118cb761bfafcad0"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726730 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bcbe26f4-fcd9-46c7-aa3a-5038b801741d","Type":"ContainerDied","Data":"c8cefb27923ca6a1b8c5bdd6a3af492257b89f880b6c87eec48e6a7441270d51"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726742 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8cefb27923ca6a1b8c5bdd6a3af492257b89f880b6c87eec48e6a7441270d51" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726753 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-54ff4c8496-8z64m" event={"ID":"2a733d85-ad49-41b8-a75b-842dce56e85c","Type":"ContainerDied","Data":"08946f5deeba339b1b854496b842f982af5b2b54721ebee20c813990c70fbe66"} Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.726926 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-5477474c5d-zs25t" podUID="6783d79d-4874-43c1-92bc-e09133e45989" containerName="keystone-api" containerID="cri-o://99025ec523a5c575412ef4a88b541702a67a5c68296982dfdc6d2fccabd2bb56" gracePeriod=30 Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.735756 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-ktw8k"] Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.789331 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="145b182b-23c5-444d-864f-0cbd2c46902d" path="/var/lib/kubelet/pods/145b182b-23c5-444d-864f-0cbd2c46902d/volumes" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.790051 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8c2c30d-a149-48f0-9995-6ef963e5f50c" path="/var/lib/kubelet/pods/c8c2c30d-a149-48f0-9995-6ef963e5f50c/volumes" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.794011 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.797903 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-ktw8k"] Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.821296 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx4dp\" (UniqueName: \"kubernetes.io/projected/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-kube-api-access-sx4dp\") pod \"keystone3eec-account-delete-v5k7n\" (UID: \"c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1\") " pod="openstack/keystone3eec-account-delete-v5k7n" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.821340 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-operator-scripts\") pod \"keystone3eec-account-delete-v5k7n\" (UID: \"c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1\") " pod="openstack/keystone3eec-account-delete-v5k7n" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.853278 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-3eec-account-create-update-tckr2"] Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.907850 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-3eec-account-create-update-tckr2"] Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.909197 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone3eec-account-delete-v5k7n"] Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.926284 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-operator-scripts\") pod \"keystone3eec-account-delete-v5k7n\" (UID: \"c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1\") " pod="openstack/keystone3eec-account-delete-v5k7n" Dec 05 07:23:29 crc kubenswrapper[4997]: I1205 07:23:29.926533 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx4dp\" (UniqueName: \"kubernetes.io/projected/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-kube-api-access-sx4dp\") pod \"keystone3eec-account-delete-v5k7n\" (UID: \"c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1\") " pod="openstack/keystone3eec-account-delete-v5k7n" Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.926996 4997 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.927049 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-operator-scripts podName:c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:30.427029805 +0000 UTC m=+1710.955937066 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-operator-scripts") pod "keystone3eec-account-delete-v5k7n" (UID: "c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1") : configmap "openstack-scripts" not found Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.936180 4997 projected.go:194] Error preparing data for projected volume kube-api-access-sx4dp for pod openstack/keystone3eec-account-delete-v5k7n: failed to fetch token: serviceaccounts "galera-openstack" not found Dec 05 07:23:29 crc kubenswrapper[4997]: E1205 07:23:29.936276 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-kube-api-access-sx4dp podName:c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:30.436250315 +0000 UTC m=+1710.965157576 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-sx4dp" (UniqueName: "kubernetes.io/projected/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-kube-api-access-sx4dp") pod "keystone3eec-account-delete-v5k7n" (UID: "c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1") : failed to fetch token: serviceaccounts "galera-openstack" not found Dec 05 07:23:30 crc kubenswrapper[4997]: E1205 07:23:30.031548 4997 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 05 07:23:30 crc kubenswrapper[4997]: E1205 07:23:30.032411 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-config-data podName:0f158874-152b-46ce-ac42-d202684853ca nodeName:}" failed. No retries permitted until 2025-12-05 07:23:38.032381831 +0000 UTC m=+1718.561289092 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-config-data") pod "rabbitmq-cell1-server-0" (UID: "0f158874-152b-46ce-ac42-d202684853ca") : configmap "rabbitmq-cell1-config-data" not found Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.039543 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="0601e553-5305-4f54-9a2d-43b43a1a5dd4" containerName="galera" containerID="cri-o://22c17242a1211bc3c93181daebd1e55ec7c7f3dfb0c90a0942488ba51f04fe30" gracePeriod=30 Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.073433 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.133249 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-combined-ca-bundle\") pod \"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed\" (UID: \"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.133563 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbz65\" (UniqueName: \"kubernetes.io/projected/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-kube-api-access-nbz65\") pod \"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed\" (UID: \"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.133709 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-config-data\") pod \"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed\" (UID: \"9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.145964 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-kube-api-access-nbz65" (OuterVolumeSpecName: "kube-api-access-nbz65") pod "9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed" (UID: "9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed"). InnerVolumeSpecName "kube-api-access-nbz65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.150555 4997 scope.go:117] "RemoveContainer" containerID="b762656e8963968dd62e28b778f1374e72ed4bfb67fbcee7edeacd3920071001" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.227281 4997 scope.go:117] "RemoveContainer" containerID="d4ab663388d68e77285b0ecaf1e1e3e7993510fd57825fc77051658b8edf094e" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.246190 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbz65\" (UniqueName: \"kubernetes.io/projected/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-kube-api-access-nbz65\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.267130 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-config-data" (OuterVolumeSpecName: "config-data") pod "9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed" (UID: "9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.271942 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed" (UID: "9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.337736 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-db8868644-mchk9" event={"ID":"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1","Type":"ContainerDied","Data":"be77616b5c8937572e9e713ab0982f57a84b605ed475f48883932d374be43cfa"} Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.338222 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be77616b5c8937572e9e713ab0982f57a84b605ed475f48883932d374be43cfa" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.345376 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.349175 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.349216 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.352266 4997 generic.go:334] "Generic (PLEG): container finished" podID="2179d060-8f4c-413d-a202-e31ebc242dfc" containerID="d14f08fb852081fdc4815d2ce697e3418dfd98d7534ab2afb3c35a5bfea2457a" exitCode=0 Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.352366 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cb79f8cdd-zkk89" event={"ID":"2179d060-8f4c-413d-a202-e31ebc242dfc","Type":"ContainerDied","Data":"d14f08fb852081fdc4815d2ce697e3418dfd98d7534ab2afb3c35a5bfea2457a"} Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.352410 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cb79f8cdd-zkk89" event={"ID":"2179d060-8f4c-413d-a202-e31ebc242dfc","Type":"ContainerDied","Data":"c89a73a98da50d8a158529471b379a982fbb6c19bed2b42f34c604c7ff17f3b3"} Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.352424 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c89a73a98da50d8a158529471b379a982fbb6c19bed2b42f34c604c7ff17f3b3" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.377315 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron84af-account-delete-pgl29" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.394039 4997 generic.go:334] "Generic (PLEG): container finished" podID="db136058-17a8-4955-8e95-576da7b9e847" containerID="fdb72dc3adf6464397dca00e5b9745e3a6b89a71985325529d858ab69ce19e3c" exitCode=0 Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.394393 4997 generic.go:334] "Generic (PLEG): container finished" podID="db136058-17a8-4955-8e95-576da7b9e847" containerID="3fc11187932225dbd83eb0de61fd45f5e409020f11e6e1a8a12baa4ea259218b" exitCode=2 Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.394250 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db136058-17a8-4955-8e95-576da7b9e847","Type":"ContainerDied","Data":"fdb72dc3adf6464397dca00e5b9745e3a6b89a71985325529d858ab69ce19e3c"} Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.395109 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db136058-17a8-4955-8e95-576da7b9e847","Type":"ContainerDied","Data":"3fc11187932225dbd83eb0de61fd45f5e409020f11e6e1a8a12baa4ea259218b"} Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.421993 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.423509 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e08a39e3-f267-496d-80b3-b12a9eef14c1","Type":"ContainerDied","Data":"89452f4406872d0491819e7c170b3af0dc88bb38d26f4432ce97d3d75d550177"} Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.423547 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89452f4406872d0491819e7c170b3af0dc88bb38d26f4432ce97d3d75d550177" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.428770 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.442040 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-54ff4c8496-8z64m" event={"ID":"2a733d85-ad49-41b8-a75b-842dce56e85c","Type":"ContainerDied","Data":"954c36632faeae217ead931414bfae62826ed6239cd9550f2be576495871cc20"} Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.442097 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="954c36632faeae217ead931414bfae62826ed6239cd9550f2be576495871cc20" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.451220 4997 scope.go:117] "RemoveContainer" containerID="5f394487f180524c3058b7242cb6bd0e8d1659b04d10a82da617c25ca61bffc0" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.451376 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-logs\") pod \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.451458 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.451568 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-internal-tls-certs\") pod \"5afeb230-4a35-4872-bffa-c882fe927643\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.451650 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68490a68-dc32-448f-90cc-c146ba49c1a8-operator-scripts\") pod \"68490a68-dc32-448f-90cc-c146ba49c1a8\" (UID: \"68490a68-dc32-448f-90cc-c146ba49c1a8\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.451712 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-combined-ca-bundle\") pod \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.451735 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-scripts\") pod \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.451789 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-httpd-run\") pod \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.451815 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-logs\") pod \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.451861 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-config-data\") pod \"5afeb230-4a35-4872-bffa-c882fe927643\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.451882 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-public-tls-certs\") pod \"5afeb230-4a35-4872-bffa-c882fe927643\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.451974 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-combined-ca-bundle\") pod \"5afeb230-4a35-4872-bffa-c882fe927643\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.452024 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-public-tls-certs\") pod \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.452050 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2lxh\" (UniqueName: \"kubernetes.io/projected/68490a68-dc32-448f-90cc-c146ba49c1a8-kube-api-access-w2lxh\") pod \"68490a68-dc32-448f-90cc-c146ba49c1a8\" (UID: \"68490a68-dc32-448f-90cc-c146ba49c1a8\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.452128 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb885\" (UniqueName: \"kubernetes.io/projected/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-kube-api-access-zb885\") pod \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.452172 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-nova-metadata-tls-certs\") pod \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.452198 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m4pw\" (UniqueName: \"kubernetes.io/projected/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-kube-api-access-6m4pw\") pod \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.452218 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5m8mj\" (UniqueName: \"kubernetes.io/projected/5afeb230-4a35-4872-bffa-c882fe927643-kube-api-access-5m8mj\") pod \"5afeb230-4a35-4872-bffa-c882fe927643\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.452285 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5afeb230-4a35-4872-bffa-c882fe927643-logs\") pod \"5afeb230-4a35-4872-bffa-c882fe927643\" (UID: \"5afeb230-4a35-4872-bffa-c882fe927643\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.452356 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-config-data\") pod \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\" (UID: \"bcbe26f4-fcd9-46c7-aa3a-5038b801741d\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.452427 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.452459 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-config-data\") pod \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.452600 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-combined-ca-bundle\") pod \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\" (UID: \"67d9e01d-5189-4aac-8f1a-b1f09fe138b8\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.453071 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx4dp\" (UniqueName: \"kubernetes.io/projected/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-kube-api-access-sx4dp\") pod \"keystone3eec-account-delete-v5k7n\" (UID: \"c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1\") " pod="openstack/keystone3eec-account-delete-v5k7n" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.453107 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-operator-scripts\") pod \"keystone3eec-account-delete-v5k7n\" (UID: \"c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1\") " pod="openstack/keystone3eec-account-delete-v5k7n" Dec 05 07:23:30 crc kubenswrapper[4997]: E1205 07:23:30.453377 4997 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 07:23:30 crc kubenswrapper[4997]: E1205 07:23:30.453434 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-operator-scripts podName:c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:31.453416366 +0000 UTC m=+1711.982323617 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-operator-scripts") pod "keystone3eec-account-delete-v5k7n" (UID: "c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1") : configmap "openstack-scripts" not found Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.459845 4997 generic.go:334] "Generic (PLEG): container finished" podID="c706bba7-965c-497f-ae7d-b7087d37d70e" containerID="b81d606450f4ec2753efb3ba9b01b31144b20d257dc2ed0c1a6d4b204cf562b4" exitCode=0 Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.460001 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c706bba7-965c-497f-ae7d-b7087d37d70e","Type":"ContainerDied","Data":"b81d606450f4ec2753efb3ba9b01b31144b20d257dc2ed0c1a6d4b204cf562b4"} Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.460041 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c706bba7-965c-497f-ae7d-b7087d37d70e","Type":"ContainerDied","Data":"94189baad8432517afd7770bd384f234f9e285019dcb23f7f78b04388f24dadd"} Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.460056 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94189baad8432517afd7770bd384f234f9e285019dcb23f7f78b04388f24dadd" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.463683 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5afeb230-4a35-4872-bffa-c882fe927643-logs" (OuterVolumeSpecName: "logs") pod "5afeb230-4a35-4872-bffa-c882fe927643" (UID: "5afeb230-4a35-4872-bffa-c882fe927643"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.472170 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68490a68-dc32-448f-90cc-c146ba49c1a8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "68490a68-dc32-448f-90cc-c146ba49c1a8" (UID: "68490a68-dc32-448f-90cc-c146ba49c1a8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.473334 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "67d9e01d-5189-4aac-8f1a-b1f09fe138b8" (UID: "67d9e01d-5189-4aac-8f1a-b1f09fe138b8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.473368 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-logs" (OuterVolumeSpecName: "logs") pod "bcbe26f4-fcd9-46c7-aa3a-5038b801741d" (UID: "bcbe26f4-fcd9-46c7-aa3a-5038b801741d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.473422 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-logs" (OuterVolumeSpecName: "logs") pod "67d9e01d-5189-4aac-8f1a-b1f09fe138b8" (UID: "67d9e01d-5189-4aac-8f1a-b1f09fe138b8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: E1205 07:23:30.476762 4997 projected.go:194] Error preparing data for projected volume kube-api-access-sx4dp for pod openstack/keystone3eec-account-delete-v5k7n: failed to fetch token: serviceaccounts "galera-openstack" not found Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.480565 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-555c8b7b77-44qw4" event={"ID":"a6bfb8b2-df28-4992-8831-2a5060323ddf","Type":"ContainerDied","Data":"a602e9160cb6a58142ca3e52d7ec8d6dd98eafd893d7a66da580770ef4ba8ac2"} Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.480670 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a602e9160cb6a58142ca3e52d7ec8d6dd98eafd893d7a66da580770ef4ba8ac2" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.482310 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 07:23:30 crc kubenswrapper[4997]: E1205 07:23:30.482785 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-kube-api-access-sx4dp podName:c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:31.479317109 +0000 UTC m=+1712.008224550 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-sx4dp" (UniqueName: "kubernetes.io/projected/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-kube-api-access-sx4dp") pod "keystone3eec-account-delete-v5k7n" (UID: "c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1") : failed to fetch token: serviceaccounts "galera-openstack" not found Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.489078 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68490a68-dc32-448f-90cc-c146ba49c1a8-kube-api-access-w2lxh" (OuterVolumeSpecName: "kube-api-access-w2lxh") pod "68490a68-dc32-448f-90cc-c146ba49c1a8" (UID: "68490a68-dc32-448f-90cc-c146ba49c1a8"). InnerVolumeSpecName "kube-api-access-w2lxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.489318 4997 generic.go:334] "Generic (PLEG): container finished" podID="f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946" containerID="d9261932ff03a509a886dc6ed5c881f90c33d3fd0c490c206516f3523389c13c" exitCode=2 Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.489346 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946","Type":"ContainerDied","Data":"d9261932ff03a509a886dc6ed5c881f90c33d3fd0c490c206516f3523389c13c"} Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.491922 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.492860 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.499326 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.509905 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.517913 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.555079 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkptj\" (UniqueName: \"kubernetes.io/projected/2a733d85-ad49-41b8-a75b-842dce56e85c-kube-api-access-vkptj\") pod \"2a733d85-ad49-41b8-a75b-842dce56e85c\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.576135 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5afeb230-4a35-4872-bffa-c882fe927643-kube-api-access-5m8mj" (OuterVolumeSpecName: "kube-api-access-5m8mj") pod "5afeb230-4a35-4872-bffa-c882fe927643" (UID: "5afeb230-4a35-4872-bffa-c882fe927643"). InnerVolumeSpecName "kube-api-access-5m8mj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.576288 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-kube-api-access-6m4pw" (OuterVolumeSpecName: "kube-api-access-6m4pw") pod "67d9e01d-5189-4aac-8f1a-b1f09fe138b8" (UID: "67d9e01d-5189-4aac-8f1a-b1f09fe138b8"). InnerVolumeSpecName "kube-api-access-6m4pw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.576962 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-kube-api-access-zb885" (OuterVolumeSpecName: "kube-api-access-zb885") pod "bcbe26f4-fcd9-46c7-aa3a-5038b801741d" (UID: "bcbe26f4-fcd9-46c7-aa3a-5038b801741d"). InnerVolumeSpecName "kube-api-access-zb885". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.578150 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.578508 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-combined-ca-bundle\") pod \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.578651 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a733d85-ad49-41b8-a75b-842dce56e85c-logs\") pod \"2a733d85-ad49-41b8-a75b-842dce56e85c\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.578774 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhpd4\" (UniqueName: \"kubernetes.io/projected/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-kube-api-access-vhpd4\") pod \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.578889 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-logs\") pod \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.579028 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-public-tls-certs\") pod \"2a733d85-ad49-41b8-a75b-842dce56e85c\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.579425 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-internal-tls-certs\") pod \"2a733d85-ad49-41b8-a75b-842dce56e85c\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.579522 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-scripts\") pod \"2a733d85-ad49-41b8-a75b-842dce56e85c\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.579634 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-config-data\") pod \"2a733d85-ad49-41b8-a75b-842dce56e85c\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.579724 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-combined-ca-bundle\") pod \"2a733d85-ad49-41b8-a75b-842dce56e85c\" (UID: \"2a733d85-ad49-41b8-a75b-842dce56e85c\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.579803 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-config-data\") pod \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.579890 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-config-data-custom\") pod \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\" (UID: \"13fc4b26-4375-4ea2-8eaa-2417d64b3cd1\") " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.578190 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "67d9e01d-5189-4aac-8f1a-b1f09fe138b8" (UID: "67d9e01d-5189-4aac-8f1a-b1f09fe138b8"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.579275 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a733d85-ad49-41b8-a75b-842dce56e85c-logs" (OuterVolumeSpecName: "logs") pod "2a733d85-ad49-41b8-a75b-842dce56e85c" (UID: "2a733d85-ad49-41b8-a75b-842dce56e85c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.579685 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-logs" (OuterVolumeSpecName: "logs") pod "13fc4b26-4375-4ea2-8eaa-2417d64b3cd1" (UID: "13fc4b26-4375-4ea2-8eaa-2417d64b3cd1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.582985 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-scripts" (OuterVolumeSpecName: "scripts") pod "67d9e01d-5189-4aac-8f1a-b1f09fe138b8" (UID: "67d9e01d-5189-4aac-8f1a-b1f09fe138b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.592031 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a733d85-ad49-41b8-a75b-842dce56e85c-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.592061 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.592071 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.592084 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68490a68-dc32-448f-90cc-c146ba49c1a8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.592095 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.592103 4997 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.592113 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.592124 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2lxh\" (UniqueName: \"kubernetes.io/projected/68490a68-dc32-448f-90cc-c146ba49c1a8-kube-api-access-w2lxh\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.592136 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb885\" (UniqueName: \"kubernetes.io/projected/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-kube-api-access-zb885\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.592144 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m4pw\" (UniqueName: \"kubernetes.io/projected/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-kube-api-access-6m4pw\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.592155 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5m8mj\" (UniqueName: \"kubernetes.io/projected/5afeb230-4a35-4872-bffa-c882fe927643-kube-api-access-5m8mj\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.592164 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5afeb230-4a35-4872-bffa-c882fe927643-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.592198 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.599749 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a733d85-ad49-41b8-a75b-842dce56e85c-kube-api-access-vkptj" (OuterVolumeSpecName: "kube-api-access-vkptj") pod "2a733d85-ad49-41b8-a75b-842dce56e85c" (UID: "2a733d85-ad49-41b8-a75b-842dce56e85c"). InnerVolumeSpecName "kube-api-access-vkptj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.615710 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.645664 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-kube-api-access-vhpd4" (OuterVolumeSpecName: "kube-api-access-vhpd4") pod "13fc4b26-4375-4ea2-8eaa-2417d64b3cd1" (UID: "13fc4b26-4375-4ea2-8eaa-2417d64b3cd1"). InnerVolumeSpecName "kube-api-access-vhpd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.647346 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "13fc4b26-4375-4ea2-8eaa-2417d64b3cd1" (UID: "13fc4b26-4375-4ea2-8eaa-2417d64b3cd1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.658602 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-scripts" (OuterVolumeSpecName: "scripts") pod "2a733d85-ad49-41b8-a75b-842dce56e85c" (UID: "2a733d85-ad49-41b8-a75b-842dce56e85c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.677383 4997 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.703615 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 05 07:23:30 crc kubenswrapper[4997]: E1205 07:23:30.710632 4997 configmap.go:193] Couldn't get configMap openstack/ovncontroller-scripts: configmap "ovncontroller-scripts" not found Dec 05 07:23:30 crc kubenswrapper[4997]: E1205 07:23:30.710772 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-scripts podName:5e4237cc-8581-4ddb-b472-e1c6a4a41cb5 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:38.710741763 +0000 UTC m=+1719.239649204 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/configmap/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-scripts") pod "ovn-controller-vnncj" (UID: "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5") : configmap "ovncontroller-scripts" not found Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.714510 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhpd4\" (UniqueName: \"kubernetes.io/projected/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-kube-api-access-vhpd4\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.714546 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.714560 4997 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.714578 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkptj\" (UniqueName: \"kubernetes.io/projected/2a733d85-ad49-41b8-a75b-842dce56e85c-kube-api-access-vkptj\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.714590 4997 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.743592 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.752691 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.774920 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.800803 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13fc4b26-4375-4ea2-8eaa-2417d64b3cd1" (UID: "13fc4b26-4375-4ea2-8eaa-2417d64b3cd1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.801474 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-config-data" (OuterVolumeSpecName: "config-data") pod "5afeb230-4a35-4872-bffa-c882fe927643" (UID: "5afeb230-4a35-4872-bffa-c882fe927643"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.808568 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="0f158874-152b-46ce-ac42-d202684853ca" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.816509 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.816552 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.833175 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67d9e01d-5189-4aac-8f1a-b1f09fe138b8" (UID: "67d9e01d-5189-4aac-8f1a-b1f09fe138b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.833410 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-config-data" (OuterVolumeSpecName: "config-data") pod "67d9e01d-5189-4aac-8f1a-b1f09fe138b8" (UID: "67d9e01d-5189-4aac-8f1a-b1f09fe138b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.836400 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-66698d9c4f-5rmpv"] Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.845060 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-66698d9c4f-5rmpv"] Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.877321 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a733d85-ad49-41b8-a75b-842dce56e85c" (UID: "2a733d85-ad49-41b8-a75b-842dce56e85c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.881543 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5afeb230-4a35-4872-bffa-c882fe927643" (UID: "5afeb230-4a35-4872-bffa-c882fe927643"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.892053 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-config-data" (OuterVolumeSpecName: "config-data") pod "bcbe26f4-fcd9-46c7-aa3a-5038b801741d" (UID: "bcbe26f4-fcd9-46c7-aa3a-5038b801741d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.922702 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.922736 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.922747 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.922757 4997 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.922767 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.951645 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5afeb230-4a35-4872-bffa-c882fe927643" (UID: "5afeb230-4a35-4872-bffa-c882fe927643"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.955511 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-config-data" (OuterVolumeSpecName: "config-data") pod "13fc4b26-4375-4ea2-8eaa-2417d64b3cd1" (UID: "13fc4b26-4375-4ea2-8eaa-2417d64b3cd1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:30 crc kubenswrapper[4997]: I1205 07:23:30.992732 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5afeb230-4a35-4872-bffa-c882fe927643" (UID: "5afeb230-4a35-4872-bffa-c882fe927643"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.031032 4997 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.031773 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5afeb230-4a35-4872-bffa-c882fe927643-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.031786 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.065138 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "67d9e01d-5189-4aac-8f1a-b1f09fe138b8" (UID: "67d9e01d-5189-4aac-8f1a-b1f09fe138b8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.084744 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bcbe26f4-fcd9-46c7-aa3a-5038b801741d" (UID: "bcbe26f4-fcd9-46c7-aa3a-5038b801741d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.105807 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2a733d85-ad49-41b8-a75b-842dce56e85c" (UID: "2a733d85-ad49-41b8-a75b-842dce56e85c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.117075 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "bcbe26f4-fcd9-46c7-aa3a-5038b801741d" (UID: "bcbe26f4-fcd9-46c7-aa3a-5038b801741d"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.132767 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-config-data" (OuterVolumeSpecName: "config-data") pod "2a733d85-ad49-41b8-a75b-842dce56e85c" (UID: "2a733d85-ad49-41b8-a75b-842dce56e85c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.154524 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.154673 4997 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.154736 4997 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67d9e01d-5189-4aac-8f1a-b1f09fe138b8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.154792 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.154846 4997 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcbe26f4-fcd9-46c7-aa3a-5038b801741d-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.193689 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.211113 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2a733d85-ad49-41b8-a75b-842dce56e85c" (UID: "2a733d85-ad49-41b8-a75b-842dce56e85c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.235712 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.235812 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-66698d9c4f-5rmpv" podUID="27ff3957-ee08-40be-a41d-02979f192fda" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.166:8080/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.235834 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-66698d9c4f-5rmpv" podUID="27ff3957-ee08-40be-a41d-02979f192fda" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.166:8080/healthcheck\": dial tcp 10.217.0.166:8080: i/o timeout" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.241009 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.266961 4997 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a733d85-ad49-41b8-a75b-842dce56e85c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: E1205 07:23:31.311584 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-sx4dp operator-scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/keystone3eec-account-delete-v5k7n" podUID="c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.352122 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.353491 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.353517 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.365966 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.400197 4997 scope.go:117] "RemoveContainer" containerID="5f2ebfbb61e660700f2975a4f58a9ce986878add969191a622893e2fcd4fd58c" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.402870 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-scripts\") pod \"e08a39e3-f267-496d-80b3-b12a9eef14c1\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.402957 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-state-metrics-tls-config\") pod \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\" (UID: \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.403005 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-internal-tls-certs\") pod \"e08a39e3-f267-496d-80b3-b12a9eef14c1\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.403061 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-internal-tls-certs\") pod \"2179d060-8f4c-413d-a202-e31ebc242dfc\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.403140 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c706bba7-965c-497f-ae7d-b7087d37d70e-config-data\") pod \"c706bba7-965c-497f-ae7d-b7087d37d70e\" (UID: \"c706bba7-965c-497f-ae7d-b7087d37d70e\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.403174 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-combined-ca-bundle\") pod \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\" (UID: \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.403221 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gn2g\" (UniqueName: \"kubernetes.io/projected/a6bfb8b2-df28-4992-8831-2a5060323ddf-kube-api-access-4gn2g\") pod \"a6bfb8b2-df28-4992-8831-2a5060323ddf\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.403269 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-public-tls-certs\") pod \"2179d060-8f4c-413d-a202-e31ebc242dfc\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.403315 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ht6m\" (UniqueName: \"kubernetes.io/projected/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-api-access-7ht6m\") pod \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\" (UID: \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.403363 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-combined-ca-bundle\") pod \"2179d060-8f4c-413d-a202-e31ebc242dfc\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.403407 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e08a39e3-f267-496d-80b3-b12a9eef14c1-httpd-run\") pod \"e08a39e3-f267-496d-80b3-b12a9eef14c1\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.403437 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c706bba7-965c-497f-ae7d-b7087d37d70e-combined-ca-bundle\") pod \"c706bba7-965c-497f-ae7d-b7087d37d70e\" (UID: \"c706bba7-965c-497f-ae7d-b7087d37d70e\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.403470 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-config-data-custom\") pod \"2179d060-8f4c-413d-a202-e31ebc242dfc\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.403535 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-state-metrics-tls-certs\") pod \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\" (UID: \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.403574 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kpvt\" (UniqueName: \"kubernetes.io/projected/e08a39e3-f267-496d-80b3-b12a9eef14c1-kube-api-access-5kpvt\") pod \"e08a39e3-f267-496d-80b3-b12a9eef14c1\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.411154 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e08a39e3-f267-496d-80b3-b12a9eef14c1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e08a39e3-f267-496d-80b3-b12a9eef14c1" (UID: "e08a39e3-f267-496d-80b3-b12a9eef14c1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.428170 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-combined-ca-bundle\") pod \"e08a39e3-f267-496d-80b3-b12a9eef14c1\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.429409 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ddxf\" (UniqueName: \"kubernetes.io/projected/2179d060-8f4c-413d-a202-e31ebc242dfc-kube-api-access-9ddxf\") pod \"2179d060-8f4c-413d-a202-e31ebc242dfc\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.429561 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"e08a39e3-f267-496d-80b3-b12a9eef14c1\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.429711 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-config-data\") pod \"e08a39e3-f267-496d-80b3-b12a9eef14c1\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.433710 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-929vr\" (UniqueName: \"kubernetes.io/projected/c706bba7-965c-497f-ae7d-b7087d37d70e-kube-api-access-929vr\") pod \"c706bba7-965c-497f-ae7d-b7087d37d70e\" (UID: \"c706bba7-965c-497f-ae7d-b7087d37d70e\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.433832 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2179d060-8f4c-413d-a202-e31ebc242dfc-logs\") pod \"2179d060-8f4c-413d-a202-e31ebc242dfc\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.433940 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-combined-ca-bundle\") pod \"a6bfb8b2-df28-4992-8831-2a5060323ddf\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.434051 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-config-data\") pod \"a6bfb8b2-df28-4992-8831-2a5060323ddf\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.434149 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-config-data\") pod \"2179d060-8f4c-413d-a202-e31ebc242dfc\" (UID: \"2179d060-8f4c-413d-a202-e31ebc242dfc\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.434234 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6bfb8b2-df28-4992-8831-2a5060323ddf-logs\") pod \"a6bfb8b2-df28-4992-8831-2a5060323ddf\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.434351 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e08a39e3-f267-496d-80b3-b12a9eef14c1-logs\") pod \"e08a39e3-f267-496d-80b3-b12a9eef14c1\" (UID: \"e08a39e3-f267-496d-80b3-b12a9eef14c1\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.434442 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-config-data-custom\") pod \"a6bfb8b2-df28-4992-8831-2a5060323ddf\" (UID: \"a6bfb8b2-df28-4992-8831-2a5060323ddf\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.435973 4997 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e08a39e3-f267-496d-80b3-b12a9eef14c1-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.447292 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell02ec3-account-delete-8p9cc" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.428538 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2179d060-8f4c-413d-a202-e31ebc242dfc" (UID: "2179d060-8f4c-413d-a202-e31ebc242dfc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.515463 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6bfb8b2-df28-4992-8831-2a5060323ddf-logs" (OuterVolumeSpecName: "logs") pod "a6bfb8b2-df28-4992-8831-2a5060323ddf" (UID: "a6bfb8b2-df28-4992-8831-2a5060323ddf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.515822 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e08a39e3-f267-496d-80b3-b12a9eef14c1-logs" (OuterVolumeSpecName: "logs") pod "e08a39e3-f267-496d-80b3-b12a9eef14c1" (UID: "e08a39e3-f267-496d-80b3-b12a9eef14c1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.447814 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e08a39e3-f267-496d-80b3-b12a9eef14c1-kube-api-access-5kpvt" (OuterVolumeSpecName: "kube-api-access-5kpvt") pod "e08a39e3-f267-496d-80b3-b12a9eef14c1" (UID: "e08a39e3-f267-496d-80b3-b12a9eef14c1"). InnerVolumeSpecName "kube-api-access-5kpvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.454029 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-scripts" (OuterVolumeSpecName: "scripts") pod "e08a39e3-f267-496d-80b3-b12a9eef14c1" (UID: "e08a39e3-f267-496d-80b3-b12a9eef14c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.469694 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2179d060-8f4c-413d-a202-e31ebc242dfc-logs" (OuterVolumeSpecName: "logs") pod "2179d060-8f4c-413d-a202-e31ebc242dfc" (UID: "2179d060-8f4c-413d-a202-e31ebc242dfc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.531758 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-api-access-7ht6m" (OuterVolumeSpecName: "kube-api-access-7ht6m") pod "f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946" (UID: "f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946"). InnerVolumeSpecName "kube-api-access-7ht6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.532316 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c706bba7-965c-497f-ae7d-b7087d37d70e-kube-api-access-929vr" (OuterVolumeSpecName: "kube-api-access-929vr") pod "c706bba7-965c-497f-ae7d-b7087d37d70e" (UID: "c706bba7-965c-497f-ae7d-b7087d37d70e"). InnerVolumeSpecName "kube-api-access-929vr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.539241 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6bfb8b2-df28-4992-8831-2a5060323ddf-kube-api-access-4gn2g" (OuterVolumeSpecName: "kube-api-access-4gn2g") pod "a6bfb8b2-df28-4992-8831-2a5060323ddf" (UID: "a6bfb8b2-df28-4992-8831-2a5060323ddf"). InnerVolumeSpecName "kube-api-access-4gn2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.573035 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf72e65a-bae4-4ce4-b65a-d5a587ef5f48-operator-scripts\") pod \"cf72e65a-bae4-4ce4-b65a-d5a587ef5f48\" (UID: \"cf72e65a-bae4-4ce4-b65a-d5a587ef5f48\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.573097 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tklqq\" (UniqueName: \"kubernetes.io/projected/cf72e65a-bae4-4ce4-b65a-d5a587ef5f48-kube-api-access-tklqq\") pod \"cf72e65a-bae4-4ce4-b65a-d5a587ef5f48\" (UID: \"cf72e65a-bae4-4ce4-b65a-d5a587ef5f48\") " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.573848 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "e08a39e3-f267-496d-80b3-b12a9eef14c1" (UID: "e08a39e3-f267-496d-80b3-b12a9eef14c1"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.573918 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c706bba7-965c-497f-ae7d-b7087d37d70e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c706bba7-965c-497f-ae7d-b7087d37d70e" (UID: "c706bba7-965c-497f-ae7d-b7087d37d70e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.574203 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2179d060-8f4c-413d-a202-e31ebc242dfc-kube-api-access-9ddxf" (OuterVolumeSpecName: "kube-api-access-9ddxf") pod "2179d060-8f4c-413d-a202-e31ebc242dfc" (UID: "2179d060-8f4c-413d-a202-e31ebc242dfc"). InnerVolumeSpecName "kube-api-access-9ddxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.573773 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx4dp\" (UniqueName: \"kubernetes.io/projected/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-kube-api-access-sx4dp\") pod \"keystone3eec-account-delete-v5k7n\" (UID: \"c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1\") " pod="openstack/keystone3eec-account-delete-v5k7n" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.574486 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-operator-scripts\") pod \"keystone3eec-account-delete-v5k7n\" (UID: \"c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1\") " pod="openstack/keystone3eec-account-delete-v5k7n" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.574326 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a6bfb8b2-df28-4992-8831-2a5060323ddf" (UID: "a6bfb8b2-df28-4992-8831-2a5060323ddf"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: E1205 07:23:31.579244 4997 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 07:23:31 crc kubenswrapper[4997]: E1205 07:23:31.579360 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-operator-scripts podName:c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:33.579334712 +0000 UTC m=+1714.108241993 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-operator-scripts") pod "keystone3eec-account-delete-v5k7n" (UID: "c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1") : configmap "openstack-scripts" not found Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.580850 4997 generic.go:334] "Generic (PLEG): container finished" podID="db136058-17a8-4955-8e95-576da7b9e847" containerID="07f1631824a73521a5df4d80b3bf4feb8dde04c7a5b2c5e5688b95c1efabb84c" exitCode=0 Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.580927 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db136058-17a8-4955-8e95-576da7b9e847","Type":"ContainerDied","Data":"07f1631824a73521a5df4d80b3bf4feb8dde04c7a5b2c5e5688b95c1efabb84c"} Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.582950 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance7904-account-delete-v5xm8" event={"ID":"8f2677aa-a856-47de-8417-5bc3d08d6ff1","Type":"ContainerStarted","Data":"67982301894febce687ec7bf69c55cab5ca1970394fb3e3692c4485a0582df66"} Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.584084 4997 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/glance7904-account-delete-v5xm8" secret="" err="secret \"galera-openstack-dockercfg-ct8h9\" not found" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.585890 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gn2g\" (UniqueName: \"kubernetes.io/projected/a6bfb8b2-df28-4992-8831-2a5060323ddf-kube-api-access-4gn2g\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.585941 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ht6m\" (UniqueName: \"kubernetes.io/projected/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-api-access-7ht6m\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.585972 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c706bba7-965c-497f-ae7d-b7087d37d70e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.585986 4997 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.585999 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kpvt\" (UniqueName: \"kubernetes.io/projected/e08a39e3-f267-496d-80b3-b12a9eef14c1-kube-api-access-5kpvt\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.586011 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ddxf\" (UniqueName: \"kubernetes.io/projected/2179d060-8f4c-413d-a202-e31ebc242dfc-kube-api-access-9ddxf\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.586065 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.586081 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-929vr\" (UniqueName: \"kubernetes.io/projected/c706bba7-965c-497f-ae7d-b7087d37d70e-kube-api-access-929vr\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.586093 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2179d060-8f4c-413d-a202-e31ebc242dfc-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.586114 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6bfb8b2-df28-4992-8831-2a5060323ddf-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.586125 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e08a39e3-f267-496d-80b3-b12a9eef14c1-logs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.586137 4997 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.586152 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.592135 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf72e65a-bae4-4ce4-b65a-d5a587ef5f48-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cf72e65a-bae4-4ce4-b65a-d5a587ef5f48" (UID: "cf72e65a-bae4-4ce4-b65a-d5a587ef5f48"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: E1205 07:23:31.609919 4997 projected.go:194] Error preparing data for projected volume kube-api-access-sx4dp for pod openstack/keystone3eec-account-delete-v5k7n: failed to fetch token: serviceaccounts "galera-openstack" not found Dec 05 07:23:31 crc kubenswrapper[4997]: E1205 07:23:31.609995 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-kube-api-access-sx4dp podName:c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:33.609973313 +0000 UTC m=+1714.138880574 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-sx4dp" (UniqueName: "kubernetes.io/projected/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-kube-api-access-sx4dp") pod "keystone3eec-account-delete-v5k7n" (UID: "c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1") : failed to fetch token: serviceaccounts "galera-openstack" not found Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.610230 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell02ec3-account-delete-8p9cc" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.610289 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell02ec3-account-delete-8p9cc" event={"ID":"cf72e65a-bae4-4ce4-b65a-d5a587ef5f48","Type":"ContainerDied","Data":"7deda0533d064e3d0aa9a66967efbd9e7313bf31bbe6f07556e6ed3b65bf764e"} Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.610353 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7deda0533d064e3d0aa9a66967efbd9e7313bf31bbe6f07556e6ed3b65bf764e" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.636147 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946","Type":"ContainerDied","Data":"f3a87fd0301e468e3f56ff45dc8706a15bba67091a215495595cb39473aa2e53"} Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.636309 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.639306 4997 generic.go:334] "Generic (PLEG): container finished" podID="1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" containerID="bd15ae01a5d3986a7e4425b35e0fa45dbfda1aea5cdf404c87ede67fab61b69c" exitCode=0 Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.639358 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79","Type":"ContainerDied","Data":"bd15ae01a5d3986a7e4425b35e0fa45dbfda1aea5cdf404c87ede67fab61b69c"} Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.641959 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf72e65a-bae4-4ce4-b65a-d5a587ef5f48-kube-api-access-tklqq" (OuterVolumeSpecName: "kube-api-access-tklqq") pod "cf72e65a-bae4-4ce4-b65a-d5a587ef5f48" (UID: "cf72e65a-bae4-4ce4-b65a-d5a587ef5f48"). InnerVolumeSpecName "kube-api-access-tklqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.648198 4997 generic.go:334] "Generic (PLEG): container finished" podID="0f158874-152b-46ce-ac42-d202684853ca" containerID="ad9f0fd9866254fcb3d271814a479f3ce3ebdaecf8def2eb7a15d7c6654387e1" exitCode=0 Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.648296 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0f158874-152b-46ce-ac42-d202684853ca","Type":"ContainerDied","Data":"ad9f0fd9866254fcb3d271814a479f3ce3ebdaecf8def2eb7a15d7c6654387e1"} Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.670062 4997 generic.go:334] "Generic (PLEG): container finished" podID="a5edac3e-6bee-494f-97ba-c4047064d02f" containerID="0793ea0018a2ecff7f4d4bdde243bfa95c8677f257b5ca9cd61b32b43ef5ef43" exitCode=1 Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.670267 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-555c8b7b77-44qw4" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.671598 4997 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapi8ba6-account-delete-t8gw7" secret="" err="secret \"galera-openstack-dockercfg-ct8h9\" not found" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.671827 4997 scope.go:117] "RemoveContainer" containerID="0793ea0018a2ecff7f4d4bdde243bfa95c8677f257b5ca9cd61b32b43ef5ef43" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.671830 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.672403 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.672593 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.675760 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi8ba6-account-delete-t8gw7" event={"ID":"a5edac3e-6bee-494f-97ba-c4047064d02f","Type":"ContainerDied","Data":"0793ea0018a2ecff7f4d4bdde243bfa95c8677f257b5ca9cd61b32b43ef5ef43"} Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.675892 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.678071 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5cb79f8cdd-zkk89" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.678148 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.678187 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone3eec-account-delete-v5k7n" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.679617 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-db8868644-mchk9" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.680016 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-54ff4c8496-8z64m" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.680130 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron84af-account-delete-pgl29" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.683193 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946" (UID: "f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.691056 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf72e65a-bae4-4ce4-b65a-d5a587ef5f48-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.691082 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tklqq\" (UniqueName: \"kubernetes.io/projected/cf72e65a-bae4-4ce4-b65a-d5a587ef5f48-kube-api-access-tklqq\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.691097 4997 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: E1205 07:23:31.691871 4997 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 07:23:31 crc kubenswrapper[4997]: E1205 07:23:31.691938 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts podName:8f2677aa-a856-47de-8417-5bc3d08d6ff1 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:32.191916035 +0000 UTC m=+1712.720823306 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts") pod "glance7904-account-delete-v5xm8" (UID: "8f2677aa-a856-47de-8417-5bc3d08d6ff1") : configmap "openstack-scripts" not found Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.699447 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance7904-account-delete-v5xm8" podStartSLOduration=9.699428508 podStartE2EDuration="9.699428508s" podCreationTimestamp="2025-12-05 07:23:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 07:23:31.630312565 +0000 UTC m=+1712.159219826" watchObservedRunningTime="2025-12-05 07:23:31.699428508 +0000 UTC m=+1712.228335769" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.737353 4997 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.778695 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946" (UID: "f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.792008 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946" (UID: "f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.792295 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0abba971-b935-4cb0-865e-b364f8521f81" path="/var/lib/kubelet/pods/0abba971-b935-4cb0-865e-b364f8521f81/volumes" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.793273 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d3915c5-30d2-43be-abd4-438cd9d8ebf7" path="/var/lib/kubelet/pods/0d3915c5-30d2-43be-abd4-438cd9d8ebf7/volumes" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.796376 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27ff3957-ee08-40be-a41d-02979f192fda" path="/var/lib/kubelet/pods/27ff3957-ee08-40be-a41d-02979f192fda/volumes" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.797497 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-combined-ca-bundle\") pod \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\" (UID: \"f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946\") " Dec 05 07:23:31 crc kubenswrapper[4997]: W1205 07:23:31.799675 4997 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946/volumes/kubernetes.io~secret/combined-ca-bundle Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.804699 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946" (UID: "f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.804747 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c706bba7-965c-497f-ae7d-b7087d37d70e-config-data" (OuterVolumeSpecName: "config-data") pod "c706bba7-965c-497f-ae7d-b7087d37d70e" (UID: "c706bba7-965c-497f-ae7d-b7087d37d70e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.806206 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c706bba7-965c-497f-ae7d-b7087d37d70e-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.806241 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.806256 4997 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.806267 4997 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: E1205 07:23:31.806360 4997 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 07:23:31 crc kubenswrapper[4997]: E1205 07:23:31.806429 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a5edac3e-6bee-494f-97ba-c4047064d02f-operator-scripts podName:a5edac3e-6bee-494f-97ba-c4047064d02f nodeName:}" failed. No retries permitted until 2025-12-05 07:23:32.306408828 +0000 UTC m=+1712.835316089 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/a5edac3e-6bee-494f-97ba-c4047064d02f-operator-scripts") pod "novaapi8ba6-account-delete-t8gw7" (UID: "a5edac3e-6bee-494f-97ba-c4047064d02f") : configmap "openstack-scripts" not found Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.808131 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="377ed320-c385-42a5-852e-31360c519c6d" path="/var/lib/kubelet/pods/377ed320-c385-42a5-852e-31360c519c6d/volumes" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.808794 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77cf9c93-8871-4190-8522-3cb69ede54dd" path="/var/lib/kubelet/pods/77cf9c93-8871-4190-8522-3cb69ede54dd/volumes" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.809528 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b" path="/var/lib/kubelet/pods/7cc1dcb3-5e6e-4b86-ac5d-e97b1818932b/volumes" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.814384 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8374cc67-e735-47f0-b310-6bebd608cece" path="/var/lib/kubelet/pods/8374cc67-e735-47f0-b310-6bebd608cece/volumes" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.815886 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed" path="/var/lib/kubelet/pods/9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed/volumes" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.816902 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe32fb42-61e6-4cd4-a75b-b684a4171224" path="/var/lib/kubelet/pods/fe32fb42-61e6-4cd4-a75b-b684a4171224/volumes" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.831780 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-config-data" (OuterVolumeSpecName: "config-data") pod "a6bfb8b2-df28-4992-8831-2a5060323ddf" (UID: "a6bfb8b2-df28-4992-8831-2a5060323ddf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.853379 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2179d060-8f4c-413d-a202-e31ebc242dfc" (UID: "2179d060-8f4c-413d-a202-e31ebc242dfc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.857434 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone3eec-account-delete-v5k7n" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.864237 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e08a39e3-f267-496d-80b3-b12a9eef14c1" (UID: "e08a39e3-f267-496d-80b3-b12a9eef14c1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.880401 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e08a39e3-f267-496d-80b3-b12a9eef14c1" (UID: "e08a39e3-f267-496d-80b3-b12a9eef14c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.880755 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a6bfb8b2-df28-4992-8831-2a5060323ddf" (UID: "a6bfb8b2-df28-4992-8831-2a5060323ddf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.889078 4997 scope.go:117] "RemoveContainer" containerID="175b0ae7934dc4fc48a3e2c7faa4ea18aa39089d2851c19371ca3e1b197fb37a" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.911265 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.911313 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.911326 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.911337 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6bfb8b2-df28-4992-8831-2a5060323ddf-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.911348 4997 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.965376 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-config-data" (OuterVolumeSpecName: "config-data") pod "2179d060-8f4c-413d-a202-e31ebc242dfc" (UID: "2179d060-8f4c-413d-a202-e31ebc242dfc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.976823 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2179d060-8f4c-413d-a202-e31ebc242dfc" (UID: "2179d060-8f4c-413d-a202-e31ebc242dfc"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.987338 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2179d060-8f4c-413d-a202-e31ebc242dfc" (UID: "2179d060-8f4c-413d-a202-e31ebc242dfc"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:31 crc kubenswrapper[4997]: I1205 07:23:31.995670 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-config-data" (OuterVolumeSpecName: "config-data") pod "e08a39e3-f267-496d-80b3-b12a9eef14c1" (UID: "e08a39e3-f267-496d-80b3-b12a9eef14c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.013483 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e08a39e3-f267-496d-80b3-b12a9eef14c1-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.013524 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.013534 4997 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.013545 4997 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2179d060-8f4c-413d-a202-e31ebc242dfc-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.193968 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-9cfz7"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.194010 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-9cfz7"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.194027 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-84af-account-create-update-7hbg6"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.194040 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron84af-account-delete-pgl29"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.194057 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-84af-account-create-update-7hbg6"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.194071 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron84af-account-delete-pgl29"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.194084 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-vj5mk"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.194095 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-vj5mk"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.194106 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-a458-account-create-update-7cfrj"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.194116 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cindera458-account-delete-vpgnx"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.194130 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-a458-account-create-update-7cfrj"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.194141 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-zmvrc"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.202260 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-zmvrc"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.217058 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-69df-account-create-update-2w6f2"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.222399 4997 scope.go:117] "RemoveContainer" containerID="8f36673252ccfc4fc6b7d58b342de4054e621d8bc6722f917b4cf09a6dd9bcaf" Dec 05 07:23:32 crc kubenswrapper[4997]: E1205 07:23:32.228236 4997 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 07:23:32 crc kubenswrapper[4997]: E1205 07:23:32.228313 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts podName:8f2677aa-a856-47de-8417-5bc3d08d6ff1 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:33.228295887 +0000 UTC m=+1713.757203148 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts") pod "glance7904-account-delete-v5xm8" (UID: "8f2677aa-a856-47de-8417-5bc3d08d6ff1") : configmap "openstack-scripts" not found Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.271539 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican69df-account-delete-tqlsz"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.279191 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-69df-account-create-update-2w6f2"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.294815 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-6t8lv"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.307816 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-6t8lv"] Dec 05 07:23:32 crc kubenswrapper[4997]: E1205 07:23:32.330058 4997 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 07:23:32 crc kubenswrapper[4997]: E1205 07:23:32.330148 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a5edac3e-6bee-494f-97ba-c4047064d02f-operator-scripts podName:a5edac3e-6bee-494f-97ba-c4047064d02f nodeName:}" failed. No retries permitted until 2025-12-05 07:23:33.330124538 +0000 UTC m=+1713.859031799 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/a5edac3e-6bee-494f-97ba-c4047064d02f-operator-scripts") pod "novaapi8ba6-account-delete-t8gw7" (UID: "a5edac3e-6bee-494f-97ba-c4047064d02f") : configmap "openstack-scripts" not found Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.336728 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement94df-account-delete-4qxc6"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.372078 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-94df-account-create-update-gsv4j"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.398504 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-94df-account-create-update-gsv4j"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.626919 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cindera458-account-delete-vpgnx" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.695725 4997 scope.go:117] "RemoveContainer" containerID="057a82bd61c9e67d2a9d3ed1d85e6973ab9ffca83e3bd81b232a535bb4ad1866" Dec 05 07:23:32 crc kubenswrapper[4997]: E1205 07:23:32.696148 4997 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Dec 05 07:23:32 crc kubenswrapper[4997]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-12-05T07:23:25Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 05 07:23:32 crc kubenswrapper[4997]: /etc/init.d/functions: line 589: 449 Alarm clock "$@" Dec 05 07:23:32 crc kubenswrapper[4997]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-vnncj" message=< Dec 05 07:23:32 crc kubenswrapper[4997]: Exiting ovn-controller (1) [FAILED] Dec 05 07:23:32 crc kubenswrapper[4997]: Killing ovn-controller (1) [ OK ] Dec 05 07:23:32 crc kubenswrapper[4997]: Killing ovn-controller (1) with SIGKILL [ OK ] Dec 05 07:23:32 crc kubenswrapper[4997]: 2025-12-05T07:23:25Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 05 07:23:32 crc kubenswrapper[4997]: /etc/init.d/functions: line 589: 449 Alarm clock "$@" Dec 05 07:23:32 crc kubenswrapper[4997]: > Dec 05 07:23:32 crc kubenswrapper[4997]: E1205 07:23:32.696199 4997 kuberuntime_container.go:691] "PreStop hook failed" err=< Dec 05 07:23:32 crc kubenswrapper[4997]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-12-05T07:23:25Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 05 07:23:32 crc kubenswrapper[4997]: /etc/init.d/functions: line 589: 449 Alarm clock "$@" Dec 05 07:23:32 crc kubenswrapper[4997]: > pod="openstack/ovn-controller-vnncj" podUID="5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" containerName="ovn-controller" containerID="cri-o://f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.696251 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-vnncj" podUID="5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" containerName="ovn-controller" containerID="cri-o://f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38" gracePeriod=22 Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.720059 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican69df-account-delete-tqlsz" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.759867 4997 generic.go:334] "Generic (PLEG): container finished" podID="a5edac3e-6bee-494f-97ba-c4047064d02f" containerID="f81fa8eed1dfdce11148625c1196e1d19831afebfa221d3caa7686de91ae7b79" exitCode=1 Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.760048 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi8ba6-account-delete-t8gw7" event={"ID":"a5edac3e-6bee-494f-97ba-c4047064d02f","Type":"ContainerDied","Data":"f81fa8eed1dfdce11148625c1196e1d19831afebfa221d3caa7686de91ae7b79"} Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.761096 4997 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapi8ba6-account-delete-t8gw7" secret="" err="secret \"galera-openstack-dockercfg-ct8h9\" not found" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.761148 4997 scope.go:117] "RemoveContainer" containerID="f81fa8eed1dfdce11148625c1196e1d19831afebfa221d3caa7686de91ae7b79" Dec 05 07:23:32 crc kubenswrapper[4997]: E1205 07:23:32.761606 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=novaapi8ba6-account-delete-t8gw7_openstack(a5edac3e-6bee-494f-97ba-c4047064d02f)\"" pod="openstack/novaapi8ba6-account-delete-t8gw7" podUID="a5edac3e-6bee-494f-97ba-c4047064d02f" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.775289 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79","Type":"ContainerDied","Data":"27a0d3d6b8ab687d87ab58840a512f00fcdb53823f791cfa7ba9a367a6e552ad"} Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.775340 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27a0d3d6b8ab687d87ab58840a512f00fcdb53823f791cfa7ba9a367a6e552ad" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.782271 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-pspnv"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.784788 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cindera458-account-delete-vpgnx" event={"ID":"b8c5aa81-a666-4ae6-a41c-fd31df260223","Type":"ContainerDied","Data":"0afdb95173171611a53b5d9cfa50ee71774a0d899adb7dedf3b9e5e5b5a7657a"} Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.784840 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0afdb95173171611a53b5d9cfa50ee71774a0d899adb7dedf3b9e5e5b5a7657a" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.789935 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cindera458-account-delete-vpgnx" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.802454 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-pspnv"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.823044 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0f158874-152b-46ce-ac42-d202684853ca","Type":"ContainerDied","Data":"9795e2ed191f04d77c06fc39017db89322459c1559b11a047ed1fee44b59593e"} Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.823099 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9795e2ed191f04d77c06fc39017db89322459c1559b11a047ed1fee44b59593e" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.836399 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-2ec3-account-create-update-w5xrv"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.849197 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8c5aa81-a666-4ae6-a41c-fd31df260223-operator-scripts\") pod \"b8c5aa81-a666-4ae6-a41c-fd31df260223\" (UID: \"b8c5aa81-a666-4ae6-a41c-fd31df260223\") " Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.849374 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d65cd492-d4bd-4489-9d28-a6c42407101a-operator-scripts\") pod \"d65cd492-d4bd-4489-9d28-a6c42407101a\" (UID: \"d65cd492-d4bd-4489-9d28-a6c42407101a\") " Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.849501 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xm6mx\" (UniqueName: \"kubernetes.io/projected/b8c5aa81-a666-4ae6-a41c-fd31df260223-kube-api-access-xm6mx\") pod \"b8c5aa81-a666-4ae6-a41c-fd31df260223\" (UID: \"b8c5aa81-a666-4ae6-a41c-fd31df260223\") " Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.849579 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nb7s\" (UniqueName: \"kubernetes.io/projected/d65cd492-d4bd-4489-9d28-a6c42407101a-kube-api-access-9nb7s\") pod \"d65cd492-d4bd-4489-9d28-a6c42407101a\" (UID: \"d65cd492-d4bd-4489-9d28-a6c42407101a\") " Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.850820 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8c5aa81-a666-4ae6-a41c-fd31df260223-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b8c5aa81-a666-4ae6-a41c-fd31df260223" (UID: "b8c5aa81-a666-4ae6-a41c-fd31df260223"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.851863 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d65cd492-d4bd-4489-9d28-a6c42407101a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d65cd492-d4bd-4489-9d28-a6c42407101a" (UID: "d65cd492-d4bd-4489-9d28-a6c42407101a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.852771 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell02ec3-account-delete-8p9cc"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.854894 4997 generic.go:334] "Generic (PLEG): container finished" podID="0601e553-5305-4f54-9a2d-43b43a1a5dd4" containerID="22c17242a1211bc3c93181daebd1e55ec7c7f3dfb0c90a0942488ba51f04fe30" exitCode=0 Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.855317 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0601e553-5305-4f54-9a2d-43b43a1a5dd4","Type":"ContainerDied","Data":"22c17242a1211bc3c93181daebd1e55ec7c7f3dfb0c90a0942488ba51f04fe30"} Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.863880 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d65cd492-d4bd-4489-9d28-a6c42407101a-kube-api-access-9nb7s" (OuterVolumeSpecName: "kube-api-access-9nb7s") pod "d65cd492-d4bd-4489-9d28-a6c42407101a" (UID: "d65cd492-d4bd-4489-9d28-a6c42407101a"). InnerVolumeSpecName "kube-api-access-9nb7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.865378 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8c5aa81-a666-4ae6-a41c-fd31df260223-kube-api-access-xm6mx" (OuterVolumeSpecName: "kube-api-access-xm6mx") pod "b8c5aa81-a666-4ae6-a41c-fd31df260223" (UID: "b8c5aa81-a666-4ae6-a41c-fd31df260223"). InnerVolumeSpecName "kube-api-access-xm6mx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.883814 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-2ec3-account-create-update-w5xrv"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.899920 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement94df-account-delete-4qxc6" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.900677 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement94df-account-delete-4qxc6" event={"ID":"c51acefb-ed8a-480c-9ab6-0345f2ff0cd3","Type":"ContainerDied","Data":"a1221a18388371ad38dd286f6a30f31ac2a036d9c6b7fc40448d3191ce825a9a"} Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.900712 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1221a18388371ad38dd286f6a30f31ac2a036d9c6b7fc40448d3191ce825a9a" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.908641 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.908797 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell02ec3-account-delete-8p9cc"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.912246 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican69df-account-delete-tqlsz" event={"ID":"d65cd492-d4bd-4489-9d28-a6c42407101a","Type":"ContainerDied","Data":"9d06463b05ff535e9822e6bdc25fcff443a779823bd9ed62a7c592c795eacbc9"} Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.912291 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d06463b05ff535e9822e6bdc25fcff443a779823bd9ed62a7c592c795eacbc9" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.912344 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican69df-account-delete-tqlsz" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.951056 4997 scope.go:117] "RemoveContainer" containerID="d9261932ff03a509a886dc6ed5c881f90c33d3fd0c490c206516f3523389c13c" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.954457 4997 generic.go:334] "Generic (PLEG): container finished" podID="ef43a612-a3df-4b50-9bf3-e5d8098cd0fe" containerID="175b217107beb960b8632016072bdd9ad602c7c079fe319f224a856dd6e14c82" exitCode=0 Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.955150 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone3eec-account-delete-v5k7n" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.955315 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-gf4w4"] Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.955371 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe","Type":"ContainerDied","Data":"175b217107beb960b8632016072bdd9ad602c7c079fe319f224a856dd6e14c82"} Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.956496 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.968072 4997 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/glance7904-account-delete-v5xm8" secret="" err="secret \"galera-openstack-dockercfg-ct8h9\" not found" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.981875 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8c5aa81-a666-4ae6-a41c-fd31df260223-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.981915 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d65cd492-d4bd-4489-9d28-a6c42407101a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.981928 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xm6mx\" (UniqueName: \"kubernetes.io/projected/b8c5aa81-a666-4ae6-a41c-fd31df260223-kube-api-access-xm6mx\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:32 crc kubenswrapper[4997]: I1205 07:23:32.981945 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nb7s\" (UniqueName: \"kubernetes.io/projected/d65cd492-d4bd-4489-9d28-a6c42407101a-kube-api-access-9nb7s\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:32.990754 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-gf4w4"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.015316 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.021883 4997 scope.go:117] "RemoveContainer" containerID="0793ea0018a2ecff7f4d4bdde243bfa95c8677f257b5ca9cd61b32b43ef5ef43" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.071802 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.077279 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-8ba6-account-create-update-26l28"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.086845 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ps2f6\" (UniqueName: \"kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-kube-api-access-ps2f6\") pod \"0f158874-152b-46ce-ac42-d202684853ca\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.086976 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-tls\") pod \"0f158874-152b-46ce-ac42-d202684853ca\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087039 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ngbr\" (UniqueName: \"kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-kube-api-access-5ngbr\") pod \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087074 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"0f158874-152b-46ce-ac42-d202684853ca\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087107 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-plugins\") pod \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087158 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087210 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-erlang-cookie\") pod \"0f158874-152b-46ce-ac42-d202684853ca\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087264 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-config-data\") pod \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087289 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6s6q\" (UniqueName: \"kubernetes.io/projected/c51acefb-ed8a-480c-9ab6-0345f2ff0cd3-kube-api-access-q6s6q\") pod \"c51acefb-ed8a-480c-9ab6-0345f2ff0cd3\" (UID: \"c51acefb-ed8a-480c-9ab6-0345f2ff0cd3\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087315 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-plugins-conf\") pod \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087351 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-pod-info\") pod \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087389 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-server-conf\") pod \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087435 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c51acefb-ed8a-480c-9ab6-0345f2ff0cd3-operator-scripts\") pod \"c51acefb-ed8a-480c-9ab6-0345f2ff0cd3\" (UID: \"c51acefb-ed8a-480c-9ab6-0345f2ff0cd3\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087474 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-plugins\") pod \"0f158874-152b-46ce-ac42-d202684853ca\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087506 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-tls\") pod \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087534 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f158874-152b-46ce-ac42-d202684853ca-pod-info\") pod \"0f158874-152b-46ce-ac42-d202684853ca\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087567 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-erlang-cookie\") pod \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087643 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-erlang-cookie-secret\") pod \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087665 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-plugins-conf\") pod \"0f158874-152b-46ce-ac42-d202684853ca\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087717 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f158874-152b-46ce-ac42-d202684853ca-erlang-cookie-secret\") pod \"0f158874-152b-46ce-ac42-d202684853ca\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087750 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-config-data\") pod \"0f158874-152b-46ce-ac42-d202684853ca\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087810 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-server-conf\") pod \"0f158874-152b-46ce-ac42-d202684853ca\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087846 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-confd\") pod \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\" (UID: \"1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.087866 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-confd\") pod \"0f158874-152b-46ce-ac42-d202684853ca\" (UID: \"0f158874-152b-46ce-ac42-d202684853ca\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.088022 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0f158874-152b-46ce-ac42-d202684853ca" (UID: "0f158874-152b-46ce-ac42-d202684853ca"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.088798 4997 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.091732 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0f158874-152b-46ce-ac42-d202684853ca" (UID: "0f158874-152b-46ce-ac42-d202684853ca"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.097172 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0f158874-152b-46ce-ac42-d202684853ca" (UID: "0f158874-152b-46ce-ac42-d202684853ca"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.097554 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" (UID: "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.097611 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi8ba6-account-delete-t8gw7"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.098572 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c51acefb-ed8a-480c-9ab6-0345f2ff0cd3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c51acefb-ed8a-480c-9ab6-0345f2ff0cd3" (UID: "c51acefb-ed8a-480c-9ab6-0345f2ff0cd3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.098742 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" (UID: "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.100134 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-kube-api-access-ps2f6" (OuterVolumeSpecName: "kube-api-access-ps2f6") pod "0f158874-152b-46ce-ac42-d202684853ca" (UID: "0f158874-152b-46ce-ac42-d202684853ca"). InnerVolumeSpecName "kube-api-access-ps2f6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.101228 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" (UID: "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.101288 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-8ba6-account-create-update-26l28"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.113337 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" (UID: "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.114869 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.122282 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0f158874-152b-46ce-ac42-d202684853ca" (UID: "0f158874-152b-46ce-ac42-d202684853ca"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.139282 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0f158874-152b-46ce-ac42-d202684853ca-pod-info" (OuterVolumeSpecName: "pod-info") pod "0f158874-152b-46ce-ac42-d202684853ca" (UID: "0f158874-152b-46ce-ac42-d202684853ca"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.139374 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.169221 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c51acefb-ed8a-480c-9ab6-0345f2ff0cd3-kube-api-access-q6s6q" (OuterVolumeSpecName: "kube-api-access-q6s6q") pod "c51acefb-ed8a-480c-9ab6-0345f2ff0cd3" (UID: "c51acefb-ed8a-480c-9ab6-0345f2ff0cd3"). InnerVolumeSpecName "kube-api-access-q6s6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.169321 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" (UID: "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.169349 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" (UID: "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.172722 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.177623 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-config-data" (OuterVolumeSpecName: "config-data") pod "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" (UID: "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.177936 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f158874-152b-46ce-ac42-d202684853ca-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0f158874-152b-46ce-ac42-d202684853ca" (UID: "0f158874-152b-46ce-ac42-d202684853ca"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.181293 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-kube-api-access-5ngbr" (OuterVolumeSpecName: "kube-api-access-5ngbr") pod "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" (UID: "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79"). InnerVolumeSpecName "kube-api-access-5ngbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.184195 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.189740 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-kolla-config\") pod \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.189926 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hrfb\" (UniqueName: \"kubernetes.io/projected/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-kube-api-access-9hrfb\") pod \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.189969 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-memcached-tls-certs\") pod \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190002 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-kolla-config\") pod \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190026 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0601e553-5305-4f54-9a2d-43b43a1a5dd4-combined-ca-bundle\") pod \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190113 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-config-data-default\") pod \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190159 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-operator-scripts\") pod \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190172 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190193 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jwls\" (UniqueName: \"kubernetes.io/projected/0601e553-5305-4f54-9a2d-43b43a1a5dd4-kube-api-access-6jwls\") pod \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190210 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-combined-ca-bundle\") pod \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190266 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-config-data\") pod \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\" (UID: \"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190300 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0601e553-5305-4f54-9a2d-43b43a1a5dd4-config-data-generated\") pod \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190344 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0601e553-5305-4f54-9a2d-43b43a1a5dd4-galera-tls-certs\") pod \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\" (UID: \"0601e553-5305-4f54-9a2d-43b43a1a5dd4\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190800 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ps2f6\" (UniqueName: \"kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-kube-api-access-ps2f6\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190820 4997 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190832 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ngbr\" (UniqueName: \"kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-kube-api-access-5ngbr\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190842 4997 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190865 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190876 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190885 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6s6q\" (UniqueName: \"kubernetes.io/projected/c51acefb-ed8a-480c-9ab6-0345f2ff0cd3-kube-api-access-q6s6q\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.193517 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "0601e553-5305-4f54-9a2d-43b43a1a5dd4" (UID: "0601e553-5305-4f54-9a2d-43b43a1a5dd4"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.201173 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "0601e553-5305-4f54-9a2d-43b43a1a5dd4" (UID: "0601e553-5305-4f54-9a2d-43b43a1a5dd4"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.201265 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-54ff4c8496-8z64m"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.202029 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-config-data" (OuterVolumeSpecName: "config-data") pod "ef43a612-a3df-4b50-9bf3-e5d8098cd0fe" (UID: "ef43a612-a3df-4b50-9bf3-e5d8098cd0fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.204798 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-pod-info" (OuterVolumeSpecName: "pod-info") pod "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" (UID: "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.205340 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0601e553-5305-4f54-9a2d-43b43a1a5dd4" (UID: "0601e553-5305-4f54-9a2d-43b43a1a5dd4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.206453 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0601e553-5305-4f54-9a2d-43b43a1a5dd4-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "0601e553-5305-4f54-9a2d-43b43a1a5dd4" (UID: "0601e553-5305-4f54-9a2d-43b43a1a5dd4"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.190895 4997 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.206550 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c51acefb-ed8a-480c-9ab6-0345f2ff0cd3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.206565 4997 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.206578 4997 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.206590 4997 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f158874-152b-46ce-ac42-d202684853ca-pod-info\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.206601 4997 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.206634 4997 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.206646 4997 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.206657 4997 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f158874-152b-46ce-ac42-d202684853ca-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.209460 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "ef43a612-a3df-4b50-9bf3-e5d8098cd0fe" (UID: "ef43a612-a3df-4b50-9bf3-e5d8098cd0fe"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.219602 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-54ff4c8496-8z64m"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.230115 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-db8868644-mchk9"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.238305 4997 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.239644 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-db8868644-mchk9"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.248187 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.267210 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 07:23:33 crc kubenswrapper[4997]: E1205 07:23:33.289330 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38 is running failed: container process not found" containerID="f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Dec 05 07:23:33 crc kubenswrapper[4997]: E1205 07:23:33.290140 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38 is running failed: container process not found" containerID="f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Dec 05 07:23:33 crc kubenswrapper[4997]: E1205 07:23:33.294182 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38 is running failed: container process not found" containerID="f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Dec 05 07:23:33 crc kubenswrapper[4997]: E1205 07:23:33.294249 4997 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-vnncj" podUID="5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" containerName="ovn-controller" Dec 05 07:23:33 crc kubenswrapper[4997]: E1205 07:23:33.295292 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:33 crc kubenswrapper[4997]: E1205 07:23:33.295849 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:33 crc kubenswrapper[4997]: E1205 07:23:33.296263 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:33 crc kubenswrapper[4997]: E1205 07:23:33.296305 4997 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sctbp" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovsdb-server" Dec 05 07:23:33 crc kubenswrapper[4997]: E1205 07:23:33.296443 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:33 crc kubenswrapper[4997]: E1205 07:23:33.298288 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:33 crc kubenswrapper[4997]: E1205 07:23:33.303565 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:33 crc kubenswrapper[4997]: E1205 07:23:33.303657 4997 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sctbp" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovs-vswitchd" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.309485 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="bcbe26f4-fcd9-46c7-aa3a-5038b801741d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": dial tcp 10.217.0.204:8775: i/o timeout" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.309630 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="bcbe26f4-fcd9-46c7-aa3a-5038b801741d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": context deadline exceeded" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.310830 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.310877 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.310888 4997 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0601e553-5305-4f54-9a2d-43b43a1a5dd4-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.310900 4997 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.310913 4997 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.310923 4997 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.310931 4997 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-pod-info\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.310958 4997 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0601e553-5305-4f54-9a2d-43b43a1a5dd4-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: E1205 07:23:33.311047 4997 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 07:23:33 crc kubenswrapper[4997]: E1205 07:23:33.311162 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts podName:8f2677aa-a856-47de-8417-5bc3d08d6ff1 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:35.311087094 +0000 UTC m=+1715.839994355 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts") pod "glance7904-account-delete-v5xm8" (UID: "8f2677aa-a856-47de-8417-5bc3d08d6ff1") : configmap "openstack-scripts" not found Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.314821 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.318793 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "0f158874-152b-46ce-ac42-d202684853ca" (UID: "0f158874-152b-46ce-ac42-d202684853ca"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.319453 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-server-conf" (OuterVolumeSpecName: "server-conf") pod "0f158874-152b-46ce-ac42-d202684853ca" (UID: "0f158874-152b-46ce-ac42-d202684853ca"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.330445 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-kube-api-access-9hrfb" (OuterVolumeSpecName: "kube-api-access-9hrfb") pod "ef43a612-a3df-4b50-9bf3-e5d8098cd0fe" (UID: "ef43a612-a3df-4b50-9bf3-e5d8098cd0fe"). InnerVolumeSpecName "kube-api-access-9hrfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.330530 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0601e553-5305-4f54-9a2d-43b43a1a5dd4-kube-api-access-6jwls" (OuterVolumeSpecName: "kube-api-access-6jwls") pod "0601e553-5305-4f54-9a2d-43b43a1a5dd4" (UID: "0601e553-5305-4f54-9a2d-43b43a1a5dd4"). InnerVolumeSpecName "kube-api-access-6jwls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.336626 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-zf87t"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.343439 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.349595 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-config-data" (OuterVolumeSpecName: "config-data") pod "0f158874-152b-46ce-ac42-d202684853ca" (UID: "0f158874-152b-46ce-ac42-d202684853ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.350715 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-zf87t"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.353944 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0601e553-5305-4f54-9a2d-43b43a1a5dd4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0601e553-5305-4f54-9a2d-43b43a1a5dd4" (UID: "0601e553-5305-4f54-9a2d-43b43a1a5dd4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.354527 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-server-conf" (OuterVolumeSpecName: "server-conf") pod "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" (UID: "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.355353 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef43a612-a3df-4b50-9bf3-e5d8098cd0fe" (UID: "ef43a612-a3df-4b50-9bf3-e5d8098cd0fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.356827 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-7904-account-create-update-j4c7h"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.363855 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-7904-account-create-update-j4c7h"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.380786 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance7904-account-delete-v5xm8"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.389128 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.395357 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "mysql-db") pod "0601e553-5305-4f54-9a2d-43b43a1a5dd4" (UID: "0601e553-5305-4f54-9a2d-43b43a1a5dd4"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.395463 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.401217 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5cb79f8cdd-zkk89"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.406961 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5cb79f8cdd-zkk89"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.414106 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.414137 4997 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f158874-152b-46ce-ac42-d202684853ca-server-conf\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.414151 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hrfb\" (UniqueName: \"kubernetes.io/projected/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-kube-api-access-9hrfb\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.414176 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.414187 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0601e553-5305-4f54-9a2d-43b43a1a5dd4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.414198 4997 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-server-conf\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.414218 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.414804 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jwls\" (UniqueName: \"kubernetes.io/projected/0601e553-5305-4f54-9a2d-43b43a1a5dd4-kube-api-access-6jwls\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.414820 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: E1205 07:23:33.414380 4997 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 07:23:33 crc kubenswrapper[4997]: E1205 07:23:33.414910 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a5edac3e-6bee-494f-97ba-c4047064d02f-operator-scripts podName:a5edac3e-6bee-494f-97ba-c4047064d02f nodeName:}" failed. No retries permitted until 2025-12-05 07:23:35.414882988 +0000 UTC m=+1715.943790249 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/a5edac3e-6bee-494f-97ba-c4047064d02f-operator-scripts") pod "novaapi8ba6-account-delete-t8gw7" (UID: "a5edac3e-6bee-494f-97ba-c4047064d02f") : configmap "openstack-scripts" not found Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.419201 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican69df-account-delete-tqlsz"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.420969 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0f158874-152b-46ce-ac42-d202684853ca" (UID: "0f158874-152b-46ce-ac42-d202684853ca"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.424994 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican69df-account-delete-tqlsz"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.458978 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "ef43a612-a3df-4b50-9bf3-e5d8098cd0fe" (UID: "ef43a612-a3df-4b50-9bf3-e5d8098cd0fe"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.480306 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0601e553-5305-4f54-9a2d-43b43a1a5dd4-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "0601e553-5305-4f54-9a2d-43b43a1a5dd4" (UID: "0601e553-5305-4f54-9a2d-43b43a1a5dd4"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.480436 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-555c8b7b77-44qw4"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.480480 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-555c8b7b77-44qw4"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.498905 4997 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.517140 4997 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0601e553-5305-4f54-9a2d-43b43a1a5dd4-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.517280 4997 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f158874-152b-46ce-ac42-d202684853ca-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.517348 4997 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.517402 4997 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.531752 4997 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.538578 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" (UID: "1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.539722 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.565829 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.594852 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone3eec-account-delete-v5k7n"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.594902 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone3eec-account-delete-v5k7n"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.611029 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cindera458-account-delete-vpgnx"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.618201 4997 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.618234 4997 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.618246 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.618257 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sx4dp\" (UniqueName: \"kubernetes.io/projected/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1-kube-api-access-sx4dp\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.629458 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cindera458-account-delete-vpgnx"] Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.636874 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-vnncj_5e4237cc-8581-4ddb-b472-e1c6a4a41cb5/ovn-controller/0.log" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.636941 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnncj" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.719668 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-ovn-controller-tls-certs\") pod \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.719863 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-run-ovn\") pod \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.719933 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-scripts\") pod \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.720137 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-run\") pod \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.720157 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dk79\" (UniqueName: \"kubernetes.io/projected/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-kube-api-access-7dk79\") pod \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.720201 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-combined-ca-bundle\") pod \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.720236 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-log-ovn\") pod \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\" (UID: \"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.721643 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-scripts" (OuterVolumeSpecName: "scripts") pod "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" (UID: "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.722050 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" (UID: "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.722111 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-run" (OuterVolumeSpecName: "var-run") pod "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" (UID: "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.722153 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" (UID: "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.725213 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-kube-api-access-7dk79" (OuterVolumeSpecName: "kube-api-access-7dk79") pod "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" (UID: "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5"). InnerVolumeSpecName "kube-api-access-7dk79". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.752241 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:23:33 crc kubenswrapper[4997]: E1205 07:23:33.753121 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.756259 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" (UID: "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.771692 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13fc4b26-4375-4ea2-8eaa-2417d64b3cd1" path="/var/lib/kubelet/pods/13fc4b26-4375-4ea2-8eaa-2417d64b3cd1/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.772338 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16042c4a-2200-4f8e-b899-6b75e2b1709b" path="/var/lib/kubelet/pods/16042c4a-2200-4f8e-b899-6b75e2b1709b/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.772951 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2179d060-8f4c-413d-a202-e31ebc242dfc" path="/var/lib/kubelet/pods/2179d060-8f4c-413d-a202-e31ebc242dfc/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.774143 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a733d85-ad49-41b8-a75b-842dce56e85c" path="/var/lib/kubelet/pods/2a733d85-ad49-41b8-a75b-842dce56e85c/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.774752 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3" path="/var/lib/kubelet/pods/4c6d7e30-12d4-4811-b3b8-a1cebd28a0b3/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.775279 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50b8f9d6-7eeb-41ea-8746-76c636b2d9d1" path="/var/lib/kubelet/pods/50b8f9d6-7eeb-41ea-8746-76c636b2d9d1/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.776656 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5afeb230-4a35-4872-bffa-c882fe927643" path="/var/lib/kubelet/pods/5afeb230-4a35-4872-bffa-c882fe927643/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.777350 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67d9e01d-5189-4aac-8f1a-b1f09fe138b8" path="/var/lib/kubelet/pods/67d9e01d-5189-4aac-8f1a-b1f09fe138b8/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.778485 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68490a68-dc32-448f-90cc-c146ba49c1a8" path="/var/lib/kubelet/pods/68490a68-dc32-448f-90cc-c146ba49c1a8/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.779428 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f0b9769-3e80-4475-b8d3-5fe6c5a265e7" path="/var/lib/kubelet/pods/6f0b9769-3e80-4475-b8d3-5fe6c5a265e7/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.780334 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56" path="/var/lib/kubelet/pods/6f5dfa02-1cb3-47ca-a8a3-53ab7980ac56/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.781004 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e06bc74-47af-4636-95c6-5c326c783d9e" path="/var/lib/kubelet/pods/8e06bc74-47af-4636-95c6-5c326c783d9e/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.782028 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3705104-b487-4dd7-b283-db1edcf99ec5" path="/var/lib/kubelet/pods/a3705104-b487-4dd7-b283-db1edcf99ec5/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.782559 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6bfb8b2-df28-4992-8831-2a5060323ddf" path="/var/lib/kubelet/pods/a6bfb8b2-df28-4992-8831-2a5060323ddf/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.783100 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a88492d3-fed2-456f-9f18-5ddfe089bd4b" path="/var/lib/kubelet/pods/a88492d3-fed2-456f-9f18-5ddfe089bd4b/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.784059 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8" path="/var/lib/kubelet/pods/b4da9b97-9f9f-4c6c-b751-6003e3f1c7e8/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.784536 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8c5aa81-a666-4ae6-a41c-fd31df260223" path="/var/lib/kubelet/pods/b8c5aa81-a666-4ae6-a41c-fd31df260223/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.785101 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcbe26f4-fcd9-46c7-aa3a-5038b801741d" path="/var/lib/kubelet/pods/bcbe26f4-fcd9-46c7-aa3a-5038b801741d/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.786065 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdbf2efb-5ab9-4018-86f0-59b9e37db599" path="/var/lib/kubelet/pods/bdbf2efb-5ab9-4018-86f0-59b9e37db599/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.786452 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1" path="/var/lib/kubelet/pods/c4f0352d-55f9-4a42-a9ff-e849cfd5bfb1/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.786825 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c66780bb-35b5-460f-bf9f-4c30e1a640c5" path="/var/lib/kubelet/pods/c66780bb-35b5-460f-bf9f-4c30e1a640c5/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.787303 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c706bba7-965c-497f-ae7d-b7087d37d70e" path="/var/lib/kubelet/pods/c706bba7-965c-497f-ae7d-b7087d37d70e/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.787828 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf72e65a-bae4-4ce4-b65a-d5a587ef5f48" path="/var/lib/kubelet/pods/cf72e65a-bae4-4ce4-b65a-d5a587ef5f48/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.788829 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d28a1824-1cc9-41be-b4af-1957b982a294" path="/var/lib/kubelet/pods/d28a1824-1cc9-41be-b4af-1957b982a294/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.789351 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d65cd492-d4bd-4489-9d28-a6c42407101a" path="/var/lib/kubelet/pods/d65cd492-d4bd-4489-9d28-a6c42407101a/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.789903 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e08a39e3-f267-496d-80b3-b12a9eef14c1" path="/var/lib/kubelet/pods/e08a39e3-f267-496d-80b3-b12a9eef14c1/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.791495 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eec54251-6dde-4c1b-a588-3747c8e81fd9" path="/var/lib/kubelet/pods/eec54251-6dde-4c1b-a588-3747c8e81fd9/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.792012 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946" path="/var/lib/kubelet/pods/f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.792464 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7fae453-348d-44cf-97dc-5e45252e0024" path="/var/lib/kubelet/pods/f7fae453-348d-44cf-97dc-5e45252e0024/volumes" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.806084 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" (UID: "5e4237cc-8581-4ddb-b472-e1c6a4a41cb5"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.824371 4997 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.824434 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.824453 4997 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-run\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.824467 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dk79\" (UniqueName: \"kubernetes.io/projected/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-kube-api-access-7dk79\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.824484 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.824516 4997 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.824530 4997 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.898406 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.925224 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-fernet-keys\") pod \"6783d79d-4874-43c1-92bc-e09133e45989\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.925280 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-combined-ca-bundle\") pod \"6783d79d-4874-43c1-92bc-e09133e45989\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.925310 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2tjr\" (UniqueName: \"kubernetes.io/projected/6783d79d-4874-43c1-92bc-e09133e45989-kube-api-access-b2tjr\") pod \"6783d79d-4874-43c1-92bc-e09133e45989\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.925333 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-credential-keys\") pod \"6783d79d-4874-43c1-92bc-e09133e45989\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.925366 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-config-data\") pod \"6783d79d-4874-43c1-92bc-e09133e45989\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.925411 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-public-tls-certs\") pod \"6783d79d-4874-43c1-92bc-e09133e45989\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.925442 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-scripts\") pod \"6783d79d-4874-43c1-92bc-e09133e45989\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.925463 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-internal-tls-certs\") pod \"6783d79d-4874-43c1-92bc-e09133e45989\" (UID: \"6783d79d-4874-43c1-92bc-e09133e45989\") " Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.928557 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6783d79d-4874-43c1-92bc-e09133e45989" (UID: "6783d79d-4874-43c1-92bc-e09133e45989"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.928739 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6783d79d-4874-43c1-92bc-e09133e45989-kube-api-access-b2tjr" (OuterVolumeSpecName: "kube-api-access-b2tjr") pod "6783d79d-4874-43c1-92bc-e09133e45989" (UID: "6783d79d-4874-43c1-92bc-e09133e45989"). InnerVolumeSpecName "kube-api-access-b2tjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.930670 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6783d79d-4874-43c1-92bc-e09133e45989" (UID: "6783d79d-4874-43c1-92bc-e09133e45989"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.931909 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-scripts" (OuterVolumeSpecName: "scripts") pod "6783d79d-4874-43c1-92bc-e09133e45989" (UID: "6783d79d-4874-43c1-92bc-e09133e45989"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.981961 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6783d79d-4874-43c1-92bc-e09133e45989" (UID: "6783d79d-4874-43c1-92bc-e09133e45989"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.988290 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ef43a612-a3df-4b50-9bf3-e5d8098cd0fe","Type":"ContainerDied","Data":"7b458905cf9b2e7f5e8005d6d0b844702a1acbf3e7f6db00f7a5af4e951d7c76"} Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.988379 4997 scope.go:117] "RemoveContainer" containerID="175b217107beb960b8632016072bdd9ad602c7c079fe319f224a856dd6e14c82" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.988567 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.999738 4997 generic.go:334] "Generic (PLEG): container finished" podID="6783d79d-4874-43c1-92bc-e09133e45989" containerID="99025ec523a5c575412ef4a88b541702a67a5c68296982dfdc6d2fccabd2bb56" exitCode=0 Dec 05 07:23:33 crc kubenswrapper[4997]: I1205 07:23:33.999913 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5477474c5d-zs25t" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.000487 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5477474c5d-zs25t" event={"ID":"6783d79d-4874-43c1-92bc-e09133e45989","Type":"ContainerDied","Data":"99025ec523a5c575412ef4a88b541702a67a5c68296982dfdc6d2fccabd2bb56"} Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.000741 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5477474c5d-zs25t" event={"ID":"6783d79d-4874-43c1-92bc-e09133e45989","Type":"ContainerDied","Data":"61c17f937766b0c229e3792ae938a47d02305aa8b1b0a03c6224aad70b14169a"} Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.005702 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6783d79d-4874-43c1-92bc-e09133e45989" (UID: "6783d79d-4874-43c1-92bc-e09133e45989"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.005942 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-config-data" (OuterVolumeSpecName: "config-data") pod "6783d79d-4874-43c1-92bc-e09133e45989" (UID: "6783d79d-4874-43c1-92bc-e09133e45989"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.007872 4997 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapi8ba6-account-delete-t8gw7" secret="" err="secret \"galera-openstack-dockercfg-ct8h9\" not found" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.007918 4997 scope.go:117] "RemoveContainer" containerID="f81fa8eed1dfdce11148625c1196e1d19831afebfa221d3caa7686de91ae7b79" Dec 05 07:23:34 crc kubenswrapper[4997]: E1205 07:23:34.008111 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=novaapi8ba6-account-delete-t8gw7_openstack(a5edac3e-6bee-494f-97ba-c4047064d02f)\"" pod="openstack/novaapi8ba6-account-delete-t8gw7" podUID="a5edac3e-6bee-494f-97ba-c4047064d02f" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.009008 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-vnncj_5e4237cc-8581-4ddb-b472-e1c6a4a41cb5/ovn-controller/0.log" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.009070 4997 generic.go:334] "Generic (PLEG): container finished" podID="5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" containerID="f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38" exitCode=137 Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.009114 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vnncj" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.009141 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vnncj" event={"ID":"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5","Type":"ContainerDied","Data":"f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38"} Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.009163 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vnncj" event={"ID":"5e4237cc-8581-4ddb-b472-e1c6a4a41cb5","Type":"ContainerDied","Data":"7206db5b438aaffeeecb9087bca02ee34c156f6a8cbf254755a36acf9ea0c826"} Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.022959 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0601e553-5305-4f54-9a2d-43b43a1a5dd4","Type":"ContainerDied","Data":"bdc5c4f8e0f8135d4200ac4b9170d56f703176ebd353118f32f1d072b5b48501"} Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.023148 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.025908 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6783d79d-4874-43c1-92bc-e09133e45989" (UID: "6783d79d-4874-43c1-92bc-e09133e45989"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.026772 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.026924 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement94df-account-delete-4qxc6" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.026966 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance7904-account-delete-v5xm8" podUID="8f2677aa-a856-47de-8417-5bc3d08d6ff1" containerName="mariadb-account-delete" containerID="cri-o://67982301894febce687ec7bf69c55cab5ca1970394fb3e3692c4485a0582df66" gracePeriod=30 Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.027014 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.028661 4997 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.028680 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.028692 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2tjr\" (UniqueName: \"kubernetes.io/projected/6783d79d-4874-43c1-92bc-e09133e45989-kube-api-access-b2tjr\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.028702 4997 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.028710 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.028718 4997 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.028729 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.028738 4997 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6783d79d-4874-43c1-92bc-e09133e45989-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.039355 4997 scope.go:117] "RemoveContainer" containerID="99025ec523a5c575412ef4a88b541702a67a5c68296982dfdc6d2fccabd2bb56" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.076955 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.077027 4997 scope.go:117] "RemoveContainer" containerID="99025ec523a5c575412ef4a88b541702a67a5c68296982dfdc6d2fccabd2bb56" Dec 05 07:23:34 crc kubenswrapper[4997]: E1205 07:23:34.078129 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99025ec523a5c575412ef4a88b541702a67a5c68296982dfdc6d2fccabd2bb56\": container with ID starting with 99025ec523a5c575412ef4a88b541702a67a5c68296982dfdc6d2fccabd2bb56 not found: ID does not exist" containerID="99025ec523a5c575412ef4a88b541702a67a5c68296982dfdc6d2fccabd2bb56" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.078200 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99025ec523a5c575412ef4a88b541702a67a5c68296982dfdc6d2fccabd2bb56"} err="failed to get container status \"99025ec523a5c575412ef4a88b541702a67a5c68296982dfdc6d2fccabd2bb56\": rpc error: code = NotFound desc = could not find container \"99025ec523a5c575412ef4a88b541702a67a5c68296982dfdc6d2fccabd2bb56\": container with ID starting with 99025ec523a5c575412ef4a88b541702a67a5c68296982dfdc6d2fccabd2bb56 not found: ID does not exist" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.078224 4997 scope.go:117] "RemoveContainer" containerID="f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.091985 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.122159 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement94df-account-delete-4qxc6"] Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.132573 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement94df-account-delete-4qxc6"] Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.144391 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.157132 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.164652 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.172694 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.179022 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.185322 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.190519 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-vnncj"] Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.195781 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-vnncj"] Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.207235 4997 scope.go:117] "RemoveContainer" containerID="f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38" Dec 05 07:23:34 crc kubenswrapper[4997]: E1205 07:23:34.207877 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38\": container with ID starting with f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38 not found: ID does not exist" containerID="f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.207955 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38"} err="failed to get container status \"f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38\": rpc error: code = NotFound desc = could not find container \"f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38\": container with ID starting with f5150cf6f1b1a8bfae552ef4680b59fd41db53be420981914e9241469b6b9b38 not found: ID does not exist" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.208007 4997 scope.go:117] "RemoveContainer" containerID="22c17242a1211bc3c93181daebd1e55ec7c7f3dfb0c90a0942488ba51f04fe30" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.236543 4997 scope.go:117] "RemoveContainer" containerID="f17f8b4ededafb62d528cf7bfb22c523561aaf354b59cd3fc4a9694a352a9d0b" Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.341811 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5477474c5d-zs25t"] Dec 05 07:23:34 crc kubenswrapper[4997]: I1205 07:23:34.352463 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-5477474c5d-zs25t"] Dec 05 07:23:35 crc kubenswrapper[4997]: E1205 07:23:35.374864 4997 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 07:23:35 crc kubenswrapper[4997]: E1205 07:23:35.374978 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts podName:8f2677aa-a856-47de-8417-5bc3d08d6ff1 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:39.37495327 +0000 UTC m=+1719.903860531 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts") pod "glance7904-account-delete-v5xm8" (UID: "8f2677aa-a856-47de-8417-5bc3d08d6ff1") : configmap "openstack-scripts" not found Dec 05 07:23:35 crc kubenswrapper[4997]: E1205 07:23:35.475571 4997 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 07:23:35 crc kubenswrapper[4997]: E1205 07:23:35.476038 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a5edac3e-6bee-494f-97ba-c4047064d02f-operator-scripts podName:a5edac3e-6bee-494f-97ba-c4047064d02f nodeName:}" failed. No retries permitted until 2025-12-05 07:23:39.47601738 +0000 UTC m=+1720.004924641 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/a5edac3e-6bee-494f-97ba-c4047064d02f-operator-scripts") pod "novaapi8ba6-account-delete-t8gw7" (UID: "a5edac3e-6bee-494f-97ba-c4047064d02f") : configmap "openstack-scripts" not found Dec 05 07:23:35 crc kubenswrapper[4997]: I1205 07:23:35.737650 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi8ba6-account-delete-t8gw7" Dec 05 07:23:35 crc kubenswrapper[4997]: I1205 07:23:35.763157 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0601e553-5305-4f54-9a2d-43b43a1a5dd4" path="/var/lib/kubelet/pods/0601e553-5305-4f54-9a2d-43b43a1a5dd4/volumes" Dec 05 07:23:35 crc kubenswrapper[4997]: I1205 07:23:35.766832 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f158874-152b-46ce-ac42-d202684853ca" path="/var/lib/kubelet/pods/0f158874-152b-46ce-ac42-d202684853ca/volumes" Dec 05 07:23:35 crc kubenswrapper[4997]: I1205 07:23:35.767570 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" path="/var/lib/kubelet/pods/1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79/volumes" Dec 05 07:23:35 crc kubenswrapper[4997]: I1205 07:23:35.769440 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" path="/var/lib/kubelet/pods/5e4237cc-8581-4ddb-b472-e1c6a4a41cb5/volumes" Dec 05 07:23:35 crc kubenswrapper[4997]: I1205 07:23:35.770547 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6783d79d-4874-43c1-92bc-e09133e45989" path="/var/lib/kubelet/pods/6783d79d-4874-43c1-92bc-e09133e45989/volumes" Dec 05 07:23:35 crc kubenswrapper[4997]: I1205 07:23:35.771408 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c51acefb-ed8a-480c-9ab6-0345f2ff0cd3" path="/var/lib/kubelet/pods/c51acefb-ed8a-480c-9ab6-0345f2ff0cd3/volumes" Dec 05 07:23:35 crc kubenswrapper[4997]: I1205 07:23:35.773403 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef43a612-a3df-4b50-9bf3-e5d8098cd0fe" path="/var/lib/kubelet/pods/ef43a612-a3df-4b50-9bf3-e5d8098cd0fe/volumes" Dec 05 07:23:35 crc kubenswrapper[4997]: I1205 07:23:35.779019 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5edac3e-6bee-494f-97ba-c4047064d02f-operator-scripts\") pod \"a5edac3e-6bee-494f-97ba-c4047064d02f\" (UID: \"a5edac3e-6bee-494f-97ba-c4047064d02f\") " Dec 05 07:23:35 crc kubenswrapper[4997]: I1205 07:23:35.779208 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlspm\" (UniqueName: \"kubernetes.io/projected/a5edac3e-6bee-494f-97ba-c4047064d02f-kube-api-access-rlspm\") pod \"a5edac3e-6bee-494f-97ba-c4047064d02f\" (UID: \"a5edac3e-6bee-494f-97ba-c4047064d02f\") " Dec 05 07:23:35 crc kubenswrapper[4997]: I1205 07:23:35.780056 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5edac3e-6bee-494f-97ba-c4047064d02f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a5edac3e-6bee-494f-97ba-c4047064d02f" (UID: "a5edac3e-6bee-494f-97ba-c4047064d02f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:35 crc kubenswrapper[4997]: I1205 07:23:35.804393 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5edac3e-6bee-494f-97ba-c4047064d02f-kube-api-access-rlspm" (OuterVolumeSpecName: "kube-api-access-rlspm") pod "a5edac3e-6bee-494f-97ba-c4047064d02f" (UID: "a5edac3e-6bee-494f-97ba-c4047064d02f"). InnerVolumeSpecName "kube-api-access-rlspm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:35 crc kubenswrapper[4997]: I1205 07:23:35.886495 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5edac3e-6bee-494f-97ba-c4047064d02f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:35 crc kubenswrapper[4997]: I1205 07:23:35.886561 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlspm\" (UniqueName: \"kubernetes.io/projected/a5edac3e-6bee-494f-97ba-c4047064d02f-kube-api-access-rlspm\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.119149 4997 generic.go:334] "Generic (PLEG): container finished" podID="db136058-17a8-4955-8e95-576da7b9e847" containerID="b3cfedfae8a42f278d8fff88f2e1586ca243b6222c804510c5b0435a9a5e1e73" exitCode=0 Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.119239 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db136058-17a8-4955-8e95-576da7b9e847","Type":"ContainerDied","Data":"b3cfedfae8a42f278d8fff88f2e1586ca243b6222c804510c5b0435a9a5e1e73"} Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.121854 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi8ba6-account-delete-t8gw7" event={"ID":"a5edac3e-6bee-494f-97ba-c4047064d02f","Type":"ContainerDied","Data":"4c10b2bcf8b062435e447eb41d03529fe258d8e27842fe2f7a56df03885dbf7e"} Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.121904 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi8ba6-account-delete-t8gw7" Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.121924 4997 scope.go:117] "RemoveContainer" containerID="f81fa8eed1dfdce11148625c1196e1d19831afebfa221d3caa7686de91ae7b79" Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.159606 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi8ba6-account-delete-t8gw7"] Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.165427 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapi8ba6-account-delete-t8gw7"] Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.250653 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.394439 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-config-data\") pod \"db136058-17a8-4955-8e95-576da7b9e847\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.394584 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-sg-core-conf-yaml\") pod \"db136058-17a8-4955-8e95-576da7b9e847\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.394642 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db136058-17a8-4955-8e95-576da7b9e847-run-httpd\") pod \"db136058-17a8-4955-8e95-576da7b9e847\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.394697 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-combined-ca-bundle\") pod \"db136058-17a8-4955-8e95-576da7b9e847\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.394723 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-scripts\") pod \"db136058-17a8-4955-8e95-576da7b9e847\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.394748 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmrcw\" (UniqueName: \"kubernetes.io/projected/db136058-17a8-4955-8e95-576da7b9e847-kube-api-access-dmrcw\") pod \"db136058-17a8-4955-8e95-576da7b9e847\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.394830 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-ceilometer-tls-certs\") pod \"db136058-17a8-4955-8e95-576da7b9e847\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.394863 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db136058-17a8-4955-8e95-576da7b9e847-log-httpd\") pod \"db136058-17a8-4955-8e95-576da7b9e847\" (UID: \"db136058-17a8-4955-8e95-576da7b9e847\") " Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.395667 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db136058-17a8-4955-8e95-576da7b9e847-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "db136058-17a8-4955-8e95-576da7b9e847" (UID: "db136058-17a8-4955-8e95-576da7b9e847"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.396531 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db136058-17a8-4955-8e95-576da7b9e847-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "db136058-17a8-4955-8e95-576da7b9e847" (UID: "db136058-17a8-4955-8e95-576da7b9e847"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.399248 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-scripts" (OuterVolumeSpecName: "scripts") pod "db136058-17a8-4955-8e95-576da7b9e847" (UID: "db136058-17a8-4955-8e95-576da7b9e847"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.409948 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db136058-17a8-4955-8e95-576da7b9e847-kube-api-access-dmrcw" (OuterVolumeSpecName: "kube-api-access-dmrcw") pod "db136058-17a8-4955-8e95-576da7b9e847" (UID: "db136058-17a8-4955-8e95-576da7b9e847"). InnerVolumeSpecName "kube-api-access-dmrcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:36 crc kubenswrapper[4997]: I1205 07:23:36.418893 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "db136058-17a8-4955-8e95-576da7b9e847" (UID: "db136058-17a8-4955-8e95-576da7b9e847"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:36.496587 4997 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:36.496652 4997 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db136058-17a8-4955-8e95-576da7b9e847-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:36.496662 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:36.496673 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmrcw\" (UniqueName: \"kubernetes.io/projected/db136058-17a8-4955-8e95-576da7b9e847-kube-api-access-dmrcw\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:36.496683 4997 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db136058-17a8-4955-8e95-576da7b9e847-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:37.172125 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "db136058-17a8-4955-8e95-576da7b9e847" (UID: "db136058-17a8-4955-8e95-576da7b9e847"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:37.179061 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: i/o timeout" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:37.197060 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:37.197861 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db136058-17a8-4955-8e95-576da7b9e847","Type":"ContainerDied","Data":"0d961eed77dab6e083f388b65d5739d3ebd5af4ac39cadc5d8e1338bcaa03957"} Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:37.197904 4997 scope.go:117] "RemoveContainer" containerID="fdb72dc3adf6464397dca00e5b9745e3a6b89a71985325529d858ab69ce19e3c" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:37.207810 4997 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:37.216153 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-config-data" (OuterVolumeSpecName: "config-data") pod "db136058-17a8-4955-8e95-576da7b9e847" (UID: "db136058-17a8-4955-8e95-576da7b9e847"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:37.218875 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db136058-17a8-4955-8e95-576da7b9e847" (UID: "db136058-17a8-4955-8e95-576da7b9e847"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:37.234851 4997 scope.go:117] "RemoveContainer" containerID="3fc11187932225dbd83eb0de61fd45f5e409020f11e6e1a8a12baa4ea259218b" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:37.300124 4997 scope.go:117] "RemoveContainer" containerID="b3cfedfae8a42f278d8fff88f2e1586ca243b6222c804510c5b0435a9a5e1e73" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:37.312045 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:37.312187 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db136058-17a8-4955-8e95-576da7b9e847-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:37.331955 4997 scope.go:117] "RemoveContainer" containerID="07f1631824a73521a5df4d80b3bf4feb8dde04c7a5b2c5e5688b95c1efabb84c" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:37.533469 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:37.545989 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:37.767001 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5edac3e-6bee-494f-97ba-c4047064d02f" path="/var/lib/kubelet/pods/a5edac3e-6bee-494f-97ba-c4047064d02f/volumes" Dec 05 07:23:37 crc kubenswrapper[4997]: I1205 07:23:37.767940 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db136058-17a8-4955-8e95-576da7b9e847" path="/var/lib/kubelet/pods/db136058-17a8-4955-8e95-576da7b9e847/volumes" Dec 05 07:23:38 crc kubenswrapper[4997]: E1205 07:23:38.295343 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:38 crc kubenswrapper[4997]: E1205 07:23:38.296654 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:38 crc kubenswrapper[4997]: E1205 07:23:38.296889 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:38 crc kubenswrapper[4997]: E1205 07:23:38.296930 4997 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sctbp" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovsdb-server" Dec 05 07:23:38 crc kubenswrapper[4997]: E1205 07:23:38.297582 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:38 crc kubenswrapper[4997]: E1205 07:23:38.301359 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:38 crc kubenswrapper[4997]: E1205 07:23:38.303155 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:38 crc kubenswrapper[4997]: E1205 07:23:38.303187 4997 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sctbp" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovs-vswitchd" Dec 05 07:23:39 crc kubenswrapper[4997]: E1205 07:23:39.376237 4997 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 07:23:39 crc kubenswrapper[4997]: E1205 07:23:39.377105 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts podName:8f2677aa-a856-47de-8417-5bc3d08d6ff1 nodeName:}" failed. No retries permitted until 2025-12-05 07:23:47.377080216 +0000 UTC m=+1727.905987477 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts") pod "glance7904-account-delete-v5xm8" (UID: "8f2677aa-a856-47de-8417-5bc3d08d6ff1") : configmap "openstack-scripts" not found Dec 05 07:23:43 crc kubenswrapper[4997]: E1205 07:23:43.295103 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:43 crc kubenswrapper[4997]: E1205 07:23:43.297194 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:43 crc kubenswrapper[4997]: E1205 07:23:43.297486 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:43 crc kubenswrapper[4997]: E1205 07:23:43.297543 4997 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sctbp" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovsdb-server" Dec 05 07:23:43 crc kubenswrapper[4997]: E1205 07:23:43.298143 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:43 crc kubenswrapper[4997]: E1205 07:23:43.299451 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:43 crc kubenswrapper[4997]: E1205 07:23:43.300594 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:43 crc kubenswrapper[4997]: E1205 07:23:43.300664 4997 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sctbp" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovs-vswitchd" Dec 05 07:23:45 crc kubenswrapper[4997]: I1205 07:23:45.938414 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-74f5b8d45c-jbkcz" podUID="c835ac5d-bd54-46d4-aab3-3ef7682f50b7" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.152:9696/\": dial tcp 10.217.0.152:9696: connect: connection refused" Dec 05 07:23:46 crc kubenswrapper[4997]: I1205 07:23:46.752703 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:23:46 crc kubenswrapper[4997]: E1205 07:23:46.753057 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.251313 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.319241 4997 generic.go:334] "Generic (PLEG): container finished" podID="c835ac5d-bd54-46d4-aab3-3ef7682f50b7" containerID="8ffd311c9d4ec727b6a0cad0412afdc5d1ea879f7894edd49a446d48d71d135f" exitCode=0 Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.319308 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-74f5b8d45c-jbkcz" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.319305 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74f5b8d45c-jbkcz" event={"ID":"c835ac5d-bd54-46d4-aab3-3ef7682f50b7","Type":"ContainerDied","Data":"8ffd311c9d4ec727b6a0cad0412afdc5d1ea879f7894edd49a446d48d71d135f"} Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.319379 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74f5b8d45c-jbkcz" event={"ID":"c835ac5d-bd54-46d4-aab3-3ef7682f50b7","Type":"ContainerDied","Data":"bd761217d241bd21cfb00057e7c3e3d35d519e845af67b85fb45144e1f373129"} Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.319405 4997 scope.go:117] "RemoveContainer" containerID="7f4731817ad8f583546eee8efc40ce792a1f83dad3127d8c88136943c5c893be" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.333172 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-ovndb-tls-certs\") pod \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.333280 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-config\") pod \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.333330 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-internal-tls-certs\") pod \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.333399 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-combined-ca-bundle\") pod \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.333478 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gplnq\" (UniqueName: \"kubernetes.io/projected/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-kube-api-access-gplnq\") pod \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.333515 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-public-tls-certs\") pod \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.333573 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-httpd-config\") pod \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\" (UID: \"c835ac5d-bd54-46d4-aab3-3ef7682f50b7\") " Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.341374 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-kube-api-access-gplnq" (OuterVolumeSpecName: "kube-api-access-gplnq") pod "c835ac5d-bd54-46d4-aab3-3ef7682f50b7" (UID: "c835ac5d-bd54-46d4-aab3-3ef7682f50b7"). InnerVolumeSpecName "kube-api-access-gplnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.342343 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "c835ac5d-bd54-46d4-aab3-3ef7682f50b7" (UID: "c835ac5d-bd54-46d4-aab3-3ef7682f50b7"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.347360 4997 scope.go:117] "RemoveContainer" containerID="8ffd311c9d4ec727b6a0cad0412afdc5d1ea879f7894edd49a446d48d71d135f" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.381319 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c835ac5d-bd54-46d4-aab3-3ef7682f50b7" (UID: "c835ac5d-bd54-46d4-aab3-3ef7682f50b7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.390647 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c835ac5d-bd54-46d4-aab3-3ef7682f50b7" (UID: "c835ac5d-bd54-46d4-aab3-3ef7682f50b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.402081 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c835ac5d-bd54-46d4-aab3-3ef7682f50b7" (UID: "c835ac5d-bd54-46d4-aab3-3ef7682f50b7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.407956 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-config" (OuterVolumeSpecName: "config") pod "c835ac5d-bd54-46d4-aab3-3ef7682f50b7" (UID: "c835ac5d-bd54-46d4-aab3-3ef7682f50b7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.408207 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "c835ac5d-bd54-46d4-aab3-3ef7682f50b7" (UID: "c835ac5d-bd54-46d4-aab3-3ef7682f50b7"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.440424 4997 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.440770 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.440879 4997 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.440938 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.441004 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gplnq\" (UniqueName: \"kubernetes.io/projected/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-kube-api-access-gplnq\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.441064 4997 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.441116 4997 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c835ac5d-bd54-46d4-aab3-3ef7682f50b7-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:47 crc kubenswrapper[4997]: E1205 07:23:47.440963 4997 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 07:23:47 crc kubenswrapper[4997]: E1205 07:23:47.441268 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts podName:8f2677aa-a856-47de-8417-5bc3d08d6ff1 nodeName:}" failed. No retries permitted until 2025-12-05 07:24:03.441249084 +0000 UTC m=+1743.970156345 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts") pod "glance7904-account-delete-v5xm8" (UID: "8f2677aa-a856-47de-8417-5bc3d08d6ff1") : configmap "openstack-scripts" not found Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.448301 4997 scope.go:117] "RemoveContainer" containerID="7f4731817ad8f583546eee8efc40ce792a1f83dad3127d8c88136943c5c893be" Dec 05 07:23:47 crc kubenswrapper[4997]: E1205 07:23:47.453415 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f4731817ad8f583546eee8efc40ce792a1f83dad3127d8c88136943c5c893be\": container with ID starting with 7f4731817ad8f583546eee8efc40ce792a1f83dad3127d8c88136943c5c893be not found: ID does not exist" containerID="7f4731817ad8f583546eee8efc40ce792a1f83dad3127d8c88136943c5c893be" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.453469 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f4731817ad8f583546eee8efc40ce792a1f83dad3127d8c88136943c5c893be"} err="failed to get container status \"7f4731817ad8f583546eee8efc40ce792a1f83dad3127d8c88136943c5c893be\": rpc error: code = NotFound desc = could not find container \"7f4731817ad8f583546eee8efc40ce792a1f83dad3127d8c88136943c5c893be\": container with ID starting with 7f4731817ad8f583546eee8efc40ce792a1f83dad3127d8c88136943c5c893be not found: ID does not exist" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.453499 4997 scope.go:117] "RemoveContainer" containerID="8ffd311c9d4ec727b6a0cad0412afdc5d1ea879f7894edd49a446d48d71d135f" Dec 05 07:23:47 crc kubenswrapper[4997]: E1205 07:23:47.454111 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ffd311c9d4ec727b6a0cad0412afdc5d1ea879f7894edd49a446d48d71d135f\": container with ID starting with 8ffd311c9d4ec727b6a0cad0412afdc5d1ea879f7894edd49a446d48d71d135f not found: ID does not exist" containerID="8ffd311c9d4ec727b6a0cad0412afdc5d1ea879f7894edd49a446d48d71d135f" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.454179 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ffd311c9d4ec727b6a0cad0412afdc5d1ea879f7894edd49a446d48d71d135f"} err="failed to get container status \"8ffd311c9d4ec727b6a0cad0412afdc5d1ea879f7894edd49a446d48d71d135f\": rpc error: code = NotFound desc = could not find container \"8ffd311c9d4ec727b6a0cad0412afdc5d1ea879f7894edd49a446d48d71d135f\": container with ID starting with 8ffd311c9d4ec727b6a0cad0412afdc5d1ea879f7894edd49a446d48d71d135f not found: ID does not exist" Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.651816 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-74f5b8d45c-jbkcz"] Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.658387 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-74f5b8d45c-jbkcz"] Dec 05 07:23:47 crc kubenswrapper[4997]: I1205 07:23:47.764762 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c835ac5d-bd54-46d4-aab3-3ef7682f50b7" path="/var/lib/kubelet/pods/c835ac5d-bd54-46d4-aab3-3ef7682f50b7/volumes" Dec 05 07:23:48 crc kubenswrapper[4997]: E1205 07:23:48.295070 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:48 crc kubenswrapper[4997]: E1205 07:23:48.295900 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:48 crc kubenswrapper[4997]: E1205 07:23:48.296092 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:48 crc kubenswrapper[4997]: E1205 07:23:48.296115 4997 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sctbp" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovsdb-server" Dec 05 07:23:48 crc kubenswrapper[4997]: E1205 07:23:48.300338 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:48 crc kubenswrapper[4997]: E1205 07:23:48.302181 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:48 crc kubenswrapper[4997]: E1205 07:23:48.305453 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:48 crc kubenswrapper[4997]: E1205 07:23:48.305486 4997 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sctbp" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovs-vswitchd" Dec 05 07:23:53 crc kubenswrapper[4997]: E1205 07:23:53.294852 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:53 crc kubenswrapper[4997]: E1205 07:23:53.295502 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:53 crc kubenswrapper[4997]: E1205 07:23:53.296126 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 07:23:53 crc kubenswrapper[4997]: E1205 07:23:53.296167 4997 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sctbp" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovsdb-server" Dec 05 07:23:53 crc kubenswrapper[4997]: E1205 07:23:53.296988 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:53 crc kubenswrapper[4997]: E1205 07:23:53.298891 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:53 crc kubenswrapper[4997]: E1205 07:23:53.300503 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 07:23:53 crc kubenswrapper[4997]: E1205 07:23:53.300545 4997 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sctbp" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovs-vswitchd" Dec 05 07:23:53 crc kubenswrapper[4997]: I1205 07:23:53.957664 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sctbp_5388a6f5-ecde-42db-9c94-a52c17e8a302/ovs-vswitchd/0.log" Dec 05 07:23:53 crc kubenswrapper[4997]: I1205 07:23:53.959251 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.062287 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5388a6f5-ecde-42db-9c94-a52c17e8a302-scripts\") pod \"5388a6f5-ecde-42db-9c94-a52c17e8a302\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.062375 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-run\") pod \"5388a6f5-ecde-42db-9c94-a52c17e8a302\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.062487 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbvgw\" (UniqueName: \"kubernetes.io/projected/5388a6f5-ecde-42db-9c94-a52c17e8a302-kube-api-access-vbvgw\") pod \"5388a6f5-ecde-42db-9c94-a52c17e8a302\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.062524 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-log\") pod \"5388a6f5-ecde-42db-9c94-a52c17e8a302\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.062524 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-run" (OuterVolumeSpecName: "var-run") pod "5388a6f5-ecde-42db-9c94-a52c17e8a302" (UID: "5388a6f5-ecde-42db-9c94-a52c17e8a302"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.062547 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-etc-ovs\") pod \"5388a6f5-ecde-42db-9c94-a52c17e8a302\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.062605 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "5388a6f5-ecde-42db-9c94-a52c17e8a302" (UID: "5388a6f5-ecde-42db-9c94-a52c17e8a302"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.062627 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-log" (OuterVolumeSpecName: "var-log") pod "5388a6f5-ecde-42db-9c94-a52c17e8a302" (UID: "5388a6f5-ecde-42db-9c94-a52c17e8a302"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.062791 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-lib\") pod \"5388a6f5-ecde-42db-9c94-a52c17e8a302\" (UID: \"5388a6f5-ecde-42db-9c94-a52c17e8a302\") " Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.062862 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-lib" (OuterVolumeSpecName: "var-lib") pod "5388a6f5-ecde-42db-9c94-a52c17e8a302" (UID: "5388a6f5-ecde-42db-9c94-a52c17e8a302"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.063763 4997 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-lib\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.063795 4997 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-run\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.063809 4997 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-var-log\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.063822 4997 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5388a6f5-ecde-42db-9c94-a52c17e8a302-etc-ovs\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.064906 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5388a6f5-ecde-42db-9c94-a52c17e8a302-scripts" (OuterVolumeSpecName: "scripts") pod "5388a6f5-ecde-42db-9c94-a52c17e8a302" (UID: "5388a6f5-ecde-42db-9c94-a52c17e8a302"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.069663 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5388a6f5-ecde-42db-9c94-a52c17e8a302-kube-api-access-vbvgw" (OuterVolumeSpecName: "kube-api-access-vbvgw") pod "5388a6f5-ecde-42db-9c94-a52c17e8a302" (UID: "5388a6f5-ecde-42db-9c94-a52c17e8a302"). InnerVolumeSpecName "kube-api-access-vbvgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.165389 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5388a6f5-ecde-42db-9c94-a52c17e8a302-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.165437 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbvgw\" (UniqueName: \"kubernetes.io/projected/5388a6f5-ecde-42db-9c94-a52c17e8a302-kube-api-access-vbvgw\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.407504 4997 generic.go:334] "Generic (PLEG): container finished" podID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerID="4fb495cc6962ca168a43dc4d03761c144cd6fa4532a0aa11050db4ee869e4f5e" exitCode=137 Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.407573 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerDied","Data":"4fb495cc6962ca168a43dc4d03761c144cd6fa4532a0aa11050db4ee869e4f5e"} Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.410291 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sctbp_5388a6f5-ecde-42db-9c94-a52c17e8a302/ovs-vswitchd/0.log" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.411503 4997 generic.go:334] "Generic (PLEG): container finished" podID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" exitCode=137 Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.411548 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sctbp" event={"ID":"5388a6f5-ecde-42db-9c94-a52c17e8a302","Type":"ContainerDied","Data":"68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb"} Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.411584 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sctbp" event={"ID":"5388a6f5-ecde-42db-9c94-a52c17e8a302","Type":"ContainerDied","Data":"24e4b91d042a02125f23e127940e3768706e261ea237a3d9728dcbc9a68346aa"} Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.411685 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-sctbp" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.411605 4997 scope.go:117] "RemoveContainer" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.455494 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-sctbp"] Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.461837 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-sctbp"] Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.467025 4997 scope.go:117] "RemoveContainer" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.493340 4997 scope.go:117] "RemoveContainer" containerID="13a5e6018eea0fdd10399142992fee1827749d6c2551be33953ed7d8cf75c944" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.533132 4997 scope.go:117] "RemoveContainer" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" Dec 05 07:23:54 crc kubenswrapper[4997]: E1205 07:23:54.533672 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb\": container with ID starting with 68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb not found: ID does not exist" containerID="68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.533717 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb"} err="failed to get container status \"68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb\": rpc error: code = NotFound desc = could not find container \"68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb\": container with ID starting with 68ef975d24d468b943d1a61ef10cc6635d59fe00851a4f86277c154d474aa9bb not found: ID does not exist" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.533778 4997 scope.go:117] "RemoveContainer" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" Dec 05 07:23:54 crc kubenswrapper[4997]: E1205 07:23:54.534110 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187\": container with ID starting with 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 not found: ID does not exist" containerID="85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.534134 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187"} err="failed to get container status \"85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187\": rpc error: code = NotFound desc = could not find container \"85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187\": container with ID starting with 85e79140c57d99cfbd53dddccb106da2290299307f6d92909ed20a2319a93187 not found: ID does not exist" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.534154 4997 scope.go:117] "RemoveContainer" containerID="13a5e6018eea0fdd10399142992fee1827749d6c2551be33953ed7d8cf75c944" Dec 05 07:23:54 crc kubenswrapper[4997]: E1205 07:23:54.534442 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13a5e6018eea0fdd10399142992fee1827749d6c2551be33953ed7d8cf75c944\": container with ID starting with 13a5e6018eea0fdd10399142992fee1827749d6c2551be33953ed7d8cf75c944 not found: ID does not exist" containerID="13a5e6018eea0fdd10399142992fee1827749d6c2551be33953ed7d8cf75c944" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.534494 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13a5e6018eea0fdd10399142992fee1827749d6c2551be33953ed7d8cf75c944"} err="failed to get container status \"13a5e6018eea0fdd10399142992fee1827749d6c2551be33953ed7d8cf75c944\": rpc error: code = NotFound desc = could not find container \"13a5e6018eea0fdd10399142992fee1827749d6c2551be33953ed7d8cf75c944\": container with ID starting with 13a5e6018eea0fdd10399142992fee1827749d6c2551be33953ed7d8cf75c944 not found: ID does not exist" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.829143 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.979986 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift\") pod \"5cc70c62-ad3e-4178-932a-d14571a06daa\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.980215 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5cc70c62-ad3e-4178-932a-d14571a06daa-lock\") pod \"5cc70c62-ad3e-4178-932a-d14571a06daa\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.980263 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxmv5\" (UniqueName: \"kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-kube-api-access-hxmv5\") pod \"5cc70c62-ad3e-4178-932a-d14571a06daa\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.980295 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5cc70c62-ad3e-4178-932a-d14571a06daa-cache\") pod \"5cc70c62-ad3e-4178-932a-d14571a06daa\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.980377 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"5cc70c62-ad3e-4178-932a-d14571a06daa\" (UID: \"5cc70c62-ad3e-4178-932a-d14571a06daa\") " Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.981010 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cc70c62-ad3e-4178-932a-d14571a06daa-lock" (OuterVolumeSpecName: "lock") pod "5cc70c62-ad3e-4178-932a-d14571a06daa" (UID: "5cc70c62-ad3e-4178-932a-d14571a06daa"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.986119 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cc70c62-ad3e-4178-932a-d14571a06daa-cache" (OuterVolumeSpecName: "cache") pod "5cc70c62-ad3e-4178-932a-d14571a06daa" (UID: "5cc70c62-ad3e-4178-932a-d14571a06daa"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.987087 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5cc70c62-ad3e-4178-932a-d14571a06daa" (UID: "5cc70c62-ad3e-4178-932a-d14571a06daa"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.987136 4997 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5cc70c62-ad3e-4178-932a-d14571a06daa-lock\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.987351 4997 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5cc70c62-ad3e-4178-932a-d14571a06daa-cache\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.989730 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-kube-api-access-hxmv5" (OuterVolumeSpecName: "kube-api-access-hxmv5") pod "5cc70c62-ad3e-4178-932a-d14571a06daa" (UID: "5cc70c62-ad3e-4178-932a-d14571a06daa"). InnerVolumeSpecName "kube-api-access-hxmv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:23:54 crc kubenswrapper[4997]: I1205 07:23:54.991005 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "swift") pod "5cc70c62-ad3e-4178-932a-d14571a06daa" (UID: "5cc70c62-ad3e-4178-932a-d14571a06daa"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.089485 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxmv5\" (UniqueName: \"kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-kube-api-access-hxmv5\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.089536 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.089548 4997 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5cc70c62-ad3e-4178-932a-d14571a06daa-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.106269 4997 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.190725 4997 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.431074 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5cc70c62-ad3e-4178-932a-d14571a06daa","Type":"ContainerDied","Data":"a042f6522375f75c48ce6fb440a64a329021e4a5723014c8733e72171176d91c"} Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.431148 4997 scope.go:117] "RemoveContainer" containerID="4fb495cc6962ca168a43dc4d03761c144cd6fa4532a0aa11050db4ee869e4f5e" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.431378 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.474875 4997 scope.go:117] "RemoveContainer" containerID="d09e7dc200cd7427ec542e14f38b16b200be9b00f5c8bcc72aa1988da898ee42" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.475830 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.483367 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.500483 4997 scope.go:117] "RemoveContainer" containerID="82fa232b03602f1c8aad22b6c099f8b323fcdf24fc81e0fba8dc78d929f37d29" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.522046 4997 scope.go:117] "RemoveContainer" containerID="6916fdc8a79724bc4f616151198996d4533bcdfd1af49ce395a01c0843637cf3" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.556175 4997 scope.go:117] "RemoveContainer" containerID="7f5cbf566dd19f6211cf0d1dae49e71d2d0c93c0e3cac1ccb8cc390c300b5e33" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.583824 4997 scope.go:117] "RemoveContainer" containerID="f483369c789ec0f4e5bb24b517217914add2901c3b684a8154f738f83e6fa37b" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.603364 4997 scope.go:117] "RemoveContainer" containerID="b87c752a856aa32bd9d97eb8af0b3d65d2a9e9d6f3d911ba2d99662c5e3deef3" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.634293 4997 scope.go:117] "RemoveContainer" containerID="732523b307a12a50490d6b3563dce9bc973897fbfa53f0e26ffcb66918c5cef4" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.656849 4997 scope.go:117] "RemoveContainer" containerID="2c3fe897d3a3071e0e6d73643c8249d5b3a509b04f3d1cf2c20850f1110f9dc0" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.681053 4997 scope.go:117] "RemoveContainer" containerID="ade129587cb10e9f61150cba8b6492aafd796f81df59fc5a8e2e4f2b826b4574" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.717328 4997 scope.go:117] "RemoveContainer" containerID="e982dc1e734d65f84320cc1a7e6b801c7989b49043d99b850c0c838b530b2638" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.750354 4997 scope.go:117] "RemoveContainer" containerID="e62736a4a4741c846c90c4aeb88e01bdf1db4768265d9eb4a765c3853650e11e" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.760701 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" path="/var/lib/kubelet/pods/5388a6f5-ecde-42db-9c94-a52c17e8a302/volumes" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.762052 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" path="/var/lib/kubelet/pods/5cc70c62-ad3e-4178-932a-d14571a06daa/volumes" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.774156 4997 scope.go:117] "RemoveContainer" containerID="7ce96074e3871701d5fda41b91a5e0c7fc16c1a429f4da49f7b22d0a58f5bcfa" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.800920 4997 scope.go:117] "RemoveContainer" containerID="5adcaef44afa1157fae6d04d911a954e618faa7d796126d8888521add2da41db" Dec 05 07:23:55 crc kubenswrapper[4997]: I1205 07:23:55.825639 4997 scope.go:117] "RemoveContainer" containerID="71f731b2d431508653bb17ac618c7642061a78120cb47e7911dfe3ee31b05713" Dec 05 07:23:59 crc kubenswrapper[4997]: I1205 07:23:59.758637 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:23:59 crc kubenswrapper[4997]: E1205 07:23:59.759357 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:24:03 crc kubenswrapper[4997]: E1205 07:24:03.442878 4997 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 07:24:03 crc kubenswrapper[4997]: E1205 07:24:03.443349 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts podName:8f2677aa-a856-47de-8417-5bc3d08d6ff1 nodeName:}" failed. No retries permitted until 2025-12-05 07:24:35.443328645 +0000 UTC m=+1775.972235906 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts") pod "glance7904-account-delete-v5xm8" (UID: "8f2677aa-a856-47de-8417-5bc3d08d6ff1") : configmap "openstack-scripts" not found Dec 05 07:24:04 crc kubenswrapper[4997]: I1205 07:24:04.570850 4997 generic.go:334] "Generic (PLEG): container finished" podID="8f2677aa-a856-47de-8417-5bc3d08d6ff1" containerID="67982301894febce687ec7bf69c55cab5ca1970394fb3e3692c4485a0582df66" exitCode=137 Dec 05 07:24:04 crc kubenswrapper[4997]: I1205 07:24:04.571324 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance7904-account-delete-v5xm8" event={"ID":"8f2677aa-a856-47de-8417-5bc3d08d6ff1","Type":"ContainerDied","Data":"67982301894febce687ec7bf69c55cab5ca1970394fb3e3692c4485a0582df66"} Dec 05 07:24:04 crc kubenswrapper[4997]: I1205 07:24:04.937923 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance7904-account-delete-v5xm8" Dec 05 07:24:05 crc kubenswrapper[4997]: I1205 07:24:05.098423 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-566jt\" (UniqueName: \"kubernetes.io/projected/8f2677aa-a856-47de-8417-5bc3d08d6ff1-kube-api-access-566jt\") pod \"8f2677aa-a856-47de-8417-5bc3d08d6ff1\" (UID: \"8f2677aa-a856-47de-8417-5bc3d08d6ff1\") " Dec 05 07:24:05 crc kubenswrapper[4997]: I1205 07:24:05.098603 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts\") pod \"8f2677aa-a856-47de-8417-5bc3d08d6ff1\" (UID: \"8f2677aa-a856-47de-8417-5bc3d08d6ff1\") " Dec 05 07:24:05 crc kubenswrapper[4997]: I1205 07:24:05.099844 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8f2677aa-a856-47de-8417-5bc3d08d6ff1" (UID: "8f2677aa-a856-47de-8417-5bc3d08d6ff1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:24:05 crc kubenswrapper[4997]: I1205 07:24:05.106032 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f2677aa-a856-47de-8417-5bc3d08d6ff1-kube-api-access-566jt" (OuterVolumeSpecName: "kube-api-access-566jt") pod "8f2677aa-a856-47de-8417-5bc3d08d6ff1" (UID: "8f2677aa-a856-47de-8417-5bc3d08d6ff1"). InnerVolumeSpecName "kube-api-access-566jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:24:05 crc kubenswrapper[4997]: I1205 07:24:05.204212 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-566jt\" (UniqueName: \"kubernetes.io/projected/8f2677aa-a856-47de-8417-5bc3d08d6ff1-kube-api-access-566jt\") on node \"crc\" DevicePath \"\"" Dec 05 07:24:05 crc kubenswrapper[4997]: I1205 07:24:05.204339 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f2677aa-a856-47de-8417-5bc3d08d6ff1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 07:24:05 crc kubenswrapper[4997]: I1205 07:24:05.587999 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance7904-account-delete-v5xm8" event={"ID":"8f2677aa-a856-47de-8417-5bc3d08d6ff1","Type":"ContainerDied","Data":"705ac0ba630b3336a44765fcb97172ba22ee01af0565e74f3793afff5a05fc9a"} Dec 05 07:24:05 crc kubenswrapper[4997]: I1205 07:24:05.588556 4997 scope.go:117] "RemoveContainer" containerID="67982301894febce687ec7bf69c55cab5ca1970394fb3e3692c4485a0582df66" Dec 05 07:24:05 crc kubenswrapper[4997]: I1205 07:24:05.588863 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance7904-account-delete-v5xm8" Dec 05 07:24:05 crc kubenswrapper[4997]: I1205 07:24:05.639318 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance7904-account-delete-v5xm8"] Dec 05 07:24:05 crc kubenswrapper[4997]: I1205 07:24:05.647303 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance7904-account-delete-v5xm8"] Dec 05 07:24:05 crc kubenswrapper[4997]: I1205 07:24:05.760926 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f2677aa-a856-47de-8417-5bc3d08d6ff1" path="/var/lib/kubelet/pods/8f2677aa-a856-47de-8417-5bc3d08d6ff1/volumes" Dec 05 07:24:07 crc kubenswrapper[4997]: I1205 07:24:07.093152 4997 scope.go:117] "RemoveContainer" containerID="3a4f08d0ae39c150584948dc3f19fc684275263ada8f51f39f56460972b674d2" Dec 05 07:24:07 crc kubenswrapper[4997]: I1205 07:24:07.125123 4997 scope.go:117] "RemoveContainer" containerID="b7780da612c7177ca63143d557d61798aed63fe6008628f0f9c5e7fba16024f1" Dec 05 07:24:07 crc kubenswrapper[4997]: I1205 07:24:07.175415 4997 scope.go:117] "RemoveContainer" containerID="c77fadb14da56ec3ffb9023a9bbf4752c6ab4d4d2977ec528e7ff29efb0be39c" Dec 05 07:24:07 crc kubenswrapper[4997]: I1205 07:24:07.198915 4997 scope.go:117] "RemoveContainer" containerID="3caac945976009c44a6d871779a90a30a1333a9614f3564c29ec3e0c616fdbe7" Dec 05 07:24:07 crc kubenswrapper[4997]: I1205 07:24:07.222358 4997 scope.go:117] "RemoveContainer" containerID="711f304ea598c5aa2085e471f825cc1a9ccb2da36ded6e28c57872ee1771bbfd" Dec 05 07:24:07 crc kubenswrapper[4997]: I1205 07:24:07.244562 4997 scope.go:117] "RemoveContainer" containerID="5d79e06e59e6f9cd89fbe9c40174b9a6f2059177958ca28bb43f3b31b4fdf3bd" Dec 05 07:24:07 crc kubenswrapper[4997]: I1205 07:24:07.293227 4997 scope.go:117] "RemoveContainer" containerID="b92bd2595a4bc2bb590760b577f8f2341220a48a3841768940be448c3738ff62" Dec 05 07:24:10 crc kubenswrapper[4997]: I1205 07:24:10.749949 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:24:10 crc kubenswrapper[4997]: E1205 07:24:10.750885 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:24:21 crc kubenswrapper[4997]: I1205 07:24:21.749431 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:24:21 crc kubenswrapper[4997]: E1205 07:24:21.750366 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:24:35 crc kubenswrapper[4997]: I1205 07:24:35.749671 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:24:35 crc kubenswrapper[4997]: E1205 07:24:35.750657 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:24:47 crc kubenswrapper[4997]: I1205 07:24:47.749114 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:24:47 crc kubenswrapper[4997]: E1205 07:24:47.750162 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:24:58 crc kubenswrapper[4997]: I1205 07:24:58.750286 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:24:58 crc kubenswrapper[4997]: E1205 07:24:58.751660 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:25:07 crc kubenswrapper[4997]: I1205 07:25:07.733380 4997 scope.go:117] "RemoveContainer" containerID="fc45b8c8dbde8e0944b28f364554fb8e350727d01cdcddc568ede9ce3f181336" Dec 05 07:25:07 crc kubenswrapper[4997]: I1205 07:25:07.765866 4997 scope.go:117] "RemoveContainer" containerID="d59ccec3588a99dd3d41c828fb0ed5ed9b64b9c1c576b222b7fac17148e89628" Dec 05 07:25:07 crc kubenswrapper[4997]: I1205 07:25:07.800438 4997 scope.go:117] "RemoveContainer" containerID="ad9f0fd9866254fcb3d271814a479f3ce3ebdaecf8def2eb7a15d7c6654387e1" Dec 05 07:25:07 crc kubenswrapper[4997]: I1205 07:25:07.824587 4997 scope.go:117] "RemoveContainer" containerID="4d6b87afcee18a48538358bac1dcd98772c8ae50d88aed51036e1d032a49b121" Dec 05 07:25:07 crc kubenswrapper[4997]: I1205 07:25:07.852551 4997 scope.go:117] "RemoveContainer" containerID="c9360bc2fe85b0fb890b926fa67438af8aa992628a1b4a88b705b4c8f00b90ea" Dec 05 07:25:07 crc kubenswrapper[4997]: I1205 07:25:07.876392 4997 scope.go:117] "RemoveContainer" containerID="d25c398bc81fe653079c2d969c3c9982a9b528001068fecd4ced24e81fd5d86e" Dec 05 07:25:07 crc kubenswrapper[4997]: I1205 07:25:07.901546 4997 scope.go:117] "RemoveContainer" containerID="2c9db8fa53e1305d4cb7a6121c6b3932ba2fb53b513d58ec4d556042d1e925ab" Dec 05 07:25:07 crc kubenswrapper[4997]: I1205 07:25:07.931976 4997 scope.go:117] "RemoveContainer" containerID="d02644e37a2cd2654ffa8c46e55402e1791c721b12088c2864e45e521fffdba0" Dec 05 07:25:07 crc kubenswrapper[4997]: I1205 07:25:07.968793 4997 scope.go:117] "RemoveContainer" containerID="4ec3032847e09b12c6c9fedc21149ebbebb9600b8ed37fb546deca7c3f8f58c7" Dec 05 07:25:08 crc kubenswrapper[4997]: I1205 07:25:08.005664 4997 scope.go:117] "RemoveContainer" containerID="bd15ae01a5d3986a7e4425b35e0fa45dbfda1aea5cdf404c87ede67fab61b69c" Dec 05 07:25:08 crc kubenswrapper[4997]: I1205 07:25:08.027044 4997 scope.go:117] "RemoveContainer" containerID="abdd9ffacada164260fbd3ad6870541d93f6c0272c1c8ff72270c8ed33bcbe2d" Dec 05 07:25:08 crc kubenswrapper[4997]: I1205 07:25:08.047021 4997 scope.go:117] "RemoveContainer" containerID="5e883daffbf9da9e1ebaee0593862fba2d57d7337e7c8104926342e03ff1d1d6" Dec 05 07:25:08 crc kubenswrapper[4997]: I1205 07:25:08.079239 4997 scope.go:117] "RemoveContainer" containerID="643884fd8f6b17c03dc7c861985fbb559e530326ef051b82964745eadd0773de" Dec 05 07:25:08 crc kubenswrapper[4997]: I1205 07:25:08.099375 4997 scope.go:117] "RemoveContainer" containerID="7e3740ba6a3041381dab8f67d89402db7bd8d90cf3554259e64bd2cbc184a290" Dec 05 07:25:08 crc kubenswrapper[4997]: I1205 07:25:08.119468 4997 scope.go:117] "RemoveContainer" containerID="92b3b79042a60525a1139389eba2978e16aefef7e8b713c4e759357afa9f7028" Dec 05 07:25:08 crc kubenswrapper[4997]: I1205 07:25:08.145382 4997 scope.go:117] "RemoveContainer" containerID="ebb950a103ad8b9026183650a563fd260355285301d70a1c13b4d592f483a882" Dec 05 07:25:08 crc kubenswrapper[4997]: I1205 07:25:08.168514 4997 scope.go:117] "RemoveContainer" containerID="cc518cfb3d28921fd9f53b3004b269e3e22458213667c45944046eb579e9e955" Dec 05 07:25:08 crc kubenswrapper[4997]: I1205 07:25:08.190421 4997 scope.go:117] "RemoveContainer" containerID="5f739def7163631384233edc764a3f8a516103dbcdc15f0624570994d94c96c8" Dec 05 07:25:08 crc kubenswrapper[4997]: I1205 07:25:08.212491 4997 scope.go:117] "RemoveContainer" containerID="4f26581bd5f07af968fd994ea8a57ae49455eb186f958e5e52364d251398e21c" Dec 05 07:25:09 crc kubenswrapper[4997]: I1205 07:25:09.753752 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:25:09 crc kubenswrapper[4997]: E1205 07:25:09.754088 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:25:23 crc kubenswrapper[4997]: I1205 07:25:23.749804 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:25:23 crc kubenswrapper[4997]: E1205 07:25:23.751160 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:25:37 crc kubenswrapper[4997]: I1205 07:25:37.750053 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:25:37 crc kubenswrapper[4997]: E1205 07:25:37.751533 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:25:51 crc kubenswrapper[4997]: I1205 07:25:51.749824 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:25:51 crc kubenswrapper[4997]: E1205 07:25:51.750993 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:26:02 crc kubenswrapper[4997]: I1205 07:26:02.749658 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:26:02 crc kubenswrapper[4997]: E1205 07:26:02.750470 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.475599 4997 scope.go:117] "RemoveContainer" containerID="161dca13d9cb4be2264cec96bcf12ea766cba1f66ad53b3908409e4612cc7f59" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.520291 4997 scope.go:117] "RemoveContainer" containerID="f82f717242295d98f71aec9f7fc80f2c5f943a9da4e85ec393c1fd2c9ccd37ab" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.580345 4997 scope.go:117] "RemoveContainer" containerID="4130ec716a6c5c2d01f70c3a03e4d4ebc152a16367604a994406cb6eda7060ce" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.621086 4997 scope.go:117] "RemoveContainer" containerID="697cbd204f70ac1631f244a00c193d1f66b8976e7fb561cf00fe1afe93eb3bb0" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.646917 4997 scope.go:117] "RemoveContainer" containerID="ba7342d2d5bfb967137e1fd9c51e48f142b7847eec802846b07240462ab579d3" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.674287 4997 scope.go:117] "RemoveContainer" containerID="5c7c977eca43b417984db669a02fedc758e9ee1c5a3cc1077d207c4c29b0e85d" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.694605 4997 scope.go:117] "RemoveContainer" containerID="8ebc72dc3d50298609d59b368904d6395e889e7151f99e044fa28283008de46e" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.713204 4997 scope.go:117] "RemoveContainer" containerID="0f86635af561a3a610b8f8d117854e4f529405ad3fb7080d7a0116d8b95d0876" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.750132 4997 scope.go:117] "RemoveContainer" containerID="2c33de8cc56e9f49bb17a4c545e112e59d5b817d9e70ca54e9b46c79c79014ba" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.767691 4997 scope.go:117] "RemoveContainer" containerID="512330c0d21f3df416e8eea2323fc059b60383128779e8e043be108be9438d25" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.789124 4997 scope.go:117] "RemoveContainer" containerID="ecc9c395cc572c039778c1c4acd19fee3798c7ced60d872d9daba5d55a266d23" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.814416 4997 scope.go:117] "RemoveContainer" containerID="d14f08fb852081fdc4815d2ce697e3418dfd98d7534ab2afb3c35a5bfea2457a" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.836251 4997 scope.go:117] "RemoveContainer" containerID="08946f5deeba339b1b854496b842f982af5b2b54721ebee20c813990c70fbe66" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.857722 4997 scope.go:117] "RemoveContainer" containerID="32e855f8095dea6a97466b7d110e63c5065aa57883dd146dc7d5cd8bf980f1a0" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.879803 4997 scope.go:117] "RemoveContainer" containerID="af3bd6a46ce774d4927e32dd1f7727318d24638c24860778360780d82ef22b47" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.897972 4997 scope.go:117] "RemoveContainer" containerID="8a63d40925c22c61026b0de2d63c2a59394ad5056824d09c3f51cc7b991b8313" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.941837 4997 scope.go:117] "RemoveContainer" containerID="b77018d13ada4b2350299017976f4a2b5e3d76cc39ecbc054f294f67cbfa4f8f" Dec 05 07:26:08 crc kubenswrapper[4997]: I1205 07:26:08.983914 4997 scope.go:117] "RemoveContainer" containerID="2d0cd2f22990c11fe1120f60bcd9583801b4644625f0d7fb00ad15a10ace632b" Dec 05 07:26:09 crc kubenswrapper[4997]: I1205 07:26:09.046256 4997 scope.go:117] "RemoveContainer" containerID="63372a7e68a9d93e249283016625478d1f7e92640812a3890997060ed1267282" Dec 05 07:26:15 crc kubenswrapper[4997]: I1205 07:26:15.749499 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:26:15 crc kubenswrapper[4997]: E1205 07:26:15.750682 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:26:28 crc kubenswrapper[4997]: I1205 07:26:28.749833 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:26:28 crc kubenswrapper[4997]: E1205 07:26:28.752202 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:26:40 crc kubenswrapper[4997]: I1205 07:26:40.749119 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:26:40 crc kubenswrapper[4997]: E1205 07:26:40.750054 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:26:53 crc kubenswrapper[4997]: I1205 07:26:53.750079 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:26:53 crc kubenswrapper[4997]: E1205 07:26:53.751433 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:27:06 crc kubenswrapper[4997]: I1205 07:27:06.750041 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:27:06 crc kubenswrapper[4997]: E1205 07:27:06.751074 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:27:09 crc kubenswrapper[4997]: I1205 07:27:09.324492 4997 scope.go:117] "RemoveContainer" containerID="025cfd66536602dc692bb16ce499c3e7d95d59c6864c476cec229bbace2c59d3" Dec 05 07:27:09 crc kubenswrapper[4997]: I1205 07:27:09.357930 4997 scope.go:117] "RemoveContainer" containerID="5eaf0178fac7b663a1feb1d79be38e9e15c37e90ce1a021fc1f087e1974f2fa5" Dec 05 07:27:09 crc kubenswrapper[4997]: I1205 07:27:09.385972 4997 scope.go:117] "RemoveContainer" containerID="49dec3f8ec745fed264ce35f45332d233e06748c10111467fac5a3e086c7c245" Dec 05 07:27:09 crc kubenswrapper[4997]: I1205 07:27:09.436221 4997 scope.go:117] "RemoveContainer" containerID="0b1d64b7f2fccb29886cde4842d5d80e274474638cdc49f408b3dfb215f85fd4" Dec 05 07:27:09 crc kubenswrapper[4997]: I1205 07:27:09.475001 4997 scope.go:117] "RemoveContainer" containerID="5624c5a4633297b9a0b3880002a7fb4866e561e1bf970698aec5d0b949bbfcc6" Dec 05 07:27:09 crc kubenswrapper[4997]: I1205 07:27:09.498913 4997 scope.go:117] "RemoveContainer" containerID="dce22783874c237d14eefd9cc90fd2f082fc0d59392932b78b0131a8307d7d6b" Dec 05 07:27:09 crc kubenswrapper[4997]: I1205 07:27:09.524124 4997 scope.go:117] "RemoveContainer" containerID="3f78e02557266ee9488a60e7e7d398025ad709b887c5b297cbc3363b33225525" Dec 05 07:27:09 crc kubenswrapper[4997]: I1205 07:27:09.585349 4997 scope.go:117] "RemoveContainer" containerID="3fac4170eed53a3f784a00877568373b3cd58e632ea290c38068fa431ff21b48" Dec 05 07:27:09 crc kubenswrapper[4997]: I1205 07:27:09.610655 4997 scope.go:117] "RemoveContainer" containerID="8f71ef8ac8eb6da4977035d50f2e232a8ea7cd965854f2b988cfb9cf8185b3a4" Dec 05 07:27:09 crc kubenswrapper[4997]: I1205 07:27:09.637175 4997 scope.go:117] "RemoveContainer" containerID="6f749a47e57e47c7006263ba4386dbf5b34d5aa77a620b464309e9c5d820056b" Dec 05 07:27:09 crc kubenswrapper[4997]: I1205 07:27:09.660706 4997 scope.go:117] "RemoveContainer" containerID="d09f8f75d78ec8de394fe2d3878fbdc0a819558c405bcad43068ea535ea0a421" Dec 05 07:27:09 crc kubenswrapper[4997]: I1205 07:27:09.683432 4997 scope.go:117] "RemoveContainer" containerID="793a4abf63252321ba4fb0dafb87c02e599c7ee609d534abf01e11eb3eff2e86" Dec 05 07:27:09 crc kubenswrapper[4997]: I1205 07:27:09.705696 4997 scope.go:117] "RemoveContainer" containerID="d4b6f46bb0d9eab2ada6b1349324da20be55e9bbe3ce62bb67888afe1b1a469c" Dec 05 07:27:18 crc kubenswrapper[4997]: I1205 07:27:18.750210 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:27:18 crc kubenswrapper[4997]: E1205 07:27:18.751701 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:27:32 crc kubenswrapper[4997]: I1205 07:27:32.749864 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:27:32 crc kubenswrapper[4997]: E1205 07:27:32.750661 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:27:44 crc kubenswrapper[4997]: I1205 07:27:44.749647 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:27:44 crc kubenswrapper[4997]: E1205 07:27:44.750805 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:27:55 crc kubenswrapper[4997]: I1205 07:27:55.749370 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:27:56 crc kubenswrapper[4997]: I1205 07:27:56.135476 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"67fe830b9e170b36bd5960d524375aebc16b702b52c6551ed6fc753345c6ff89"} Dec 05 07:28:09 crc kubenswrapper[4997]: I1205 07:28:09.900198 4997 scope.go:117] "RemoveContainer" containerID="6c7e34087d4bfaea3711d104c62c9dd6772529963b74b9d73df375927d1f000e" Dec 05 07:28:09 crc kubenswrapper[4997]: I1205 07:28:09.949504 4997 scope.go:117] "RemoveContainer" containerID="b81d606450f4ec2753efb3ba9b01b31144b20d257dc2ed0c1a6d4b204cf562b4" Dec 05 07:28:09 crc kubenswrapper[4997]: I1205 07:28:09.973550 4997 scope.go:117] "RemoveContainer" containerID="8262a289c6746a6163cf170184afb1a33d7c8739d56d972fe6ac01db023fe654" Dec 05 07:28:09 crc kubenswrapper[4997]: I1205 07:28:09.994520 4997 scope.go:117] "RemoveContainer" containerID="159ce5f13204b37468ae891f1ae556c7d3978744ca70fe3f088c098ee1f00ac9" Dec 05 07:28:10 crc kubenswrapper[4997]: I1205 07:28:10.048949 4997 scope.go:117] "RemoveContainer" containerID="bd53b88774db5676c0ecbd766a894c0787029e64044e69e6c842f108a8b45f92" Dec 05 07:29:10 crc kubenswrapper[4997]: I1205 07:29:10.136838 4997 scope.go:117] "RemoveContainer" containerID="96ac136074ed0bcb85702f5bfe22b13a7895276ce80e0413118cb761bfafcad0" Dec 05 07:29:10 crc kubenswrapper[4997]: I1205 07:29:10.170963 4997 scope.go:117] "RemoveContainer" containerID="e1a5fbc1a91ff3e0d6d191e2796f352f99e1dd7df658b4c7a2370f2df1211989" Dec 05 07:29:10 crc kubenswrapper[4997]: I1205 07:29:10.202798 4997 scope.go:117] "RemoveContainer" containerID="db7397da7d823a2ab29a061ca9792ba46bd6e1c49ac1ae528d2649ad3365c16a" Dec 05 07:29:10 crc kubenswrapper[4997]: I1205 07:29:10.239529 4997 scope.go:117] "RemoveContainer" containerID="27d910165cbca8f6eb45e1a947fe3daf0e9d27a117dff110f3fc7aeef2a06961" Dec 05 07:29:10 crc kubenswrapper[4997]: I1205 07:29:10.276320 4997 scope.go:117] "RemoveContainer" containerID="96f6c22ff0cd4cb36b8fadc88840e34b4298c41452961e008b5dd7fc23ef4889" Dec 05 07:29:10 crc kubenswrapper[4997]: I1205 07:29:10.307909 4997 scope.go:117] "RemoveContainer" containerID="834d649b80dd8fee860f1e640b0ea6a36d58f4fc6fd2c37271b4e7ce17650c8d" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.153654 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl"] Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.154979 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d65cd492-d4bd-4489-9d28-a6c42407101a" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155003 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d65cd492-d4bd-4489-9d28-a6c42407101a" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155050 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-server" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155059 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-server" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155072 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" containerName="setup-container" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155080 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" containerName="setup-container" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155091 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c706bba7-965c-497f-ae7d-b7087d37d70e" containerName="nova-cell0-conductor-conductor" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155100 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c706bba7-965c-497f-ae7d-b7087d37d70e" containerName="nova-cell0-conductor-conductor" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155113 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovs-vswitchd" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155121 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovs-vswitchd" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155138 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6bfb8b2-df28-4992-8831-2a5060323ddf" containerName="barbican-worker" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155148 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6bfb8b2-df28-4992-8831-2a5060323ddf" containerName="barbican-worker" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155160 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c835ac5d-bd54-46d4-aab3-3ef7682f50b7" containerName="neutron-httpd" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155167 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c835ac5d-bd54-46d4-aab3-3ef7682f50b7" containerName="neutron-httpd" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155183 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5afeb230-4a35-4872-bffa-c882fe927643" containerName="nova-api-log" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155191 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5afeb230-4a35-4872-bffa-c882fe927643" containerName="nova-api-log" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155202 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db136058-17a8-4955-8e95-576da7b9e847" containerName="proxy-httpd" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155209 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="db136058-17a8-4955-8e95-576da7b9e847" containerName="proxy-httpd" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155224 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="account-auditor" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155232 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="account-auditor" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155243 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="container-auditor" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155251 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="container-auditor" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155266 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f158874-152b-46ce-ac42-d202684853ca" containerName="setup-container" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155275 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f158874-152b-46ce-ac42-d202684853ca" containerName="setup-container" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155289 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef43a612-a3df-4b50-9bf3-e5d8098cd0fe" containerName="memcached" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155297 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef43a612-a3df-4b50-9bf3-e5d8098cd0fe" containerName="memcached" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155305 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-expirer" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155313 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-expirer" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155353 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d9e01d-5189-4aac-8f1a-b1f09fe138b8" containerName="glance-log" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155363 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d9e01d-5189-4aac-8f1a-b1f09fe138b8" containerName="glance-log" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155379 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db136058-17a8-4955-8e95-576da7b9e847" containerName="sg-core" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155386 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="db136058-17a8-4955-8e95-576da7b9e847" containerName="sg-core" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155402 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovsdb-server-init" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155411 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovsdb-server-init" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155424 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e08a39e3-f267-496d-80b3-b12a9eef14c1" containerName="glance-log" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155432 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e08a39e3-f267-496d-80b3-b12a9eef14c1" containerName="glance-log" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155448 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-auditor" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155456 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-auditor" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155467 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d9e01d-5189-4aac-8f1a-b1f09fe138b8" containerName="glance-httpd" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155474 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d9e01d-5189-4aac-8f1a-b1f09fe138b8" containerName="glance-httpd" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155486 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8c5aa81-a666-4ae6-a41c-fd31df260223" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155494 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8c5aa81-a666-4ae6-a41c-fd31df260223" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155509 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="container-replicator" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155517 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="container-replicator" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155531 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" containerName="rabbitmq" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155539 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" containerName="rabbitmq" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155553 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0601e553-5305-4f54-9a2d-43b43a1a5dd4" containerName="mysql-bootstrap" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155561 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0601e553-5305-4f54-9a2d-43b43a1a5dd4" containerName="mysql-bootstrap" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155575 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" containerName="ovn-controller" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155583 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" containerName="ovn-controller" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155593 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="account-reaper" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155603 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="account-reaper" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155632 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-replicator" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155642 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-replicator" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155660 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="account-server" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155669 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="account-server" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155683 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5afeb230-4a35-4872-bffa-c882fe927643" containerName="nova-api-api" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155692 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5afeb230-4a35-4872-bffa-c882fe927643" containerName="nova-api-api" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155704 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="rsync" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155712 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="rsync" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155728 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c835ac5d-bd54-46d4-aab3-3ef7682f50b7" containerName="neutron-api" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155737 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c835ac5d-bd54-46d4-aab3-3ef7682f50b7" containerName="neutron-api" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155747 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcbe26f4-fcd9-46c7-aa3a-5038b801741d" containerName="nova-metadata-log" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155755 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcbe26f4-fcd9-46c7-aa3a-5038b801741d" containerName="nova-metadata-log" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155771 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf72e65a-bae4-4ce4-b65a-d5a587ef5f48" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155779 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf72e65a-bae4-4ce4-b65a-d5a587ef5f48" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155794 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db136058-17a8-4955-8e95-576da7b9e847" containerName="ceilometer-central-agent" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155803 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="db136058-17a8-4955-8e95-576da7b9e847" containerName="ceilometer-central-agent" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155814 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="swift-recon-cron" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155822 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="swift-recon-cron" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155834 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="container-server" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155841 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="container-server" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155850 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2179d060-8f4c-413d-a202-e31ebc242dfc" containerName="barbican-api" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155857 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2179d060-8f4c-413d-a202-e31ebc242dfc" containerName="barbican-api" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155873 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="account-replicator" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155880 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="account-replicator" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155895 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2179d060-8f4c-413d-a202-e31ebc242dfc" containerName="barbican-api-log" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155903 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2179d060-8f4c-413d-a202-e31ebc242dfc" containerName="barbican-api-log" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155914 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0601e553-5305-4f54-9a2d-43b43a1a5dd4" containerName="galera" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155921 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0601e553-5305-4f54-9a2d-43b43a1a5dd4" containerName="galera" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155934 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-updater" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155940 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-updater" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155955 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="container-updater" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155962 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="container-updater" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155973 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6783d79d-4874-43c1-92bc-e09133e45989" containerName="keystone-api" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155981 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6783d79d-4874-43c1-92bc-e09133e45989" containerName="keystone-api" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.155990 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13fc4b26-4375-4ea2-8eaa-2417d64b3cd1" containerName="barbican-keystone-listener-log" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.155997 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="13fc4b26-4375-4ea2-8eaa-2417d64b3cd1" containerName="barbican-keystone-listener-log" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.156010 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f2677aa-a856-47de-8417-5bc3d08d6ff1" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156017 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f2677aa-a856-47de-8417-5bc3d08d6ff1" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.156030 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946" containerName="kube-state-metrics" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156038 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946" containerName="kube-state-metrics" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.156051 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed" containerName="nova-scheduler-scheduler" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156059 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed" containerName="nova-scheduler-scheduler" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.156068 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a733d85-ad49-41b8-a75b-842dce56e85c" containerName="placement-api" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156077 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a733d85-ad49-41b8-a75b-842dce56e85c" containerName="placement-api" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.156087 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c51acefb-ed8a-480c-9ab6-0345f2ff0cd3" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156095 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c51acefb-ed8a-480c-9ab6-0345f2ff0cd3" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.156104 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68490a68-dc32-448f-90cc-c146ba49c1a8" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156111 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="68490a68-dc32-448f-90cc-c146ba49c1a8" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.156120 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcbe26f4-fcd9-46c7-aa3a-5038b801741d" containerName="nova-metadata-metadata" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156127 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcbe26f4-fcd9-46c7-aa3a-5038b801741d" containerName="nova-metadata-metadata" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.156137 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovsdb-server" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156151 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovsdb-server" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.156160 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f158874-152b-46ce-ac42-d202684853ca" containerName="rabbitmq" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156167 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f158874-152b-46ce-ac42-d202684853ca" containerName="rabbitmq" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.156176 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5edac3e-6bee-494f-97ba-c4047064d02f" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156184 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5edac3e-6bee-494f-97ba-c4047064d02f" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.156195 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e08a39e3-f267-496d-80b3-b12a9eef14c1" containerName="glance-httpd" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156202 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e08a39e3-f267-496d-80b3-b12a9eef14c1" containerName="glance-httpd" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.156212 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db136058-17a8-4955-8e95-576da7b9e847" containerName="ceilometer-notification-agent" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156221 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="db136058-17a8-4955-8e95-576da7b9e847" containerName="ceilometer-notification-agent" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.156232 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5edac3e-6bee-494f-97ba-c4047064d02f" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156240 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5edac3e-6bee-494f-97ba-c4047064d02f" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.156253 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a733d85-ad49-41b8-a75b-842dce56e85c" containerName="placement-log" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156261 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a733d85-ad49-41b8-a75b-842dce56e85c" containerName="placement-log" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.156270 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6bfb8b2-df28-4992-8831-2a5060323ddf" containerName="barbican-worker-log" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156278 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6bfb8b2-df28-4992-8831-2a5060323ddf" containerName="barbican-worker-log" Dec 05 07:30:00 crc kubenswrapper[4997]: E1205 07:30:00.156288 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13fc4b26-4375-4ea2-8eaa-2417d64b3cd1" containerName="barbican-keystone-listener" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156296 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="13fc4b26-4375-4ea2-8eaa-2417d64b3cd1" containerName="barbican-keystone-listener" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156485 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="account-auditor" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156497 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="13fc4b26-4375-4ea2-8eaa-2417d64b3cd1" containerName="barbican-keystone-listener" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156507 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-expirer" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156515 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a733d85-ad49-41b8-a75b-842dce56e85c" containerName="placement-log" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156523 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a733d85-ad49-41b8-a75b-842dce56e85c" containerName="placement-api" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156533 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f2677aa-a856-47de-8417-5bc3d08d6ff1" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156542 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ec01dcc-e42e-4ad1-bafb-fd9284a3ec79" containerName="rabbitmq" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156548 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="swift-recon-cron" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156557 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e08a39e3-f267-496d-80b3-b12a9eef14c1" containerName="glance-httpd" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156568 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="db136058-17a8-4955-8e95-576da7b9e847" containerName="ceilometer-central-agent" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156578 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="account-replicator" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156586 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="account-reaper" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156595 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0601e553-5305-4f54-9a2d-43b43a1a5dd4" containerName="galera" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156604 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="db136058-17a8-4955-8e95-576da7b9e847" containerName="ceilometer-notification-agent" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156636 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f158874-152b-46ce-ac42-d202684853ca" containerName="rabbitmq" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156649 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6bfb8b2-df28-4992-8831-2a5060323ddf" containerName="barbican-worker" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156658 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-replicator" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156667 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8c5aa81-a666-4ae6-a41c-fd31df260223" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156681 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcbe26f4-fcd9-46c7-aa3a-5038b801741d" containerName="nova-metadata-metadata" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156692 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovsdb-server" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156700 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5388a6f5-ecde-42db-9c94-a52c17e8a302" containerName="ovs-vswitchd" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156708 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5edac3e-6bee-494f-97ba-c4047064d02f" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156715 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-auditor" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156722 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-updater" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156731 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf72e65a-bae4-4ce4-b65a-d5a587ef5f48" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156743 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e08a39e3-f267-496d-80b3-b12a9eef14c1" containerName="glance-log" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156754 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6783d79d-4874-43c1-92bc-e09133e45989" containerName="keystone-api" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156763 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d9e01d-5189-4aac-8f1a-b1f09fe138b8" containerName="glance-httpd" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156773 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="c835ac5d-bd54-46d4-aab3-3ef7682f50b7" containerName="neutron-httpd" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156782 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="9efeea50-b18e-4cbe-b4b6-f5af06d9f8ed" containerName="nova-scheduler-scheduler" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156792 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcbe26f4-fcd9-46c7-aa3a-5038b801741d" containerName="nova-metadata-log" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156801 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="c835ac5d-bd54-46d4-aab3-3ef7682f50b7" containerName="neutron-api" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156810 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="68490a68-dc32-448f-90cc-c146ba49c1a8" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156819 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="container-replicator" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156830 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="db136058-17a8-4955-8e95-576da7b9e847" containerName="sg-core" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156840 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="13fc4b26-4375-4ea2-8eaa-2417d64b3cd1" containerName="barbican-keystone-listener-log" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156852 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5afeb230-4a35-4872-bffa-c882fe927643" containerName="nova-api-api" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156866 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="rsync" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156873 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2179d060-8f4c-413d-a202-e31ebc242dfc" containerName="barbican-api-log" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156880 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e4237cc-8581-4ddb-b472-e1c6a4a41cb5" containerName="ovn-controller" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156889 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="c706bba7-965c-497f-ae7d-b7087d37d70e" containerName="nova-cell0-conductor-conductor" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156898 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="container-server" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156906 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="c51acefb-ed8a-480c-9ab6-0345f2ff0cd3" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156915 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="object-server" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156924 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="db136058-17a8-4955-8e95-576da7b9e847" containerName="proxy-httpd" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156934 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d9e01d-5189-4aac-8f1a-b1f09fe138b8" containerName="glance-log" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156947 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="container-updater" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156956 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="account-server" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156966 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7f9ee20-f544-4c9f-ae7c-c65cd6a7b946" containerName="kube-state-metrics" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156974 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6bfb8b2-df28-4992-8831-2a5060323ddf" containerName="barbican-worker-log" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156984 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef43a612-a3df-4b50-9bf3-e5d8098cd0fe" containerName="memcached" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.156992 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2179d060-8f4c-413d-a202-e31ebc242dfc" containerName="barbican-api" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.157001 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="d65cd492-d4bd-4489-9d28-a6c42407101a" containerName="mariadb-account-delete" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.157008 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc70c62-ad3e-4178-932a-d14571a06daa" containerName="container-auditor" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.157015 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5afeb230-4a35-4872-bffa-c882fe927643" containerName="nova-api-log" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.157596 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.161031 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.168566 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.190656 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl"] Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.330744 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fns9p\" (UniqueName: \"kubernetes.io/projected/8f3272ac-8f40-4c96-95b0-31b7e78e6318-kube-api-access-fns9p\") pod \"collect-profiles-29415330-h9pcl\" (UID: \"8f3272ac-8f40-4c96-95b0-31b7e78e6318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.330850 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8f3272ac-8f40-4c96-95b0-31b7e78e6318-secret-volume\") pod \"collect-profiles-29415330-h9pcl\" (UID: \"8f3272ac-8f40-4c96-95b0-31b7e78e6318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.331877 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8f3272ac-8f40-4c96-95b0-31b7e78e6318-config-volume\") pod \"collect-profiles-29415330-h9pcl\" (UID: \"8f3272ac-8f40-4c96-95b0-31b7e78e6318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.433967 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fns9p\" (UniqueName: \"kubernetes.io/projected/8f3272ac-8f40-4c96-95b0-31b7e78e6318-kube-api-access-fns9p\") pod \"collect-profiles-29415330-h9pcl\" (UID: \"8f3272ac-8f40-4c96-95b0-31b7e78e6318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.434066 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8f3272ac-8f40-4c96-95b0-31b7e78e6318-secret-volume\") pod \"collect-profiles-29415330-h9pcl\" (UID: \"8f3272ac-8f40-4c96-95b0-31b7e78e6318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.434119 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8f3272ac-8f40-4c96-95b0-31b7e78e6318-config-volume\") pod \"collect-profiles-29415330-h9pcl\" (UID: \"8f3272ac-8f40-4c96-95b0-31b7e78e6318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.435237 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8f3272ac-8f40-4c96-95b0-31b7e78e6318-config-volume\") pod \"collect-profiles-29415330-h9pcl\" (UID: \"8f3272ac-8f40-4c96-95b0-31b7e78e6318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.451390 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8f3272ac-8f40-4c96-95b0-31b7e78e6318-secret-volume\") pod \"collect-profiles-29415330-h9pcl\" (UID: \"8f3272ac-8f40-4c96-95b0-31b7e78e6318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.453046 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fns9p\" (UniqueName: \"kubernetes.io/projected/8f3272ac-8f40-4c96-95b0-31b7e78e6318-kube-api-access-fns9p\") pod \"collect-profiles-29415330-h9pcl\" (UID: \"8f3272ac-8f40-4c96-95b0-31b7e78e6318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.481834 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl" Dec 05 07:30:00 crc kubenswrapper[4997]: I1205 07:30:00.729629 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl"] Dec 05 07:30:01 crc kubenswrapper[4997]: I1205 07:30:01.497629 4997 generic.go:334] "Generic (PLEG): container finished" podID="8f3272ac-8f40-4c96-95b0-31b7e78e6318" containerID="224eab4b424ce3861377ade439c8d1bea513ab185533a7173886823a2fa6936e" exitCode=0 Dec 05 07:30:01 crc kubenswrapper[4997]: I1205 07:30:01.497778 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl" event={"ID":"8f3272ac-8f40-4c96-95b0-31b7e78e6318","Type":"ContainerDied","Data":"224eab4b424ce3861377ade439c8d1bea513ab185533a7173886823a2fa6936e"} Dec 05 07:30:01 crc kubenswrapper[4997]: I1205 07:30:01.498088 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl" event={"ID":"8f3272ac-8f40-4c96-95b0-31b7e78e6318","Type":"ContainerStarted","Data":"6ccd9abb2dca471664d07efc01eaf05a34d2d8b73ea0df2ed8c359d64f850ae5"} Dec 05 07:30:02 crc kubenswrapper[4997]: I1205 07:30:02.832521 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl" Dec 05 07:30:02 crc kubenswrapper[4997]: I1205 07:30:02.878435 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8f3272ac-8f40-4c96-95b0-31b7e78e6318-secret-volume\") pod \"8f3272ac-8f40-4c96-95b0-31b7e78e6318\" (UID: \"8f3272ac-8f40-4c96-95b0-31b7e78e6318\") " Dec 05 07:30:02 crc kubenswrapper[4997]: I1205 07:30:02.879922 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8f3272ac-8f40-4c96-95b0-31b7e78e6318-config-volume\") pod \"8f3272ac-8f40-4c96-95b0-31b7e78e6318\" (UID: \"8f3272ac-8f40-4c96-95b0-31b7e78e6318\") " Dec 05 07:30:02 crc kubenswrapper[4997]: I1205 07:30:02.880085 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fns9p\" (UniqueName: \"kubernetes.io/projected/8f3272ac-8f40-4c96-95b0-31b7e78e6318-kube-api-access-fns9p\") pod \"8f3272ac-8f40-4c96-95b0-31b7e78e6318\" (UID: \"8f3272ac-8f40-4c96-95b0-31b7e78e6318\") " Dec 05 07:30:02 crc kubenswrapper[4997]: I1205 07:30:02.880973 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f3272ac-8f40-4c96-95b0-31b7e78e6318-config-volume" (OuterVolumeSpecName: "config-volume") pod "8f3272ac-8f40-4c96-95b0-31b7e78e6318" (UID: "8f3272ac-8f40-4c96-95b0-31b7e78e6318"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:30:02 crc kubenswrapper[4997]: I1205 07:30:02.882107 4997 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8f3272ac-8f40-4c96-95b0-31b7e78e6318-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 07:30:02 crc kubenswrapper[4997]: I1205 07:30:02.889552 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f3272ac-8f40-4c96-95b0-31b7e78e6318-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8f3272ac-8f40-4c96-95b0-31b7e78e6318" (UID: "8f3272ac-8f40-4c96-95b0-31b7e78e6318"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:30:02 crc kubenswrapper[4997]: I1205 07:30:02.889607 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f3272ac-8f40-4c96-95b0-31b7e78e6318-kube-api-access-fns9p" (OuterVolumeSpecName: "kube-api-access-fns9p") pod "8f3272ac-8f40-4c96-95b0-31b7e78e6318" (UID: "8f3272ac-8f40-4c96-95b0-31b7e78e6318"). InnerVolumeSpecName "kube-api-access-fns9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:30:02 crc kubenswrapper[4997]: I1205 07:30:02.984478 4997 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8f3272ac-8f40-4c96-95b0-31b7e78e6318-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 07:30:02 crc kubenswrapper[4997]: I1205 07:30:02.984531 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fns9p\" (UniqueName: \"kubernetes.io/projected/8f3272ac-8f40-4c96-95b0-31b7e78e6318-kube-api-access-fns9p\") on node \"crc\" DevicePath \"\"" Dec 05 07:30:03 crc kubenswrapper[4997]: I1205 07:30:03.530959 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl" event={"ID":"8f3272ac-8f40-4c96-95b0-31b7e78e6318","Type":"ContainerDied","Data":"6ccd9abb2dca471664d07efc01eaf05a34d2d8b73ea0df2ed8c359d64f850ae5"} Dec 05 07:30:03 crc kubenswrapper[4997]: I1205 07:30:03.531053 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ccd9abb2dca471664d07efc01eaf05a34d2d8b73ea0df2ed8c359d64f850ae5" Dec 05 07:30:03 crc kubenswrapper[4997]: I1205 07:30:03.531150 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl" Dec 05 07:30:03 crc kubenswrapper[4997]: E1205 07:30:03.691257 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f3272ac_8f40_4c96_95b0_31b7e78e6318.slice/crio-6ccd9abb2dca471664d07efc01eaf05a34d2d8b73ea0df2ed8c359d64f850ae5\": RecentStats: unable to find data in memory cache]" Dec 05 07:30:03 crc kubenswrapper[4997]: I1205 07:30:03.937891 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92"] Dec 05 07:30:03 crc kubenswrapper[4997]: I1205 07:30:03.943664 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415285-28q92"] Dec 05 07:30:05 crc kubenswrapper[4997]: I1205 07:30:05.764855 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f692422b-da9f-4775-825e-ee198e5b61b8" path="/var/lib/kubelet/pods/f692422b-da9f-4775-825e-ee198e5b61b8/volumes" Dec 05 07:30:10 crc kubenswrapper[4997]: I1205 07:30:10.442833 4997 scope.go:117] "RemoveContainer" containerID="84c9f79e02c9e635085a99dbc7faaaf90bfa04701ecbeee66bafb3d0bf7cf96b" Dec 05 07:30:10 crc kubenswrapper[4997]: I1205 07:30:10.467956 4997 scope.go:117] "RemoveContainer" containerID="331b3a19e5c1b6a99d65e10dc3dc2975027b515da4a33c5e24c9c58c4f701c55" Dec 05 07:30:10 crc kubenswrapper[4997]: I1205 07:30:10.517558 4997 scope.go:117] "RemoveContainer" containerID="8a2e07a2742d3964e8caee339bfc34622988235ccfcc24681f0a0082129f5702" Dec 05 07:30:10 crc kubenswrapper[4997]: I1205 07:30:10.553006 4997 scope.go:117] "RemoveContainer" containerID="69e70f6cd56b05b725830aa87f03ff2323fde92fedc1b17a8fc26c0917b692da" Dec 05 07:30:10 crc kubenswrapper[4997]: I1205 07:30:10.586811 4997 scope.go:117] "RemoveContainer" containerID="769b1779b305cc0d585e4dba78fb1ee7190f831ce52ad7c2bc764a5532dc31e6" Dec 05 07:30:10 crc kubenswrapper[4997]: I1205 07:30:10.640387 4997 scope.go:117] "RemoveContainer" containerID="e0d6bff0be26dd0adebe52af15c98c304087e7208454b30644f8e9193c0a2c1f" Dec 05 07:30:19 crc kubenswrapper[4997]: I1205 07:30:19.770978 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:30:19 crc kubenswrapper[4997]: I1205 07:30:19.771803 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:30:49 crc kubenswrapper[4997]: I1205 07:30:49.770717 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:30:49 crc kubenswrapper[4997]: I1205 07:30:49.772055 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:31:19 crc kubenswrapper[4997]: I1205 07:31:19.770233 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:31:19 crc kubenswrapper[4997]: I1205 07:31:19.771210 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:31:19 crc kubenswrapper[4997]: I1205 07:31:19.771302 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 07:31:19 crc kubenswrapper[4997]: I1205 07:31:19.771957 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"67fe830b9e170b36bd5960d524375aebc16b702b52c6551ed6fc753345c6ff89"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 07:31:19 crc kubenswrapper[4997]: I1205 07:31:19.772033 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://67fe830b9e170b36bd5960d524375aebc16b702b52c6551ed6fc753345c6ff89" gracePeriod=600 Dec 05 07:31:20 crc kubenswrapper[4997]: I1205 07:31:20.321667 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="67fe830b9e170b36bd5960d524375aebc16b702b52c6551ed6fc753345c6ff89" exitCode=0 Dec 05 07:31:20 crc kubenswrapper[4997]: I1205 07:31:20.321880 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"67fe830b9e170b36bd5960d524375aebc16b702b52c6551ed6fc753345c6ff89"} Dec 05 07:31:20 crc kubenswrapper[4997]: I1205 07:31:20.322127 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899"} Dec 05 07:31:20 crc kubenswrapper[4997]: I1205 07:31:20.322158 4997 scope.go:117] "RemoveContainer" containerID="7a5c2c8b3c95035f127450d063e2762f89ed7fb384554a453439056bfaf28c08" Dec 05 07:31:58 crc kubenswrapper[4997]: I1205 07:31:58.264042 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tzmkw"] Dec 05 07:31:58 crc kubenswrapper[4997]: E1205 07:31:58.266099 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f3272ac-8f40-4c96-95b0-31b7e78e6318" containerName="collect-profiles" Dec 05 07:31:58 crc kubenswrapper[4997]: I1205 07:31:58.266165 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f3272ac-8f40-4c96-95b0-31b7e78e6318" containerName="collect-profiles" Dec 05 07:31:58 crc kubenswrapper[4997]: I1205 07:31:58.266514 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5edac3e-6bee-494f-97ba-c4047064d02f" containerName="mariadb-account-delete" Dec 05 07:31:58 crc kubenswrapper[4997]: I1205 07:31:58.266540 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f3272ac-8f40-4c96-95b0-31b7e78e6318" containerName="collect-profiles" Dec 05 07:31:58 crc kubenswrapper[4997]: I1205 07:31:58.269309 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tzmkw" Dec 05 07:31:58 crc kubenswrapper[4997]: I1205 07:31:58.293045 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tzmkw"] Dec 05 07:31:58 crc kubenswrapper[4997]: I1205 07:31:58.411326 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-catalog-content\") pod \"redhat-operators-tzmkw\" (UID: \"4958f0cc-3cc3-47e5-a754-0b3ab586c91b\") " pod="openshift-marketplace/redhat-operators-tzmkw" Dec 05 07:31:58 crc kubenswrapper[4997]: I1205 07:31:58.411706 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-utilities\") pod \"redhat-operators-tzmkw\" (UID: \"4958f0cc-3cc3-47e5-a754-0b3ab586c91b\") " pod="openshift-marketplace/redhat-operators-tzmkw" Dec 05 07:31:58 crc kubenswrapper[4997]: I1205 07:31:58.412195 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nttj\" (UniqueName: \"kubernetes.io/projected/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-kube-api-access-7nttj\") pod \"redhat-operators-tzmkw\" (UID: \"4958f0cc-3cc3-47e5-a754-0b3ab586c91b\") " pod="openshift-marketplace/redhat-operators-tzmkw" Dec 05 07:31:58 crc kubenswrapper[4997]: I1205 07:31:58.514569 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-catalog-content\") pod \"redhat-operators-tzmkw\" (UID: \"4958f0cc-3cc3-47e5-a754-0b3ab586c91b\") " pod="openshift-marketplace/redhat-operators-tzmkw" Dec 05 07:31:58 crc kubenswrapper[4997]: I1205 07:31:58.514663 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-utilities\") pod \"redhat-operators-tzmkw\" (UID: \"4958f0cc-3cc3-47e5-a754-0b3ab586c91b\") " pod="openshift-marketplace/redhat-operators-tzmkw" Dec 05 07:31:58 crc kubenswrapper[4997]: I1205 07:31:58.514720 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nttj\" (UniqueName: \"kubernetes.io/projected/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-kube-api-access-7nttj\") pod \"redhat-operators-tzmkw\" (UID: \"4958f0cc-3cc3-47e5-a754-0b3ab586c91b\") " pod="openshift-marketplace/redhat-operators-tzmkw" Dec 05 07:31:58 crc kubenswrapper[4997]: I1205 07:31:58.515373 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-catalog-content\") pod \"redhat-operators-tzmkw\" (UID: \"4958f0cc-3cc3-47e5-a754-0b3ab586c91b\") " pod="openshift-marketplace/redhat-operators-tzmkw" Dec 05 07:31:58 crc kubenswrapper[4997]: I1205 07:31:58.515390 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-utilities\") pod \"redhat-operators-tzmkw\" (UID: \"4958f0cc-3cc3-47e5-a754-0b3ab586c91b\") " pod="openshift-marketplace/redhat-operators-tzmkw" Dec 05 07:31:58 crc kubenswrapper[4997]: I1205 07:31:58.540427 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nttj\" (UniqueName: \"kubernetes.io/projected/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-kube-api-access-7nttj\") pod \"redhat-operators-tzmkw\" (UID: \"4958f0cc-3cc3-47e5-a754-0b3ab586c91b\") " pod="openshift-marketplace/redhat-operators-tzmkw" Dec 05 07:31:58 crc kubenswrapper[4997]: I1205 07:31:58.630987 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tzmkw" Dec 05 07:31:58 crc kubenswrapper[4997]: I1205 07:31:58.905254 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tzmkw"] Dec 05 07:31:59 crc kubenswrapper[4997]: I1205 07:31:59.753142 4997 generic.go:334] "Generic (PLEG): container finished" podID="4958f0cc-3cc3-47e5-a754-0b3ab586c91b" containerID="a3784e1a6dfe29da8902b609c53a23e723f3d3b110bb53ab9ebc036b9f739633" exitCode=0 Dec 05 07:31:59 crc kubenswrapper[4997]: I1205 07:31:59.763285 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 07:31:59 crc kubenswrapper[4997]: I1205 07:31:59.768786 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tzmkw" event={"ID":"4958f0cc-3cc3-47e5-a754-0b3ab586c91b","Type":"ContainerDied","Data":"a3784e1a6dfe29da8902b609c53a23e723f3d3b110bb53ab9ebc036b9f739633"} Dec 05 07:31:59 crc kubenswrapper[4997]: I1205 07:31:59.768835 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tzmkw" event={"ID":"4958f0cc-3cc3-47e5-a754-0b3ab586c91b","Type":"ContainerStarted","Data":"23623063a96eb718e52cc035f01d01bd0cbd18eb4b2921707db721da6205c584"} Dec 05 07:32:00 crc kubenswrapper[4997]: I1205 07:32:00.770101 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tzmkw" event={"ID":"4958f0cc-3cc3-47e5-a754-0b3ab586c91b","Type":"ContainerStarted","Data":"4d40e37af9d79716c9c82e7e54cf520361585d28a544e7aaf599f66655e6816b"} Dec 05 07:32:01 crc kubenswrapper[4997]: I1205 07:32:01.787998 4997 generic.go:334] "Generic (PLEG): container finished" podID="4958f0cc-3cc3-47e5-a754-0b3ab586c91b" containerID="4d40e37af9d79716c9c82e7e54cf520361585d28a544e7aaf599f66655e6816b" exitCode=0 Dec 05 07:32:01 crc kubenswrapper[4997]: I1205 07:32:01.788092 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tzmkw" event={"ID":"4958f0cc-3cc3-47e5-a754-0b3ab586c91b","Type":"ContainerDied","Data":"4d40e37af9d79716c9c82e7e54cf520361585d28a544e7aaf599f66655e6816b"} Dec 05 07:32:02 crc kubenswrapper[4997]: I1205 07:32:02.801179 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tzmkw" event={"ID":"4958f0cc-3cc3-47e5-a754-0b3ab586c91b","Type":"ContainerStarted","Data":"ed3cb7b5695e5740b8c3954c55395276729b0785e57296c588772acdd94a7b0c"} Dec 05 07:32:02 crc kubenswrapper[4997]: I1205 07:32:02.836658 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tzmkw" podStartSLOduration=2.38313604 podStartE2EDuration="4.836600931s" podCreationTimestamp="2025-12-05 07:31:58 +0000 UTC" firstStartedPulling="2025-12-05 07:31:59.762986433 +0000 UTC m=+2220.291893694" lastFinishedPulling="2025-12-05 07:32:02.216451294 +0000 UTC m=+2222.745358585" observedRunningTime="2025-12-05 07:32:02.829107179 +0000 UTC m=+2223.358014540" watchObservedRunningTime="2025-12-05 07:32:02.836600931 +0000 UTC m=+2223.365508212" Dec 05 07:32:08 crc kubenswrapper[4997]: I1205 07:32:08.631898 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tzmkw" Dec 05 07:32:08 crc kubenswrapper[4997]: I1205 07:32:08.632466 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tzmkw" Dec 05 07:32:09 crc kubenswrapper[4997]: I1205 07:32:09.703469 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tzmkw" podUID="4958f0cc-3cc3-47e5-a754-0b3ab586c91b" containerName="registry-server" probeResult="failure" output=< Dec 05 07:32:09 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 07:32:09 crc kubenswrapper[4997]: > Dec 05 07:32:18 crc kubenswrapper[4997]: I1205 07:32:18.683146 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tzmkw" Dec 05 07:32:18 crc kubenswrapper[4997]: I1205 07:32:18.771029 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tzmkw" Dec 05 07:32:18 crc kubenswrapper[4997]: I1205 07:32:18.934877 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tzmkw"] Dec 05 07:32:19 crc kubenswrapper[4997]: I1205 07:32:19.997892 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tzmkw" podUID="4958f0cc-3cc3-47e5-a754-0b3ab586c91b" containerName="registry-server" containerID="cri-o://ed3cb7b5695e5740b8c3954c55395276729b0785e57296c588772acdd94a7b0c" gracePeriod=2 Dec 05 07:32:20 crc kubenswrapper[4997]: I1205 07:32:20.486951 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tzmkw" Dec 05 07:32:20 crc kubenswrapper[4997]: I1205 07:32:20.506020 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-utilities\") pod \"4958f0cc-3cc3-47e5-a754-0b3ab586c91b\" (UID: \"4958f0cc-3cc3-47e5-a754-0b3ab586c91b\") " Dec 05 07:32:20 crc kubenswrapper[4997]: I1205 07:32:20.506141 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-catalog-content\") pod \"4958f0cc-3cc3-47e5-a754-0b3ab586c91b\" (UID: \"4958f0cc-3cc3-47e5-a754-0b3ab586c91b\") " Dec 05 07:32:20 crc kubenswrapper[4997]: I1205 07:32:20.506462 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nttj\" (UniqueName: \"kubernetes.io/projected/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-kube-api-access-7nttj\") pod \"4958f0cc-3cc3-47e5-a754-0b3ab586c91b\" (UID: \"4958f0cc-3cc3-47e5-a754-0b3ab586c91b\") " Dec 05 07:32:20 crc kubenswrapper[4997]: I1205 07:32:20.511190 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-utilities" (OuterVolumeSpecName: "utilities") pod "4958f0cc-3cc3-47e5-a754-0b3ab586c91b" (UID: "4958f0cc-3cc3-47e5-a754-0b3ab586c91b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:32:20 crc kubenswrapper[4997]: I1205 07:32:20.518990 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-kube-api-access-7nttj" (OuterVolumeSpecName: "kube-api-access-7nttj") pod "4958f0cc-3cc3-47e5-a754-0b3ab586c91b" (UID: "4958f0cc-3cc3-47e5-a754-0b3ab586c91b"). InnerVolumeSpecName "kube-api-access-7nttj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:32:20 crc kubenswrapper[4997]: I1205 07:32:20.609227 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:32:20 crc kubenswrapper[4997]: I1205 07:32:20.609314 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nttj\" (UniqueName: \"kubernetes.io/projected/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-kube-api-access-7nttj\") on node \"crc\" DevicePath \"\"" Dec 05 07:32:20 crc kubenswrapper[4997]: I1205 07:32:20.636052 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4958f0cc-3cc3-47e5-a754-0b3ab586c91b" (UID: "4958f0cc-3cc3-47e5-a754-0b3ab586c91b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:32:20 crc kubenswrapper[4997]: I1205 07:32:20.710458 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4958f0cc-3cc3-47e5-a754-0b3ab586c91b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:32:21 crc kubenswrapper[4997]: I1205 07:32:21.011288 4997 generic.go:334] "Generic (PLEG): container finished" podID="4958f0cc-3cc3-47e5-a754-0b3ab586c91b" containerID="ed3cb7b5695e5740b8c3954c55395276729b0785e57296c588772acdd94a7b0c" exitCode=0 Dec 05 07:32:21 crc kubenswrapper[4997]: I1205 07:32:21.011389 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tzmkw" Dec 05 07:32:21 crc kubenswrapper[4997]: I1205 07:32:21.011366 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tzmkw" event={"ID":"4958f0cc-3cc3-47e5-a754-0b3ab586c91b","Type":"ContainerDied","Data":"ed3cb7b5695e5740b8c3954c55395276729b0785e57296c588772acdd94a7b0c"} Dec 05 07:32:21 crc kubenswrapper[4997]: I1205 07:32:21.011582 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tzmkw" event={"ID":"4958f0cc-3cc3-47e5-a754-0b3ab586c91b","Type":"ContainerDied","Data":"23623063a96eb718e52cc035f01d01bd0cbd18eb4b2921707db721da6205c584"} Dec 05 07:32:21 crc kubenswrapper[4997]: I1205 07:32:21.011653 4997 scope.go:117] "RemoveContainer" containerID="ed3cb7b5695e5740b8c3954c55395276729b0785e57296c588772acdd94a7b0c" Dec 05 07:32:21 crc kubenswrapper[4997]: I1205 07:32:21.040710 4997 scope.go:117] "RemoveContainer" containerID="4d40e37af9d79716c9c82e7e54cf520361585d28a544e7aaf599f66655e6816b" Dec 05 07:32:21 crc kubenswrapper[4997]: I1205 07:32:21.096473 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tzmkw"] Dec 05 07:32:21 crc kubenswrapper[4997]: I1205 07:32:21.097395 4997 scope.go:117] "RemoveContainer" containerID="a3784e1a6dfe29da8902b609c53a23e723f3d3b110bb53ab9ebc036b9f739633" Dec 05 07:32:21 crc kubenswrapper[4997]: I1205 07:32:21.107401 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tzmkw"] Dec 05 07:32:21 crc kubenswrapper[4997]: I1205 07:32:21.123452 4997 scope.go:117] "RemoveContainer" containerID="ed3cb7b5695e5740b8c3954c55395276729b0785e57296c588772acdd94a7b0c" Dec 05 07:32:21 crc kubenswrapper[4997]: E1205 07:32:21.124059 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed3cb7b5695e5740b8c3954c55395276729b0785e57296c588772acdd94a7b0c\": container with ID starting with ed3cb7b5695e5740b8c3954c55395276729b0785e57296c588772acdd94a7b0c not found: ID does not exist" containerID="ed3cb7b5695e5740b8c3954c55395276729b0785e57296c588772acdd94a7b0c" Dec 05 07:32:21 crc kubenswrapper[4997]: I1205 07:32:21.124127 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed3cb7b5695e5740b8c3954c55395276729b0785e57296c588772acdd94a7b0c"} err="failed to get container status \"ed3cb7b5695e5740b8c3954c55395276729b0785e57296c588772acdd94a7b0c\": rpc error: code = NotFound desc = could not find container \"ed3cb7b5695e5740b8c3954c55395276729b0785e57296c588772acdd94a7b0c\": container with ID starting with ed3cb7b5695e5740b8c3954c55395276729b0785e57296c588772acdd94a7b0c not found: ID does not exist" Dec 05 07:32:21 crc kubenswrapper[4997]: I1205 07:32:21.124193 4997 scope.go:117] "RemoveContainer" containerID="4d40e37af9d79716c9c82e7e54cf520361585d28a544e7aaf599f66655e6816b" Dec 05 07:32:21 crc kubenswrapper[4997]: E1205 07:32:21.125019 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d40e37af9d79716c9c82e7e54cf520361585d28a544e7aaf599f66655e6816b\": container with ID starting with 4d40e37af9d79716c9c82e7e54cf520361585d28a544e7aaf599f66655e6816b not found: ID does not exist" containerID="4d40e37af9d79716c9c82e7e54cf520361585d28a544e7aaf599f66655e6816b" Dec 05 07:32:21 crc kubenswrapper[4997]: I1205 07:32:21.125076 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d40e37af9d79716c9c82e7e54cf520361585d28a544e7aaf599f66655e6816b"} err="failed to get container status \"4d40e37af9d79716c9c82e7e54cf520361585d28a544e7aaf599f66655e6816b\": rpc error: code = NotFound desc = could not find container \"4d40e37af9d79716c9c82e7e54cf520361585d28a544e7aaf599f66655e6816b\": container with ID starting with 4d40e37af9d79716c9c82e7e54cf520361585d28a544e7aaf599f66655e6816b not found: ID does not exist" Dec 05 07:32:21 crc kubenswrapper[4997]: I1205 07:32:21.125115 4997 scope.go:117] "RemoveContainer" containerID="a3784e1a6dfe29da8902b609c53a23e723f3d3b110bb53ab9ebc036b9f739633" Dec 05 07:32:21 crc kubenswrapper[4997]: E1205 07:32:21.125517 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3784e1a6dfe29da8902b609c53a23e723f3d3b110bb53ab9ebc036b9f739633\": container with ID starting with a3784e1a6dfe29da8902b609c53a23e723f3d3b110bb53ab9ebc036b9f739633 not found: ID does not exist" containerID="a3784e1a6dfe29da8902b609c53a23e723f3d3b110bb53ab9ebc036b9f739633" Dec 05 07:32:21 crc kubenswrapper[4997]: I1205 07:32:21.125564 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3784e1a6dfe29da8902b609c53a23e723f3d3b110bb53ab9ebc036b9f739633"} err="failed to get container status \"a3784e1a6dfe29da8902b609c53a23e723f3d3b110bb53ab9ebc036b9f739633\": rpc error: code = NotFound desc = could not find container \"a3784e1a6dfe29da8902b609c53a23e723f3d3b110bb53ab9ebc036b9f739633\": container with ID starting with a3784e1a6dfe29da8902b609c53a23e723f3d3b110bb53ab9ebc036b9f739633 not found: ID does not exist" Dec 05 07:32:21 crc kubenswrapper[4997]: I1205 07:32:21.763083 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4958f0cc-3cc3-47e5-a754-0b3ab586c91b" path="/var/lib/kubelet/pods/4958f0cc-3cc3-47e5-a754-0b3ab586c91b/volumes" Dec 05 07:32:23 crc kubenswrapper[4997]: I1205 07:32:23.362941 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8j94t"] Dec 05 07:32:23 crc kubenswrapper[4997]: E1205 07:32:23.364646 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4958f0cc-3cc3-47e5-a754-0b3ab586c91b" containerName="registry-server" Dec 05 07:32:23 crc kubenswrapper[4997]: I1205 07:32:23.364684 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4958f0cc-3cc3-47e5-a754-0b3ab586c91b" containerName="registry-server" Dec 05 07:32:23 crc kubenswrapper[4997]: E1205 07:32:23.364750 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4958f0cc-3cc3-47e5-a754-0b3ab586c91b" containerName="extract-utilities" Dec 05 07:32:23 crc kubenswrapper[4997]: I1205 07:32:23.364769 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4958f0cc-3cc3-47e5-a754-0b3ab586c91b" containerName="extract-utilities" Dec 05 07:32:23 crc kubenswrapper[4997]: E1205 07:32:23.364794 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4958f0cc-3cc3-47e5-a754-0b3ab586c91b" containerName="extract-content" Dec 05 07:32:23 crc kubenswrapper[4997]: I1205 07:32:23.364814 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4958f0cc-3cc3-47e5-a754-0b3ab586c91b" containerName="extract-content" Dec 05 07:32:23 crc kubenswrapper[4997]: I1205 07:32:23.365238 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="4958f0cc-3cc3-47e5-a754-0b3ab586c91b" containerName="registry-server" Dec 05 07:32:23 crc kubenswrapper[4997]: I1205 07:32:23.368382 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8j94t" Dec 05 07:32:23 crc kubenswrapper[4997]: I1205 07:32:23.386059 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8j94t"] Dec 05 07:32:23 crc kubenswrapper[4997]: I1205 07:32:23.454263 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-utilities\") pod \"community-operators-8j94t\" (UID: \"a6cbc45e-db81-42f4-a90c-e1b296e70bfd\") " pod="openshift-marketplace/community-operators-8j94t" Dec 05 07:32:23 crc kubenswrapper[4997]: I1205 07:32:23.454356 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbf2x\" (UniqueName: \"kubernetes.io/projected/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-kube-api-access-jbf2x\") pod \"community-operators-8j94t\" (UID: \"a6cbc45e-db81-42f4-a90c-e1b296e70bfd\") " pod="openshift-marketplace/community-operators-8j94t" Dec 05 07:32:23 crc kubenswrapper[4997]: I1205 07:32:23.454387 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-catalog-content\") pod \"community-operators-8j94t\" (UID: \"a6cbc45e-db81-42f4-a90c-e1b296e70bfd\") " pod="openshift-marketplace/community-operators-8j94t" Dec 05 07:32:23 crc kubenswrapper[4997]: I1205 07:32:23.555120 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-utilities\") pod \"community-operators-8j94t\" (UID: \"a6cbc45e-db81-42f4-a90c-e1b296e70bfd\") " pod="openshift-marketplace/community-operators-8j94t" Dec 05 07:32:23 crc kubenswrapper[4997]: I1205 07:32:23.555210 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbf2x\" (UniqueName: \"kubernetes.io/projected/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-kube-api-access-jbf2x\") pod \"community-operators-8j94t\" (UID: \"a6cbc45e-db81-42f4-a90c-e1b296e70bfd\") " pod="openshift-marketplace/community-operators-8j94t" Dec 05 07:32:23 crc kubenswrapper[4997]: I1205 07:32:23.555242 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-catalog-content\") pod \"community-operators-8j94t\" (UID: \"a6cbc45e-db81-42f4-a90c-e1b296e70bfd\") " pod="openshift-marketplace/community-operators-8j94t" Dec 05 07:32:23 crc kubenswrapper[4997]: I1205 07:32:23.555885 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-utilities\") pod \"community-operators-8j94t\" (UID: \"a6cbc45e-db81-42f4-a90c-e1b296e70bfd\") " pod="openshift-marketplace/community-operators-8j94t" Dec 05 07:32:23 crc kubenswrapper[4997]: I1205 07:32:23.555922 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-catalog-content\") pod \"community-operators-8j94t\" (UID: \"a6cbc45e-db81-42f4-a90c-e1b296e70bfd\") " pod="openshift-marketplace/community-operators-8j94t" Dec 05 07:32:23 crc kubenswrapper[4997]: I1205 07:32:23.576194 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbf2x\" (UniqueName: \"kubernetes.io/projected/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-kube-api-access-jbf2x\") pod \"community-operators-8j94t\" (UID: \"a6cbc45e-db81-42f4-a90c-e1b296e70bfd\") " pod="openshift-marketplace/community-operators-8j94t" Dec 05 07:32:23 crc kubenswrapper[4997]: I1205 07:32:23.712385 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8j94t" Dec 05 07:32:24 crc kubenswrapper[4997]: I1205 07:32:24.250372 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8j94t"] Dec 05 07:32:25 crc kubenswrapper[4997]: I1205 07:32:25.056292 4997 generic.go:334] "Generic (PLEG): container finished" podID="a6cbc45e-db81-42f4-a90c-e1b296e70bfd" containerID="b16db9d67cd0ff418238014d8bd6b2725694b11ebe496cb71a003e76a11e0c1d" exitCode=0 Dec 05 07:32:25 crc kubenswrapper[4997]: I1205 07:32:25.056398 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8j94t" event={"ID":"a6cbc45e-db81-42f4-a90c-e1b296e70bfd","Type":"ContainerDied","Data":"b16db9d67cd0ff418238014d8bd6b2725694b11ebe496cb71a003e76a11e0c1d"} Dec 05 07:32:25 crc kubenswrapper[4997]: I1205 07:32:25.056821 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8j94t" event={"ID":"a6cbc45e-db81-42f4-a90c-e1b296e70bfd","Type":"ContainerStarted","Data":"28cc451aeeebe4bae78f5280f64ed6910df8604a6e112bae2a2bf35f2864d115"} Dec 05 07:32:25 crc kubenswrapper[4997]: I1205 07:32:25.545164 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zhhx7"] Dec 05 07:32:25 crc kubenswrapper[4997]: I1205 07:32:25.547380 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhhx7" Dec 05 07:32:25 crc kubenswrapper[4997]: I1205 07:32:25.557692 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhhx7"] Dec 05 07:32:25 crc kubenswrapper[4997]: I1205 07:32:25.593584 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31f64fe5-a126-42c6-9ebc-be65835acce1-catalog-content\") pod \"redhat-marketplace-zhhx7\" (UID: \"31f64fe5-a126-42c6-9ebc-be65835acce1\") " pod="openshift-marketplace/redhat-marketplace-zhhx7" Dec 05 07:32:25 crc kubenswrapper[4997]: I1205 07:32:25.593682 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tp2z\" (UniqueName: \"kubernetes.io/projected/31f64fe5-a126-42c6-9ebc-be65835acce1-kube-api-access-7tp2z\") pod \"redhat-marketplace-zhhx7\" (UID: \"31f64fe5-a126-42c6-9ebc-be65835acce1\") " pod="openshift-marketplace/redhat-marketplace-zhhx7" Dec 05 07:32:25 crc kubenswrapper[4997]: I1205 07:32:25.593774 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31f64fe5-a126-42c6-9ebc-be65835acce1-utilities\") pod \"redhat-marketplace-zhhx7\" (UID: \"31f64fe5-a126-42c6-9ebc-be65835acce1\") " pod="openshift-marketplace/redhat-marketplace-zhhx7" Dec 05 07:32:25 crc kubenswrapper[4997]: I1205 07:32:25.695736 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31f64fe5-a126-42c6-9ebc-be65835acce1-utilities\") pod \"redhat-marketplace-zhhx7\" (UID: \"31f64fe5-a126-42c6-9ebc-be65835acce1\") " pod="openshift-marketplace/redhat-marketplace-zhhx7" Dec 05 07:32:25 crc kubenswrapper[4997]: I1205 07:32:25.695862 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31f64fe5-a126-42c6-9ebc-be65835acce1-catalog-content\") pod \"redhat-marketplace-zhhx7\" (UID: \"31f64fe5-a126-42c6-9ebc-be65835acce1\") " pod="openshift-marketplace/redhat-marketplace-zhhx7" Dec 05 07:32:25 crc kubenswrapper[4997]: I1205 07:32:25.695895 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tp2z\" (UniqueName: \"kubernetes.io/projected/31f64fe5-a126-42c6-9ebc-be65835acce1-kube-api-access-7tp2z\") pod \"redhat-marketplace-zhhx7\" (UID: \"31f64fe5-a126-42c6-9ebc-be65835acce1\") " pod="openshift-marketplace/redhat-marketplace-zhhx7" Dec 05 07:32:25 crc kubenswrapper[4997]: I1205 07:32:25.697225 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31f64fe5-a126-42c6-9ebc-be65835acce1-catalog-content\") pod \"redhat-marketplace-zhhx7\" (UID: \"31f64fe5-a126-42c6-9ebc-be65835acce1\") " pod="openshift-marketplace/redhat-marketplace-zhhx7" Dec 05 07:32:25 crc kubenswrapper[4997]: I1205 07:32:25.697258 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31f64fe5-a126-42c6-9ebc-be65835acce1-utilities\") pod \"redhat-marketplace-zhhx7\" (UID: \"31f64fe5-a126-42c6-9ebc-be65835acce1\") " pod="openshift-marketplace/redhat-marketplace-zhhx7" Dec 05 07:32:25 crc kubenswrapper[4997]: I1205 07:32:25.733232 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tp2z\" (UniqueName: \"kubernetes.io/projected/31f64fe5-a126-42c6-9ebc-be65835acce1-kube-api-access-7tp2z\") pod \"redhat-marketplace-zhhx7\" (UID: \"31f64fe5-a126-42c6-9ebc-be65835acce1\") " pod="openshift-marketplace/redhat-marketplace-zhhx7" Dec 05 07:32:25 crc kubenswrapper[4997]: I1205 07:32:25.879828 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhhx7" Dec 05 07:32:26 crc kubenswrapper[4997]: I1205 07:32:26.072426 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8j94t" event={"ID":"a6cbc45e-db81-42f4-a90c-e1b296e70bfd","Type":"ContainerStarted","Data":"66d7a96ec32124e11061d0071b20c6db4933c493dc206566a5ab143cdf83ee38"} Dec 05 07:32:26 crc kubenswrapper[4997]: I1205 07:32:26.395392 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhhx7"] Dec 05 07:32:26 crc kubenswrapper[4997]: W1205 07:32:26.402358 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31f64fe5_a126_42c6_9ebc_be65835acce1.slice/crio-bbc4024bae350757e2a2b4639132f901813bc3f4b504ef4da33c450f85d0a63d WatchSource:0}: Error finding container bbc4024bae350757e2a2b4639132f901813bc3f4b504ef4da33c450f85d0a63d: Status 404 returned error can't find the container with id bbc4024bae350757e2a2b4639132f901813bc3f4b504ef4da33c450f85d0a63d Dec 05 07:32:27 crc kubenswrapper[4997]: I1205 07:32:27.091872 4997 generic.go:334] "Generic (PLEG): container finished" podID="a6cbc45e-db81-42f4-a90c-e1b296e70bfd" containerID="66d7a96ec32124e11061d0071b20c6db4933c493dc206566a5ab143cdf83ee38" exitCode=0 Dec 05 07:32:27 crc kubenswrapper[4997]: I1205 07:32:27.092747 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8j94t" event={"ID":"a6cbc45e-db81-42f4-a90c-e1b296e70bfd","Type":"ContainerDied","Data":"66d7a96ec32124e11061d0071b20c6db4933c493dc206566a5ab143cdf83ee38"} Dec 05 07:32:27 crc kubenswrapper[4997]: I1205 07:32:27.098152 4997 generic.go:334] "Generic (PLEG): container finished" podID="31f64fe5-a126-42c6-9ebc-be65835acce1" containerID="ad214e9f229646bff0aaff1d7e91114f477decf53429c509455da535ab1591a1" exitCode=0 Dec 05 07:32:27 crc kubenswrapper[4997]: I1205 07:32:27.098199 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhhx7" event={"ID":"31f64fe5-a126-42c6-9ebc-be65835acce1","Type":"ContainerDied","Data":"ad214e9f229646bff0aaff1d7e91114f477decf53429c509455da535ab1591a1"} Dec 05 07:32:27 crc kubenswrapper[4997]: I1205 07:32:27.098229 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhhx7" event={"ID":"31f64fe5-a126-42c6-9ebc-be65835acce1","Type":"ContainerStarted","Data":"bbc4024bae350757e2a2b4639132f901813bc3f4b504ef4da33c450f85d0a63d"} Dec 05 07:32:28 crc kubenswrapper[4997]: I1205 07:32:28.110670 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8j94t" event={"ID":"a6cbc45e-db81-42f4-a90c-e1b296e70bfd","Type":"ContainerStarted","Data":"b2fea1c3b7572f8d519a05b2e07fc942b646bf6bf4338bb6ceac7663746d2e65"} Dec 05 07:32:28 crc kubenswrapper[4997]: I1205 07:32:28.114224 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhhx7" event={"ID":"31f64fe5-a126-42c6-9ebc-be65835acce1","Type":"ContainerStarted","Data":"0f6fca32994589c29db55c220579c38b6651393cf3a25ba17596c21677d23ff0"} Dec 05 07:32:28 crc kubenswrapper[4997]: I1205 07:32:28.132284 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8j94t" podStartSLOduration=2.669949799 podStartE2EDuration="5.132252641s" podCreationTimestamp="2025-12-05 07:32:23 +0000 UTC" firstStartedPulling="2025-12-05 07:32:25.058022075 +0000 UTC m=+2245.586929336" lastFinishedPulling="2025-12-05 07:32:27.520324907 +0000 UTC m=+2248.049232178" observedRunningTime="2025-12-05 07:32:28.127914704 +0000 UTC m=+2248.656821975" watchObservedRunningTime="2025-12-05 07:32:28.132252641 +0000 UTC m=+2248.661159902" Dec 05 07:32:29 crc kubenswrapper[4997]: I1205 07:32:29.128525 4997 generic.go:334] "Generic (PLEG): container finished" podID="31f64fe5-a126-42c6-9ebc-be65835acce1" containerID="0f6fca32994589c29db55c220579c38b6651393cf3a25ba17596c21677d23ff0" exitCode=0 Dec 05 07:32:29 crc kubenswrapper[4997]: I1205 07:32:29.128660 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhhx7" event={"ID":"31f64fe5-a126-42c6-9ebc-be65835acce1","Type":"ContainerDied","Data":"0f6fca32994589c29db55c220579c38b6651393cf3a25ba17596c21677d23ff0"} Dec 05 07:32:30 crc kubenswrapper[4997]: I1205 07:32:30.137395 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhhx7" event={"ID":"31f64fe5-a126-42c6-9ebc-be65835acce1","Type":"ContainerStarted","Data":"d1922a162435ea293865e86f48d980a8bb9d0ec18acd0f37d73e26bc01e9d910"} Dec 05 07:32:30 crc kubenswrapper[4997]: I1205 07:32:30.163969 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zhhx7" podStartSLOduration=2.606661199 podStartE2EDuration="5.163942097s" podCreationTimestamp="2025-12-05 07:32:25 +0000 UTC" firstStartedPulling="2025-12-05 07:32:27.103454032 +0000 UTC m=+2247.632361293" lastFinishedPulling="2025-12-05 07:32:29.66073492 +0000 UTC m=+2250.189642191" observedRunningTime="2025-12-05 07:32:30.157938844 +0000 UTC m=+2250.686846105" watchObservedRunningTime="2025-12-05 07:32:30.163942097 +0000 UTC m=+2250.692849358" Dec 05 07:32:33 crc kubenswrapper[4997]: I1205 07:32:33.713470 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8j94t" Dec 05 07:32:33 crc kubenswrapper[4997]: I1205 07:32:33.715401 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8j94t" Dec 05 07:32:33 crc kubenswrapper[4997]: I1205 07:32:33.803747 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8j94t" Dec 05 07:32:34 crc kubenswrapper[4997]: I1205 07:32:34.231474 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8j94t" Dec 05 07:32:35 crc kubenswrapper[4997]: I1205 07:32:35.336526 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8j94t"] Dec 05 07:32:35 crc kubenswrapper[4997]: I1205 07:32:35.880223 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zhhx7" Dec 05 07:32:35 crc kubenswrapper[4997]: I1205 07:32:35.880300 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zhhx7" Dec 05 07:32:35 crc kubenswrapper[4997]: I1205 07:32:35.930255 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zhhx7" Dec 05 07:32:36 crc kubenswrapper[4997]: I1205 07:32:36.195633 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8j94t" podUID="a6cbc45e-db81-42f4-a90c-e1b296e70bfd" containerName="registry-server" containerID="cri-o://b2fea1c3b7572f8d519a05b2e07fc942b646bf6bf4338bb6ceac7663746d2e65" gracePeriod=2 Dec 05 07:32:36 crc kubenswrapper[4997]: I1205 07:32:36.276029 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zhhx7" Dec 05 07:32:37 crc kubenswrapper[4997]: I1205 07:32:37.734684 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhhx7"] Dec 05 07:32:37 crc kubenswrapper[4997]: I1205 07:32:37.784199 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8j94t" Dec 05 07:32:37 crc kubenswrapper[4997]: I1205 07:32:37.914978 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-catalog-content\") pod \"a6cbc45e-db81-42f4-a90c-e1b296e70bfd\" (UID: \"a6cbc45e-db81-42f4-a90c-e1b296e70bfd\") " Dec 05 07:32:37 crc kubenswrapper[4997]: I1205 07:32:37.915464 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-utilities\") pod \"a6cbc45e-db81-42f4-a90c-e1b296e70bfd\" (UID: \"a6cbc45e-db81-42f4-a90c-e1b296e70bfd\") " Dec 05 07:32:37 crc kubenswrapper[4997]: I1205 07:32:37.915756 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbf2x\" (UniqueName: \"kubernetes.io/projected/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-kube-api-access-jbf2x\") pod \"a6cbc45e-db81-42f4-a90c-e1b296e70bfd\" (UID: \"a6cbc45e-db81-42f4-a90c-e1b296e70bfd\") " Dec 05 07:32:37 crc kubenswrapper[4997]: I1205 07:32:37.918082 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-utilities" (OuterVolumeSpecName: "utilities") pod "a6cbc45e-db81-42f4-a90c-e1b296e70bfd" (UID: "a6cbc45e-db81-42f4-a90c-e1b296e70bfd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:32:37 crc kubenswrapper[4997]: I1205 07:32:37.927045 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-kube-api-access-jbf2x" (OuterVolumeSpecName: "kube-api-access-jbf2x") pod "a6cbc45e-db81-42f4-a90c-e1b296e70bfd" (UID: "a6cbc45e-db81-42f4-a90c-e1b296e70bfd"). InnerVolumeSpecName "kube-api-access-jbf2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:32:37 crc kubenswrapper[4997]: I1205 07:32:37.993651 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a6cbc45e-db81-42f4-a90c-e1b296e70bfd" (UID: "a6cbc45e-db81-42f4-a90c-e1b296e70bfd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.018118 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.018177 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.018194 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbf2x\" (UniqueName: \"kubernetes.io/projected/a6cbc45e-db81-42f4-a90c-e1b296e70bfd-kube-api-access-jbf2x\") on node \"crc\" DevicePath \"\"" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.217778 4997 generic.go:334] "Generic (PLEG): container finished" podID="a6cbc45e-db81-42f4-a90c-e1b296e70bfd" containerID="b2fea1c3b7572f8d519a05b2e07fc942b646bf6bf4338bb6ceac7663746d2e65" exitCode=0 Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.217837 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8j94t" event={"ID":"a6cbc45e-db81-42f4-a90c-e1b296e70bfd","Type":"ContainerDied","Data":"b2fea1c3b7572f8d519a05b2e07fc942b646bf6bf4338bb6ceac7663746d2e65"} Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.217899 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8j94t" event={"ID":"a6cbc45e-db81-42f4-a90c-e1b296e70bfd","Type":"ContainerDied","Data":"28cc451aeeebe4bae78f5280f64ed6910df8604a6e112bae2a2bf35f2864d115"} Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.217927 4997 scope.go:117] "RemoveContainer" containerID="b2fea1c3b7572f8d519a05b2e07fc942b646bf6bf4338bb6ceac7663746d2e65" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.217930 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8j94t" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.218130 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zhhx7" podUID="31f64fe5-a126-42c6-9ebc-be65835acce1" containerName="registry-server" containerID="cri-o://d1922a162435ea293865e86f48d980a8bb9d0ec18acd0f37d73e26bc01e9d910" gracePeriod=2 Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.254141 4997 scope.go:117] "RemoveContainer" containerID="66d7a96ec32124e11061d0071b20c6db4933c493dc206566a5ab143cdf83ee38" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.273190 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8j94t"] Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.302783 4997 scope.go:117] "RemoveContainer" containerID="b16db9d67cd0ff418238014d8bd6b2725694b11ebe496cb71a003e76a11e0c1d" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.305142 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8j94t"] Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.334319 4997 scope.go:117] "RemoveContainer" containerID="b2fea1c3b7572f8d519a05b2e07fc942b646bf6bf4338bb6ceac7663746d2e65" Dec 05 07:32:38 crc kubenswrapper[4997]: E1205 07:32:38.335215 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2fea1c3b7572f8d519a05b2e07fc942b646bf6bf4338bb6ceac7663746d2e65\": container with ID starting with b2fea1c3b7572f8d519a05b2e07fc942b646bf6bf4338bb6ceac7663746d2e65 not found: ID does not exist" containerID="b2fea1c3b7572f8d519a05b2e07fc942b646bf6bf4338bb6ceac7663746d2e65" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.335339 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2fea1c3b7572f8d519a05b2e07fc942b646bf6bf4338bb6ceac7663746d2e65"} err="failed to get container status \"b2fea1c3b7572f8d519a05b2e07fc942b646bf6bf4338bb6ceac7663746d2e65\": rpc error: code = NotFound desc = could not find container \"b2fea1c3b7572f8d519a05b2e07fc942b646bf6bf4338bb6ceac7663746d2e65\": container with ID starting with b2fea1c3b7572f8d519a05b2e07fc942b646bf6bf4338bb6ceac7663746d2e65 not found: ID does not exist" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.335408 4997 scope.go:117] "RemoveContainer" containerID="66d7a96ec32124e11061d0071b20c6db4933c493dc206566a5ab143cdf83ee38" Dec 05 07:32:38 crc kubenswrapper[4997]: E1205 07:32:38.335862 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66d7a96ec32124e11061d0071b20c6db4933c493dc206566a5ab143cdf83ee38\": container with ID starting with 66d7a96ec32124e11061d0071b20c6db4933c493dc206566a5ab143cdf83ee38 not found: ID does not exist" containerID="66d7a96ec32124e11061d0071b20c6db4933c493dc206566a5ab143cdf83ee38" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.335917 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66d7a96ec32124e11061d0071b20c6db4933c493dc206566a5ab143cdf83ee38"} err="failed to get container status \"66d7a96ec32124e11061d0071b20c6db4933c493dc206566a5ab143cdf83ee38\": rpc error: code = NotFound desc = could not find container \"66d7a96ec32124e11061d0071b20c6db4933c493dc206566a5ab143cdf83ee38\": container with ID starting with 66d7a96ec32124e11061d0071b20c6db4933c493dc206566a5ab143cdf83ee38 not found: ID does not exist" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.335960 4997 scope.go:117] "RemoveContainer" containerID="b16db9d67cd0ff418238014d8bd6b2725694b11ebe496cb71a003e76a11e0c1d" Dec 05 07:32:38 crc kubenswrapper[4997]: E1205 07:32:38.336500 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b16db9d67cd0ff418238014d8bd6b2725694b11ebe496cb71a003e76a11e0c1d\": container with ID starting with b16db9d67cd0ff418238014d8bd6b2725694b11ebe496cb71a003e76a11e0c1d not found: ID does not exist" containerID="b16db9d67cd0ff418238014d8bd6b2725694b11ebe496cb71a003e76a11e0c1d" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.336525 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b16db9d67cd0ff418238014d8bd6b2725694b11ebe496cb71a003e76a11e0c1d"} err="failed to get container status \"b16db9d67cd0ff418238014d8bd6b2725694b11ebe496cb71a003e76a11e0c1d\": rpc error: code = NotFound desc = could not find container \"b16db9d67cd0ff418238014d8bd6b2725694b11ebe496cb71a003e76a11e0c1d\": container with ID starting with b16db9d67cd0ff418238014d8bd6b2725694b11ebe496cb71a003e76a11e0c1d not found: ID does not exist" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.630122 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhhx7" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.737384 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tp2z\" (UniqueName: \"kubernetes.io/projected/31f64fe5-a126-42c6-9ebc-be65835acce1-kube-api-access-7tp2z\") pod \"31f64fe5-a126-42c6-9ebc-be65835acce1\" (UID: \"31f64fe5-a126-42c6-9ebc-be65835acce1\") " Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.737580 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31f64fe5-a126-42c6-9ebc-be65835acce1-catalog-content\") pod \"31f64fe5-a126-42c6-9ebc-be65835acce1\" (UID: \"31f64fe5-a126-42c6-9ebc-be65835acce1\") " Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.737694 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31f64fe5-a126-42c6-9ebc-be65835acce1-utilities\") pod \"31f64fe5-a126-42c6-9ebc-be65835acce1\" (UID: \"31f64fe5-a126-42c6-9ebc-be65835acce1\") " Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.739110 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31f64fe5-a126-42c6-9ebc-be65835acce1-utilities" (OuterVolumeSpecName: "utilities") pod "31f64fe5-a126-42c6-9ebc-be65835acce1" (UID: "31f64fe5-a126-42c6-9ebc-be65835acce1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.744191 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31f64fe5-a126-42c6-9ebc-be65835acce1-kube-api-access-7tp2z" (OuterVolumeSpecName: "kube-api-access-7tp2z") pod "31f64fe5-a126-42c6-9ebc-be65835acce1" (UID: "31f64fe5-a126-42c6-9ebc-be65835acce1"). InnerVolumeSpecName "kube-api-access-7tp2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.770456 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31f64fe5-a126-42c6-9ebc-be65835acce1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "31f64fe5-a126-42c6-9ebc-be65835acce1" (UID: "31f64fe5-a126-42c6-9ebc-be65835acce1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.839728 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31f64fe5-a126-42c6-9ebc-be65835acce1-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.839780 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tp2z\" (UniqueName: \"kubernetes.io/projected/31f64fe5-a126-42c6-9ebc-be65835acce1-kube-api-access-7tp2z\") on node \"crc\" DevicePath \"\"" Dec 05 07:32:38 crc kubenswrapper[4997]: I1205 07:32:38.839796 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31f64fe5-a126-42c6-9ebc-be65835acce1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:32:39 crc kubenswrapper[4997]: I1205 07:32:39.233213 4997 generic.go:334] "Generic (PLEG): container finished" podID="31f64fe5-a126-42c6-9ebc-be65835acce1" containerID="d1922a162435ea293865e86f48d980a8bb9d0ec18acd0f37d73e26bc01e9d910" exitCode=0 Dec 05 07:32:39 crc kubenswrapper[4997]: I1205 07:32:39.233276 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhhx7" Dec 05 07:32:39 crc kubenswrapper[4997]: I1205 07:32:39.233294 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhhx7" event={"ID":"31f64fe5-a126-42c6-9ebc-be65835acce1","Type":"ContainerDied","Data":"d1922a162435ea293865e86f48d980a8bb9d0ec18acd0f37d73e26bc01e9d910"} Dec 05 07:32:39 crc kubenswrapper[4997]: I1205 07:32:39.233349 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhhx7" event={"ID":"31f64fe5-a126-42c6-9ebc-be65835acce1","Type":"ContainerDied","Data":"bbc4024bae350757e2a2b4639132f901813bc3f4b504ef4da33c450f85d0a63d"} Dec 05 07:32:39 crc kubenswrapper[4997]: I1205 07:32:39.233382 4997 scope.go:117] "RemoveContainer" containerID="d1922a162435ea293865e86f48d980a8bb9d0ec18acd0f37d73e26bc01e9d910" Dec 05 07:32:39 crc kubenswrapper[4997]: I1205 07:32:39.274378 4997 scope.go:117] "RemoveContainer" containerID="0f6fca32994589c29db55c220579c38b6651393cf3a25ba17596c21677d23ff0" Dec 05 07:32:39 crc kubenswrapper[4997]: I1205 07:32:39.284600 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhhx7"] Dec 05 07:32:39 crc kubenswrapper[4997]: I1205 07:32:39.292681 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhhx7"] Dec 05 07:32:39 crc kubenswrapper[4997]: I1205 07:32:39.297064 4997 scope.go:117] "RemoveContainer" containerID="ad214e9f229646bff0aaff1d7e91114f477decf53429c509455da535ab1591a1" Dec 05 07:32:39 crc kubenswrapper[4997]: I1205 07:32:39.330143 4997 scope.go:117] "RemoveContainer" containerID="d1922a162435ea293865e86f48d980a8bb9d0ec18acd0f37d73e26bc01e9d910" Dec 05 07:32:39 crc kubenswrapper[4997]: E1205 07:32:39.330652 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1922a162435ea293865e86f48d980a8bb9d0ec18acd0f37d73e26bc01e9d910\": container with ID starting with d1922a162435ea293865e86f48d980a8bb9d0ec18acd0f37d73e26bc01e9d910 not found: ID does not exist" containerID="d1922a162435ea293865e86f48d980a8bb9d0ec18acd0f37d73e26bc01e9d910" Dec 05 07:32:39 crc kubenswrapper[4997]: I1205 07:32:39.330686 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1922a162435ea293865e86f48d980a8bb9d0ec18acd0f37d73e26bc01e9d910"} err="failed to get container status \"d1922a162435ea293865e86f48d980a8bb9d0ec18acd0f37d73e26bc01e9d910\": rpc error: code = NotFound desc = could not find container \"d1922a162435ea293865e86f48d980a8bb9d0ec18acd0f37d73e26bc01e9d910\": container with ID starting with d1922a162435ea293865e86f48d980a8bb9d0ec18acd0f37d73e26bc01e9d910 not found: ID does not exist" Dec 05 07:32:39 crc kubenswrapper[4997]: I1205 07:32:39.330715 4997 scope.go:117] "RemoveContainer" containerID="0f6fca32994589c29db55c220579c38b6651393cf3a25ba17596c21677d23ff0" Dec 05 07:32:39 crc kubenswrapper[4997]: E1205 07:32:39.331038 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f6fca32994589c29db55c220579c38b6651393cf3a25ba17596c21677d23ff0\": container with ID starting with 0f6fca32994589c29db55c220579c38b6651393cf3a25ba17596c21677d23ff0 not found: ID does not exist" containerID="0f6fca32994589c29db55c220579c38b6651393cf3a25ba17596c21677d23ff0" Dec 05 07:32:39 crc kubenswrapper[4997]: I1205 07:32:39.331082 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f6fca32994589c29db55c220579c38b6651393cf3a25ba17596c21677d23ff0"} err="failed to get container status \"0f6fca32994589c29db55c220579c38b6651393cf3a25ba17596c21677d23ff0\": rpc error: code = NotFound desc = could not find container \"0f6fca32994589c29db55c220579c38b6651393cf3a25ba17596c21677d23ff0\": container with ID starting with 0f6fca32994589c29db55c220579c38b6651393cf3a25ba17596c21677d23ff0 not found: ID does not exist" Dec 05 07:32:39 crc kubenswrapper[4997]: I1205 07:32:39.331118 4997 scope.go:117] "RemoveContainer" containerID="ad214e9f229646bff0aaff1d7e91114f477decf53429c509455da535ab1591a1" Dec 05 07:32:39 crc kubenswrapper[4997]: E1205 07:32:39.331390 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad214e9f229646bff0aaff1d7e91114f477decf53429c509455da535ab1591a1\": container with ID starting with ad214e9f229646bff0aaff1d7e91114f477decf53429c509455da535ab1591a1 not found: ID does not exist" containerID="ad214e9f229646bff0aaff1d7e91114f477decf53429c509455da535ab1591a1" Dec 05 07:32:39 crc kubenswrapper[4997]: I1205 07:32:39.331425 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad214e9f229646bff0aaff1d7e91114f477decf53429c509455da535ab1591a1"} err="failed to get container status \"ad214e9f229646bff0aaff1d7e91114f477decf53429c509455da535ab1591a1\": rpc error: code = NotFound desc = could not find container \"ad214e9f229646bff0aaff1d7e91114f477decf53429c509455da535ab1591a1\": container with ID starting with ad214e9f229646bff0aaff1d7e91114f477decf53429c509455da535ab1591a1 not found: ID does not exist" Dec 05 07:32:39 crc kubenswrapper[4997]: I1205 07:32:39.769021 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31f64fe5-a126-42c6-9ebc-be65835acce1" path="/var/lib/kubelet/pods/31f64fe5-a126-42c6-9ebc-be65835acce1/volumes" Dec 05 07:32:39 crc kubenswrapper[4997]: I1205 07:32:39.770409 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6cbc45e-db81-42f4-a90c-e1b296e70bfd" path="/var/lib/kubelet/pods/a6cbc45e-db81-42f4-a90c-e1b296e70bfd/volumes" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.280197 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wrnn4"] Dec 05 07:33:18 crc kubenswrapper[4997]: E1205 07:33:18.281885 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6cbc45e-db81-42f4-a90c-e1b296e70bfd" containerName="registry-server" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.281914 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6cbc45e-db81-42f4-a90c-e1b296e70bfd" containerName="registry-server" Dec 05 07:33:18 crc kubenswrapper[4997]: E1205 07:33:18.281932 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6cbc45e-db81-42f4-a90c-e1b296e70bfd" containerName="extract-utilities" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.281941 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6cbc45e-db81-42f4-a90c-e1b296e70bfd" containerName="extract-utilities" Dec 05 07:33:18 crc kubenswrapper[4997]: E1205 07:33:18.281959 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6cbc45e-db81-42f4-a90c-e1b296e70bfd" containerName="extract-content" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.281970 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6cbc45e-db81-42f4-a90c-e1b296e70bfd" containerName="extract-content" Dec 05 07:33:18 crc kubenswrapper[4997]: E1205 07:33:18.281997 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31f64fe5-a126-42c6-9ebc-be65835acce1" containerName="registry-server" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.282008 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="31f64fe5-a126-42c6-9ebc-be65835acce1" containerName="registry-server" Dec 05 07:33:18 crc kubenswrapper[4997]: E1205 07:33:18.282025 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31f64fe5-a126-42c6-9ebc-be65835acce1" containerName="extract-content" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.282037 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="31f64fe5-a126-42c6-9ebc-be65835acce1" containerName="extract-content" Dec 05 07:33:18 crc kubenswrapper[4997]: E1205 07:33:18.282066 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31f64fe5-a126-42c6-9ebc-be65835acce1" containerName="extract-utilities" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.282077 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="31f64fe5-a126-42c6-9ebc-be65835acce1" containerName="extract-utilities" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.282321 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="31f64fe5-a126-42c6-9ebc-be65835acce1" containerName="registry-server" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.282347 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6cbc45e-db81-42f4-a90c-e1b296e70bfd" containerName="registry-server" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.284215 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wrnn4" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.291649 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wrnn4"] Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.392049 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/287fa72b-6f52-472f-8178-519dec66712a-utilities\") pod \"certified-operators-wrnn4\" (UID: \"287fa72b-6f52-472f-8178-519dec66712a\") " pod="openshift-marketplace/certified-operators-wrnn4" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.392116 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4sw9\" (UniqueName: \"kubernetes.io/projected/287fa72b-6f52-472f-8178-519dec66712a-kube-api-access-m4sw9\") pod \"certified-operators-wrnn4\" (UID: \"287fa72b-6f52-472f-8178-519dec66712a\") " pod="openshift-marketplace/certified-operators-wrnn4" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.392260 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/287fa72b-6f52-472f-8178-519dec66712a-catalog-content\") pod \"certified-operators-wrnn4\" (UID: \"287fa72b-6f52-472f-8178-519dec66712a\") " pod="openshift-marketplace/certified-operators-wrnn4" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.494145 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/287fa72b-6f52-472f-8178-519dec66712a-utilities\") pod \"certified-operators-wrnn4\" (UID: \"287fa72b-6f52-472f-8178-519dec66712a\") " pod="openshift-marketplace/certified-operators-wrnn4" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.494542 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4sw9\" (UniqueName: \"kubernetes.io/projected/287fa72b-6f52-472f-8178-519dec66712a-kube-api-access-m4sw9\") pod \"certified-operators-wrnn4\" (UID: \"287fa72b-6f52-472f-8178-519dec66712a\") " pod="openshift-marketplace/certified-operators-wrnn4" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.494793 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/287fa72b-6f52-472f-8178-519dec66712a-catalog-content\") pod \"certified-operators-wrnn4\" (UID: \"287fa72b-6f52-472f-8178-519dec66712a\") " pod="openshift-marketplace/certified-operators-wrnn4" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.495055 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/287fa72b-6f52-472f-8178-519dec66712a-utilities\") pod \"certified-operators-wrnn4\" (UID: \"287fa72b-6f52-472f-8178-519dec66712a\") " pod="openshift-marketplace/certified-operators-wrnn4" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.495143 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/287fa72b-6f52-472f-8178-519dec66712a-catalog-content\") pod \"certified-operators-wrnn4\" (UID: \"287fa72b-6f52-472f-8178-519dec66712a\") " pod="openshift-marketplace/certified-operators-wrnn4" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.527017 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4sw9\" (UniqueName: \"kubernetes.io/projected/287fa72b-6f52-472f-8178-519dec66712a-kube-api-access-m4sw9\") pod \"certified-operators-wrnn4\" (UID: \"287fa72b-6f52-472f-8178-519dec66712a\") " pod="openshift-marketplace/certified-operators-wrnn4" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.613966 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wrnn4" Dec 05 07:33:18 crc kubenswrapper[4997]: I1205 07:33:18.932667 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wrnn4"] Dec 05 07:33:19 crc kubenswrapper[4997]: I1205 07:33:19.676791 4997 generic.go:334] "Generic (PLEG): container finished" podID="287fa72b-6f52-472f-8178-519dec66712a" containerID="bcc9f34a74fb5b36b855a21fefef3b5c0f4c2067b496a827e7cd6914441dd241" exitCode=0 Dec 05 07:33:19 crc kubenswrapper[4997]: I1205 07:33:19.677032 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wrnn4" event={"ID":"287fa72b-6f52-472f-8178-519dec66712a","Type":"ContainerDied","Data":"bcc9f34a74fb5b36b855a21fefef3b5c0f4c2067b496a827e7cd6914441dd241"} Dec 05 07:33:19 crc kubenswrapper[4997]: I1205 07:33:19.677203 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wrnn4" event={"ID":"287fa72b-6f52-472f-8178-519dec66712a","Type":"ContainerStarted","Data":"97edcca745baedde49b73a2650742d3e5ebb08b3bc889a75d884480d14a6b94b"} Dec 05 07:33:20 crc kubenswrapper[4997]: I1205 07:33:20.686221 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wrnn4" event={"ID":"287fa72b-6f52-472f-8178-519dec66712a","Type":"ContainerStarted","Data":"bca1960575a8385ba4aea3b543f55e5ce34331e4a811162e0d0eb49884a53baa"} Dec 05 07:33:21 crc kubenswrapper[4997]: I1205 07:33:21.697301 4997 generic.go:334] "Generic (PLEG): container finished" podID="287fa72b-6f52-472f-8178-519dec66712a" containerID="bca1960575a8385ba4aea3b543f55e5ce34331e4a811162e0d0eb49884a53baa" exitCode=0 Dec 05 07:33:21 crc kubenswrapper[4997]: I1205 07:33:21.697367 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wrnn4" event={"ID":"287fa72b-6f52-472f-8178-519dec66712a","Type":"ContainerDied","Data":"bca1960575a8385ba4aea3b543f55e5ce34331e4a811162e0d0eb49884a53baa"} Dec 05 07:33:22 crc kubenswrapper[4997]: I1205 07:33:22.711367 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wrnn4" event={"ID":"287fa72b-6f52-472f-8178-519dec66712a","Type":"ContainerStarted","Data":"64b483de4e5057ab3544126edfa8474f1145cbc0662e49a14ba73ca8e4c28ccb"} Dec 05 07:33:22 crc kubenswrapper[4997]: I1205 07:33:22.756482 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wrnn4" podStartSLOduration=2.346689363 podStartE2EDuration="4.756455584s" podCreationTimestamp="2025-12-05 07:33:18 +0000 UTC" firstStartedPulling="2025-12-05 07:33:19.678873772 +0000 UTC m=+2300.207781043" lastFinishedPulling="2025-12-05 07:33:22.088640003 +0000 UTC m=+2302.617547264" observedRunningTime="2025-12-05 07:33:22.742023045 +0000 UTC m=+2303.270930346" watchObservedRunningTime="2025-12-05 07:33:22.756455584 +0000 UTC m=+2303.285362845" Dec 05 07:33:28 crc kubenswrapper[4997]: I1205 07:33:28.615039 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wrnn4" Dec 05 07:33:28 crc kubenswrapper[4997]: I1205 07:33:28.617496 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wrnn4" Dec 05 07:33:28 crc kubenswrapper[4997]: I1205 07:33:28.673486 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wrnn4" Dec 05 07:33:28 crc kubenswrapper[4997]: I1205 07:33:28.837259 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wrnn4" Dec 05 07:33:29 crc kubenswrapper[4997]: I1205 07:33:29.453501 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wrnn4"] Dec 05 07:33:30 crc kubenswrapper[4997]: I1205 07:33:30.799388 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wrnn4" podUID="287fa72b-6f52-472f-8178-519dec66712a" containerName="registry-server" containerID="cri-o://64b483de4e5057ab3544126edfa8474f1145cbc0662e49a14ba73ca8e4c28ccb" gracePeriod=2 Dec 05 07:33:31 crc kubenswrapper[4997]: I1205 07:33:31.810178 4997 generic.go:334] "Generic (PLEG): container finished" podID="287fa72b-6f52-472f-8178-519dec66712a" containerID="64b483de4e5057ab3544126edfa8474f1145cbc0662e49a14ba73ca8e4c28ccb" exitCode=0 Dec 05 07:33:31 crc kubenswrapper[4997]: I1205 07:33:31.810275 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wrnn4" event={"ID":"287fa72b-6f52-472f-8178-519dec66712a","Type":"ContainerDied","Data":"64b483de4e5057ab3544126edfa8474f1145cbc0662e49a14ba73ca8e4c28ccb"} Dec 05 07:33:32 crc kubenswrapper[4997]: I1205 07:33:32.366645 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wrnn4" Dec 05 07:33:32 crc kubenswrapper[4997]: I1205 07:33:32.539032 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4sw9\" (UniqueName: \"kubernetes.io/projected/287fa72b-6f52-472f-8178-519dec66712a-kube-api-access-m4sw9\") pod \"287fa72b-6f52-472f-8178-519dec66712a\" (UID: \"287fa72b-6f52-472f-8178-519dec66712a\") " Dec 05 07:33:32 crc kubenswrapper[4997]: I1205 07:33:32.539092 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/287fa72b-6f52-472f-8178-519dec66712a-catalog-content\") pod \"287fa72b-6f52-472f-8178-519dec66712a\" (UID: \"287fa72b-6f52-472f-8178-519dec66712a\") " Dec 05 07:33:32 crc kubenswrapper[4997]: I1205 07:33:32.539284 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/287fa72b-6f52-472f-8178-519dec66712a-utilities\") pod \"287fa72b-6f52-472f-8178-519dec66712a\" (UID: \"287fa72b-6f52-472f-8178-519dec66712a\") " Dec 05 07:33:32 crc kubenswrapper[4997]: I1205 07:33:32.540624 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/287fa72b-6f52-472f-8178-519dec66712a-utilities" (OuterVolumeSpecName: "utilities") pod "287fa72b-6f52-472f-8178-519dec66712a" (UID: "287fa72b-6f52-472f-8178-519dec66712a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:33:32 crc kubenswrapper[4997]: I1205 07:33:32.551006 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/287fa72b-6f52-472f-8178-519dec66712a-kube-api-access-m4sw9" (OuterVolumeSpecName: "kube-api-access-m4sw9") pod "287fa72b-6f52-472f-8178-519dec66712a" (UID: "287fa72b-6f52-472f-8178-519dec66712a"). InnerVolumeSpecName "kube-api-access-m4sw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:33:32 crc kubenswrapper[4997]: I1205 07:33:32.592645 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/287fa72b-6f52-472f-8178-519dec66712a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "287fa72b-6f52-472f-8178-519dec66712a" (UID: "287fa72b-6f52-472f-8178-519dec66712a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:33:32 crc kubenswrapper[4997]: I1205 07:33:32.641888 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/287fa72b-6f52-472f-8178-519dec66712a-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:33:32 crc kubenswrapper[4997]: I1205 07:33:32.641928 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4sw9\" (UniqueName: \"kubernetes.io/projected/287fa72b-6f52-472f-8178-519dec66712a-kube-api-access-m4sw9\") on node \"crc\" DevicePath \"\"" Dec 05 07:33:32 crc kubenswrapper[4997]: I1205 07:33:32.641941 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/287fa72b-6f52-472f-8178-519dec66712a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:33:32 crc kubenswrapper[4997]: I1205 07:33:32.822788 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wrnn4" event={"ID":"287fa72b-6f52-472f-8178-519dec66712a","Type":"ContainerDied","Data":"97edcca745baedde49b73a2650742d3e5ebb08b3bc889a75d884480d14a6b94b"} Dec 05 07:33:32 crc kubenswrapper[4997]: I1205 07:33:32.822878 4997 scope.go:117] "RemoveContainer" containerID="64b483de4e5057ab3544126edfa8474f1145cbc0662e49a14ba73ca8e4c28ccb" Dec 05 07:33:32 crc kubenswrapper[4997]: I1205 07:33:32.822892 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wrnn4" Dec 05 07:33:32 crc kubenswrapper[4997]: I1205 07:33:32.855343 4997 scope.go:117] "RemoveContainer" containerID="bca1960575a8385ba4aea3b543f55e5ce34331e4a811162e0d0eb49884a53baa" Dec 05 07:33:32 crc kubenswrapper[4997]: I1205 07:33:32.881346 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wrnn4"] Dec 05 07:33:32 crc kubenswrapper[4997]: I1205 07:33:32.895439 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wrnn4"] Dec 05 07:33:32 crc kubenswrapper[4997]: I1205 07:33:32.905879 4997 scope.go:117] "RemoveContainer" containerID="bcc9f34a74fb5b36b855a21fefef3b5c0f4c2067b496a827e7cd6914441dd241" Dec 05 07:33:33 crc kubenswrapper[4997]: I1205 07:33:33.761448 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="287fa72b-6f52-472f-8178-519dec66712a" path="/var/lib/kubelet/pods/287fa72b-6f52-472f-8178-519dec66712a/volumes" Dec 05 07:33:49 crc kubenswrapper[4997]: I1205 07:33:49.769823 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:33:49 crc kubenswrapper[4997]: I1205 07:33:49.770843 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:34:19 crc kubenswrapper[4997]: I1205 07:34:19.771043 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:34:19 crc kubenswrapper[4997]: I1205 07:34:19.771814 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:34:49 crc kubenswrapper[4997]: I1205 07:34:49.770669 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:34:49 crc kubenswrapper[4997]: I1205 07:34:49.772806 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:34:49 crc kubenswrapper[4997]: I1205 07:34:49.772893 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 07:34:49 crc kubenswrapper[4997]: I1205 07:34:49.773544 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 07:34:49 crc kubenswrapper[4997]: I1205 07:34:49.773639 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" gracePeriod=600 Dec 05 07:34:49 crc kubenswrapper[4997]: E1205 07:34:49.914227 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:34:50 crc kubenswrapper[4997]: I1205 07:34:50.582551 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" exitCode=0 Dec 05 07:34:50 crc kubenswrapper[4997]: I1205 07:34:50.582651 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899"} Dec 05 07:34:50 crc kubenswrapper[4997]: I1205 07:34:50.583139 4997 scope.go:117] "RemoveContainer" containerID="67fe830b9e170b36bd5960d524375aebc16b702b52c6551ed6fc753345c6ff89" Dec 05 07:34:50 crc kubenswrapper[4997]: I1205 07:34:50.583812 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:34:50 crc kubenswrapper[4997]: E1205 07:34:50.584147 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:35:01 crc kubenswrapper[4997]: I1205 07:35:01.749739 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:35:01 crc kubenswrapper[4997]: E1205 07:35:01.750734 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:35:12 crc kubenswrapper[4997]: I1205 07:35:12.749869 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:35:12 crc kubenswrapper[4997]: E1205 07:35:12.751021 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:35:24 crc kubenswrapper[4997]: I1205 07:35:24.749679 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:35:24 crc kubenswrapper[4997]: E1205 07:35:24.750867 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:35:38 crc kubenswrapper[4997]: I1205 07:35:38.750516 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:35:38 crc kubenswrapper[4997]: E1205 07:35:38.752354 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:35:52 crc kubenswrapper[4997]: I1205 07:35:52.749946 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:35:52 crc kubenswrapper[4997]: E1205 07:35:52.750983 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:36:05 crc kubenswrapper[4997]: I1205 07:36:05.750256 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:36:05 crc kubenswrapper[4997]: E1205 07:36:05.753534 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:36:18 crc kubenswrapper[4997]: I1205 07:36:18.750418 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:36:18 crc kubenswrapper[4997]: E1205 07:36:18.751270 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:36:31 crc kubenswrapper[4997]: I1205 07:36:31.750299 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:36:31 crc kubenswrapper[4997]: E1205 07:36:31.751822 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:36:42 crc kubenswrapper[4997]: I1205 07:36:42.749528 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:36:42 crc kubenswrapper[4997]: E1205 07:36:42.750926 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:36:54 crc kubenswrapper[4997]: I1205 07:36:54.750319 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:36:54 crc kubenswrapper[4997]: E1205 07:36:54.751537 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:37:05 crc kubenswrapper[4997]: I1205 07:37:05.749784 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:37:05 crc kubenswrapper[4997]: E1205 07:37:05.751191 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:37:20 crc kubenswrapper[4997]: I1205 07:37:20.749196 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:37:20 crc kubenswrapper[4997]: E1205 07:37:20.750563 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:37:33 crc kubenswrapper[4997]: I1205 07:37:33.749140 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:37:33 crc kubenswrapper[4997]: E1205 07:37:33.750245 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:37:45 crc kubenswrapper[4997]: I1205 07:37:45.749224 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:37:45 crc kubenswrapper[4997]: E1205 07:37:45.750113 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:37:58 crc kubenswrapper[4997]: I1205 07:37:58.749959 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:37:58 crc kubenswrapper[4997]: E1205 07:37:58.751074 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:38:09 crc kubenswrapper[4997]: I1205 07:38:09.757977 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:38:09 crc kubenswrapper[4997]: E1205 07:38:09.758958 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:38:24 crc kubenswrapper[4997]: I1205 07:38:24.749232 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:38:24 crc kubenswrapper[4997]: E1205 07:38:24.750499 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:38:38 crc kubenswrapper[4997]: I1205 07:38:38.750798 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:38:38 crc kubenswrapper[4997]: E1205 07:38:38.752192 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:38:52 crc kubenswrapper[4997]: I1205 07:38:52.750762 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:38:52 crc kubenswrapper[4997]: E1205 07:38:52.752359 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:39:04 crc kubenswrapper[4997]: I1205 07:39:04.749708 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:39:04 crc kubenswrapper[4997]: E1205 07:39:04.751015 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:39:17 crc kubenswrapper[4997]: I1205 07:39:17.749143 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:39:17 crc kubenswrapper[4997]: E1205 07:39:17.751994 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:39:31 crc kubenswrapper[4997]: I1205 07:39:31.748651 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:39:31 crc kubenswrapper[4997]: E1205 07:39:31.749252 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:39:42 crc kubenswrapper[4997]: I1205 07:39:42.748897 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:39:42 crc kubenswrapper[4997]: E1205 07:39:42.749671 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:39:54 crc kubenswrapper[4997]: I1205 07:39:54.749292 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:39:55 crc kubenswrapper[4997]: I1205 07:39:55.675390 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"f5170be035deef2d3f0f609623faa4e7d32aea6b730d5f3deee9491682e1ccbf"} Dec 05 07:42:19 crc kubenswrapper[4997]: I1205 07:42:19.770527 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:42:19 crc kubenswrapper[4997]: I1205 07:42:19.771371 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:42:37 crc kubenswrapper[4997]: I1205 07:42:37.986501 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ztn29"] Dec 05 07:42:37 crc kubenswrapper[4997]: E1205 07:42:37.989356 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="287fa72b-6f52-472f-8178-519dec66712a" containerName="extract-content" Dec 05 07:42:37 crc kubenswrapper[4997]: I1205 07:42:37.989390 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="287fa72b-6f52-472f-8178-519dec66712a" containerName="extract-content" Dec 05 07:42:37 crc kubenswrapper[4997]: E1205 07:42:37.989412 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="287fa72b-6f52-472f-8178-519dec66712a" containerName="registry-server" Dec 05 07:42:37 crc kubenswrapper[4997]: I1205 07:42:37.989420 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="287fa72b-6f52-472f-8178-519dec66712a" containerName="registry-server" Dec 05 07:42:37 crc kubenswrapper[4997]: E1205 07:42:37.989439 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="287fa72b-6f52-472f-8178-519dec66712a" containerName="extract-utilities" Dec 05 07:42:37 crc kubenswrapper[4997]: I1205 07:42:37.989447 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="287fa72b-6f52-472f-8178-519dec66712a" containerName="extract-utilities" Dec 05 07:42:37 crc kubenswrapper[4997]: I1205 07:42:37.989577 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="287fa72b-6f52-472f-8178-519dec66712a" containerName="registry-server" Dec 05 07:42:37 crc kubenswrapper[4997]: I1205 07:42:37.990726 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ztn29" Dec 05 07:42:38 crc kubenswrapper[4997]: I1205 07:42:38.003581 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ztn29"] Dec 05 07:42:38 crc kubenswrapper[4997]: I1205 07:42:38.073704 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1101ad6-9f6c-4083-a8cc-a1df2400108f-catalog-content\") pod \"community-operators-ztn29\" (UID: \"b1101ad6-9f6c-4083-a8cc-a1df2400108f\") " pod="openshift-marketplace/community-operators-ztn29" Dec 05 07:42:38 crc kubenswrapper[4997]: I1205 07:42:38.073918 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnlg8\" (UniqueName: \"kubernetes.io/projected/b1101ad6-9f6c-4083-a8cc-a1df2400108f-kube-api-access-bnlg8\") pod \"community-operators-ztn29\" (UID: \"b1101ad6-9f6c-4083-a8cc-a1df2400108f\") " pod="openshift-marketplace/community-operators-ztn29" Dec 05 07:42:38 crc kubenswrapper[4997]: I1205 07:42:38.074013 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1101ad6-9f6c-4083-a8cc-a1df2400108f-utilities\") pod \"community-operators-ztn29\" (UID: \"b1101ad6-9f6c-4083-a8cc-a1df2400108f\") " pod="openshift-marketplace/community-operators-ztn29" Dec 05 07:42:38 crc kubenswrapper[4997]: I1205 07:42:38.176032 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnlg8\" (UniqueName: \"kubernetes.io/projected/b1101ad6-9f6c-4083-a8cc-a1df2400108f-kube-api-access-bnlg8\") pod \"community-operators-ztn29\" (UID: \"b1101ad6-9f6c-4083-a8cc-a1df2400108f\") " pod="openshift-marketplace/community-operators-ztn29" Dec 05 07:42:38 crc kubenswrapper[4997]: I1205 07:42:38.176099 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1101ad6-9f6c-4083-a8cc-a1df2400108f-utilities\") pod \"community-operators-ztn29\" (UID: \"b1101ad6-9f6c-4083-a8cc-a1df2400108f\") " pod="openshift-marketplace/community-operators-ztn29" Dec 05 07:42:38 crc kubenswrapper[4997]: I1205 07:42:38.176168 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1101ad6-9f6c-4083-a8cc-a1df2400108f-catalog-content\") pod \"community-operators-ztn29\" (UID: \"b1101ad6-9f6c-4083-a8cc-a1df2400108f\") " pod="openshift-marketplace/community-operators-ztn29" Dec 05 07:42:38 crc kubenswrapper[4997]: I1205 07:42:38.176749 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1101ad6-9f6c-4083-a8cc-a1df2400108f-catalog-content\") pod \"community-operators-ztn29\" (UID: \"b1101ad6-9f6c-4083-a8cc-a1df2400108f\") " pod="openshift-marketplace/community-operators-ztn29" Dec 05 07:42:38 crc kubenswrapper[4997]: I1205 07:42:38.176857 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1101ad6-9f6c-4083-a8cc-a1df2400108f-utilities\") pod \"community-operators-ztn29\" (UID: \"b1101ad6-9f6c-4083-a8cc-a1df2400108f\") " pod="openshift-marketplace/community-operators-ztn29" Dec 05 07:42:38 crc kubenswrapper[4997]: I1205 07:42:38.200370 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnlg8\" (UniqueName: \"kubernetes.io/projected/b1101ad6-9f6c-4083-a8cc-a1df2400108f-kube-api-access-bnlg8\") pod \"community-operators-ztn29\" (UID: \"b1101ad6-9f6c-4083-a8cc-a1df2400108f\") " pod="openshift-marketplace/community-operators-ztn29" Dec 05 07:42:38 crc kubenswrapper[4997]: I1205 07:42:38.319167 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ztn29" Dec 05 07:42:38 crc kubenswrapper[4997]: I1205 07:42:38.899714 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ztn29"] Dec 05 07:42:39 crc kubenswrapper[4997]: I1205 07:42:39.197804 4997 generic.go:334] "Generic (PLEG): container finished" podID="b1101ad6-9f6c-4083-a8cc-a1df2400108f" containerID="fb7c8c1b767815cc86382421feba788de6baec153d9d67059c98a8275f554982" exitCode=0 Dec 05 07:42:39 crc kubenswrapper[4997]: I1205 07:42:39.197856 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztn29" event={"ID":"b1101ad6-9f6c-4083-a8cc-a1df2400108f","Type":"ContainerDied","Data":"fb7c8c1b767815cc86382421feba788de6baec153d9d67059c98a8275f554982"} Dec 05 07:42:39 crc kubenswrapper[4997]: I1205 07:42:39.197885 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztn29" event={"ID":"b1101ad6-9f6c-4083-a8cc-a1df2400108f","Type":"ContainerStarted","Data":"d6bf5379a1476c1c59b6bd3a04180f633de238fd55d7c7cf708af874d0c74c0d"} Dec 05 07:42:39 crc kubenswrapper[4997]: I1205 07:42:39.200950 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 07:42:40 crc kubenswrapper[4997]: I1205 07:42:40.210424 4997 generic.go:334] "Generic (PLEG): container finished" podID="b1101ad6-9f6c-4083-a8cc-a1df2400108f" containerID="ceba9cb1237f3811065138065ba9eb2432b7b53ee26812070b782476bb4066ea" exitCode=0 Dec 05 07:42:40 crc kubenswrapper[4997]: I1205 07:42:40.210548 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztn29" event={"ID":"b1101ad6-9f6c-4083-a8cc-a1df2400108f","Type":"ContainerDied","Data":"ceba9cb1237f3811065138065ba9eb2432b7b53ee26812070b782476bb4066ea"} Dec 05 07:42:41 crc kubenswrapper[4997]: I1205 07:42:41.224363 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztn29" event={"ID":"b1101ad6-9f6c-4083-a8cc-a1df2400108f","Type":"ContainerStarted","Data":"9d45fa09708a2d6d83166f78ca9d28abe6fc303fab25058cb1251ebffe0a0aa1"} Dec 05 07:42:41 crc kubenswrapper[4997]: I1205 07:42:41.273529 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ztn29" podStartSLOduration=2.853353629 podStartE2EDuration="4.273493815s" podCreationTimestamp="2025-12-05 07:42:37 +0000 UTC" firstStartedPulling="2025-12-05 07:42:39.20066772 +0000 UTC m=+2859.729574991" lastFinishedPulling="2025-12-05 07:42:40.620807906 +0000 UTC m=+2861.149715177" observedRunningTime="2025-12-05 07:42:41.253013133 +0000 UTC m=+2861.781920414" watchObservedRunningTime="2025-12-05 07:42:41.273493815 +0000 UTC m=+2861.802401136" Dec 05 07:42:48 crc kubenswrapper[4997]: I1205 07:42:48.319505 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ztn29" Dec 05 07:42:48 crc kubenswrapper[4997]: I1205 07:42:48.320335 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ztn29" Dec 05 07:42:48 crc kubenswrapper[4997]: I1205 07:42:48.395938 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ztn29" Dec 05 07:42:49 crc kubenswrapper[4997]: I1205 07:42:49.355051 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ztn29" Dec 05 07:42:49 crc kubenswrapper[4997]: I1205 07:42:49.409394 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ztn29"] Dec 05 07:42:49 crc kubenswrapper[4997]: I1205 07:42:49.770265 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:42:49 crc kubenswrapper[4997]: I1205 07:42:49.770345 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:42:51 crc kubenswrapper[4997]: I1205 07:42:51.320650 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ztn29" podUID="b1101ad6-9f6c-4083-a8cc-a1df2400108f" containerName="registry-server" containerID="cri-o://9d45fa09708a2d6d83166f78ca9d28abe6fc303fab25058cb1251ebffe0a0aa1" gracePeriod=2 Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.326318 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ztn29" Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.331967 4997 generic.go:334] "Generic (PLEG): container finished" podID="b1101ad6-9f6c-4083-a8cc-a1df2400108f" containerID="9d45fa09708a2d6d83166f78ca9d28abe6fc303fab25058cb1251ebffe0a0aa1" exitCode=0 Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.332039 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztn29" event={"ID":"b1101ad6-9f6c-4083-a8cc-a1df2400108f","Type":"ContainerDied","Data":"9d45fa09708a2d6d83166f78ca9d28abe6fc303fab25058cb1251ebffe0a0aa1"} Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.332050 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ztn29" Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.332091 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztn29" event={"ID":"b1101ad6-9f6c-4083-a8cc-a1df2400108f","Type":"ContainerDied","Data":"d6bf5379a1476c1c59b6bd3a04180f633de238fd55d7c7cf708af874d0c74c0d"} Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.332117 4997 scope.go:117] "RemoveContainer" containerID="9d45fa09708a2d6d83166f78ca9d28abe6fc303fab25058cb1251ebffe0a0aa1" Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.376059 4997 scope.go:117] "RemoveContainer" containerID="ceba9cb1237f3811065138065ba9eb2432b7b53ee26812070b782476bb4066ea" Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.405884 4997 scope.go:117] "RemoveContainer" containerID="fb7c8c1b767815cc86382421feba788de6baec153d9d67059c98a8275f554982" Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.419969 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1101ad6-9f6c-4083-a8cc-a1df2400108f-catalog-content\") pod \"b1101ad6-9f6c-4083-a8cc-a1df2400108f\" (UID: \"b1101ad6-9f6c-4083-a8cc-a1df2400108f\") " Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.420083 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnlg8\" (UniqueName: \"kubernetes.io/projected/b1101ad6-9f6c-4083-a8cc-a1df2400108f-kube-api-access-bnlg8\") pod \"b1101ad6-9f6c-4083-a8cc-a1df2400108f\" (UID: \"b1101ad6-9f6c-4083-a8cc-a1df2400108f\") " Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.420174 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1101ad6-9f6c-4083-a8cc-a1df2400108f-utilities\") pod \"b1101ad6-9f6c-4083-a8cc-a1df2400108f\" (UID: \"b1101ad6-9f6c-4083-a8cc-a1df2400108f\") " Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.421858 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1101ad6-9f6c-4083-a8cc-a1df2400108f-utilities" (OuterVolumeSpecName: "utilities") pod "b1101ad6-9f6c-4083-a8cc-a1df2400108f" (UID: "b1101ad6-9f6c-4083-a8cc-a1df2400108f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.428896 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1101ad6-9f6c-4083-a8cc-a1df2400108f-kube-api-access-bnlg8" (OuterVolumeSpecName: "kube-api-access-bnlg8") pod "b1101ad6-9f6c-4083-a8cc-a1df2400108f" (UID: "b1101ad6-9f6c-4083-a8cc-a1df2400108f"). InnerVolumeSpecName "kube-api-access-bnlg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.430632 4997 scope.go:117] "RemoveContainer" containerID="9d45fa09708a2d6d83166f78ca9d28abe6fc303fab25058cb1251ebffe0a0aa1" Dec 05 07:42:52 crc kubenswrapper[4997]: E1205 07:42:52.431172 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d45fa09708a2d6d83166f78ca9d28abe6fc303fab25058cb1251ebffe0a0aa1\": container with ID starting with 9d45fa09708a2d6d83166f78ca9d28abe6fc303fab25058cb1251ebffe0a0aa1 not found: ID does not exist" containerID="9d45fa09708a2d6d83166f78ca9d28abe6fc303fab25058cb1251ebffe0a0aa1" Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.431217 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d45fa09708a2d6d83166f78ca9d28abe6fc303fab25058cb1251ebffe0a0aa1"} err="failed to get container status \"9d45fa09708a2d6d83166f78ca9d28abe6fc303fab25058cb1251ebffe0a0aa1\": rpc error: code = NotFound desc = could not find container \"9d45fa09708a2d6d83166f78ca9d28abe6fc303fab25058cb1251ebffe0a0aa1\": container with ID starting with 9d45fa09708a2d6d83166f78ca9d28abe6fc303fab25058cb1251ebffe0a0aa1 not found: ID does not exist" Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.431245 4997 scope.go:117] "RemoveContainer" containerID="ceba9cb1237f3811065138065ba9eb2432b7b53ee26812070b782476bb4066ea" Dec 05 07:42:52 crc kubenswrapper[4997]: E1205 07:42:52.431763 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ceba9cb1237f3811065138065ba9eb2432b7b53ee26812070b782476bb4066ea\": container with ID starting with ceba9cb1237f3811065138065ba9eb2432b7b53ee26812070b782476bb4066ea not found: ID does not exist" containerID="ceba9cb1237f3811065138065ba9eb2432b7b53ee26812070b782476bb4066ea" Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.431816 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceba9cb1237f3811065138065ba9eb2432b7b53ee26812070b782476bb4066ea"} err="failed to get container status \"ceba9cb1237f3811065138065ba9eb2432b7b53ee26812070b782476bb4066ea\": rpc error: code = NotFound desc = could not find container \"ceba9cb1237f3811065138065ba9eb2432b7b53ee26812070b782476bb4066ea\": container with ID starting with ceba9cb1237f3811065138065ba9eb2432b7b53ee26812070b782476bb4066ea not found: ID does not exist" Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.431851 4997 scope.go:117] "RemoveContainer" containerID="fb7c8c1b767815cc86382421feba788de6baec153d9d67059c98a8275f554982" Dec 05 07:42:52 crc kubenswrapper[4997]: E1205 07:42:52.432184 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb7c8c1b767815cc86382421feba788de6baec153d9d67059c98a8275f554982\": container with ID starting with fb7c8c1b767815cc86382421feba788de6baec153d9d67059c98a8275f554982 not found: ID does not exist" containerID="fb7c8c1b767815cc86382421feba788de6baec153d9d67059c98a8275f554982" Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.432218 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb7c8c1b767815cc86382421feba788de6baec153d9d67059c98a8275f554982"} err="failed to get container status \"fb7c8c1b767815cc86382421feba788de6baec153d9d67059c98a8275f554982\": rpc error: code = NotFound desc = could not find container \"fb7c8c1b767815cc86382421feba788de6baec153d9d67059c98a8275f554982\": container with ID starting with fb7c8c1b767815cc86382421feba788de6baec153d9d67059c98a8275f554982 not found: ID does not exist" Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.483859 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1101ad6-9f6c-4083-a8cc-a1df2400108f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1101ad6-9f6c-4083-a8cc-a1df2400108f" (UID: "b1101ad6-9f6c-4083-a8cc-a1df2400108f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.523000 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1101ad6-9f6c-4083-a8cc-a1df2400108f-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.523077 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1101ad6-9f6c-4083-a8cc-a1df2400108f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.523109 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnlg8\" (UniqueName: \"kubernetes.io/projected/b1101ad6-9f6c-4083-a8cc-a1df2400108f-kube-api-access-bnlg8\") on node \"crc\" DevicePath \"\"" Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.689671 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ztn29"] Dec 05 07:42:52 crc kubenswrapper[4997]: I1205 07:42:52.697959 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ztn29"] Dec 05 07:42:53 crc kubenswrapper[4997]: I1205 07:42:53.768166 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1101ad6-9f6c-4083-a8cc-a1df2400108f" path="/var/lib/kubelet/pods/b1101ad6-9f6c-4083-a8cc-a1df2400108f/volumes" Dec 05 07:43:04 crc kubenswrapper[4997]: I1205 07:43:04.935025 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cnlvf"] Dec 05 07:43:04 crc kubenswrapper[4997]: E1205 07:43:04.939116 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1101ad6-9f6c-4083-a8cc-a1df2400108f" containerName="registry-server" Dec 05 07:43:04 crc kubenswrapper[4997]: I1205 07:43:04.939170 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1101ad6-9f6c-4083-a8cc-a1df2400108f" containerName="registry-server" Dec 05 07:43:04 crc kubenswrapper[4997]: E1205 07:43:04.939197 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1101ad6-9f6c-4083-a8cc-a1df2400108f" containerName="extract-utilities" Dec 05 07:43:04 crc kubenswrapper[4997]: I1205 07:43:04.939215 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1101ad6-9f6c-4083-a8cc-a1df2400108f" containerName="extract-utilities" Dec 05 07:43:04 crc kubenswrapper[4997]: E1205 07:43:04.939255 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1101ad6-9f6c-4083-a8cc-a1df2400108f" containerName="extract-content" Dec 05 07:43:04 crc kubenswrapper[4997]: I1205 07:43:04.939271 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1101ad6-9f6c-4083-a8cc-a1df2400108f" containerName="extract-content" Dec 05 07:43:04 crc kubenswrapper[4997]: I1205 07:43:04.939610 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1101ad6-9f6c-4083-a8cc-a1df2400108f" containerName="registry-server" Dec 05 07:43:04 crc kubenswrapper[4997]: I1205 07:43:04.942839 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cnlvf" Dec 05 07:43:04 crc kubenswrapper[4997]: I1205 07:43:04.953208 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cnlvf"] Dec 05 07:43:05 crc kubenswrapper[4997]: I1205 07:43:05.030691 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-utilities\") pod \"redhat-marketplace-cnlvf\" (UID: \"e31bc66d-fde2-4b7f-abcf-8c50aa39c943\") " pod="openshift-marketplace/redhat-marketplace-cnlvf" Dec 05 07:43:05 crc kubenswrapper[4997]: I1205 07:43:05.030772 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-catalog-content\") pod \"redhat-marketplace-cnlvf\" (UID: \"e31bc66d-fde2-4b7f-abcf-8c50aa39c943\") " pod="openshift-marketplace/redhat-marketplace-cnlvf" Dec 05 07:43:05 crc kubenswrapper[4997]: I1205 07:43:05.030819 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqcjk\" (UniqueName: \"kubernetes.io/projected/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-kube-api-access-sqcjk\") pod \"redhat-marketplace-cnlvf\" (UID: \"e31bc66d-fde2-4b7f-abcf-8c50aa39c943\") " pod="openshift-marketplace/redhat-marketplace-cnlvf" Dec 05 07:43:05 crc kubenswrapper[4997]: I1205 07:43:05.132364 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-utilities\") pod \"redhat-marketplace-cnlvf\" (UID: \"e31bc66d-fde2-4b7f-abcf-8c50aa39c943\") " pod="openshift-marketplace/redhat-marketplace-cnlvf" Dec 05 07:43:05 crc kubenswrapper[4997]: I1205 07:43:05.132431 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-catalog-content\") pod \"redhat-marketplace-cnlvf\" (UID: \"e31bc66d-fde2-4b7f-abcf-8c50aa39c943\") " pod="openshift-marketplace/redhat-marketplace-cnlvf" Dec 05 07:43:05 crc kubenswrapper[4997]: I1205 07:43:05.132470 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqcjk\" (UniqueName: \"kubernetes.io/projected/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-kube-api-access-sqcjk\") pod \"redhat-marketplace-cnlvf\" (UID: \"e31bc66d-fde2-4b7f-abcf-8c50aa39c943\") " pod="openshift-marketplace/redhat-marketplace-cnlvf" Dec 05 07:43:05 crc kubenswrapper[4997]: I1205 07:43:05.133037 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-utilities\") pod \"redhat-marketplace-cnlvf\" (UID: \"e31bc66d-fde2-4b7f-abcf-8c50aa39c943\") " pod="openshift-marketplace/redhat-marketplace-cnlvf" Dec 05 07:43:05 crc kubenswrapper[4997]: I1205 07:43:05.133128 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-catalog-content\") pod \"redhat-marketplace-cnlvf\" (UID: \"e31bc66d-fde2-4b7f-abcf-8c50aa39c943\") " pod="openshift-marketplace/redhat-marketplace-cnlvf" Dec 05 07:43:05 crc kubenswrapper[4997]: I1205 07:43:05.155435 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqcjk\" (UniqueName: \"kubernetes.io/projected/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-kube-api-access-sqcjk\") pod \"redhat-marketplace-cnlvf\" (UID: \"e31bc66d-fde2-4b7f-abcf-8c50aa39c943\") " pod="openshift-marketplace/redhat-marketplace-cnlvf" Dec 05 07:43:05 crc kubenswrapper[4997]: I1205 07:43:05.276448 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cnlvf" Dec 05 07:43:05 crc kubenswrapper[4997]: I1205 07:43:05.746812 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cnlvf"] Dec 05 07:43:06 crc kubenswrapper[4997]: I1205 07:43:06.466308 4997 generic.go:334] "Generic (PLEG): container finished" podID="e31bc66d-fde2-4b7f-abcf-8c50aa39c943" containerID="1df3bcb5111576815fc5645567632b7c7fab688dde00513f0c0bd69773ed4d16" exitCode=0 Dec 05 07:43:06 crc kubenswrapper[4997]: I1205 07:43:06.466547 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cnlvf" event={"ID":"e31bc66d-fde2-4b7f-abcf-8c50aa39c943","Type":"ContainerDied","Data":"1df3bcb5111576815fc5645567632b7c7fab688dde00513f0c0bd69773ed4d16"} Dec 05 07:43:06 crc kubenswrapper[4997]: I1205 07:43:06.466819 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cnlvf" event={"ID":"e31bc66d-fde2-4b7f-abcf-8c50aa39c943","Type":"ContainerStarted","Data":"84265c221479174a3ad08c6949b43cce3083570fa5e2f6487dae344074629bee"} Dec 05 07:43:08 crc kubenswrapper[4997]: I1205 07:43:08.488805 4997 generic.go:334] "Generic (PLEG): container finished" podID="e31bc66d-fde2-4b7f-abcf-8c50aa39c943" containerID="ba29c18d9529ac577babe41b39e59734d98f9f64a46a921850521a874df46a2c" exitCode=0 Dec 05 07:43:08 crc kubenswrapper[4997]: I1205 07:43:08.488938 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cnlvf" event={"ID":"e31bc66d-fde2-4b7f-abcf-8c50aa39c943","Type":"ContainerDied","Data":"ba29c18d9529ac577babe41b39e59734d98f9f64a46a921850521a874df46a2c"} Dec 05 07:43:09 crc kubenswrapper[4997]: I1205 07:43:09.501825 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cnlvf" event={"ID":"e31bc66d-fde2-4b7f-abcf-8c50aa39c943","Type":"ContainerStarted","Data":"10102b20931a8f4cce5fbec13c99260644e22ffbd3a1257983bde1bf14ee1408"} Dec 05 07:43:09 crc kubenswrapper[4997]: I1205 07:43:09.537248 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cnlvf" podStartSLOduration=3.101176843 podStartE2EDuration="5.537222813s" podCreationTimestamp="2025-12-05 07:43:04 +0000 UTC" firstStartedPulling="2025-12-05 07:43:06.46949255 +0000 UTC m=+2886.998399841" lastFinishedPulling="2025-12-05 07:43:08.90553851 +0000 UTC m=+2889.434445811" observedRunningTime="2025-12-05 07:43:09.526223317 +0000 UTC m=+2890.055130608" watchObservedRunningTime="2025-12-05 07:43:09.537222813 +0000 UTC m=+2890.066130084" Dec 05 07:43:15 crc kubenswrapper[4997]: I1205 07:43:15.277461 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cnlvf" Dec 05 07:43:15 crc kubenswrapper[4997]: I1205 07:43:15.277756 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cnlvf" Dec 05 07:43:15 crc kubenswrapper[4997]: I1205 07:43:15.329538 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cnlvf" Dec 05 07:43:15 crc kubenswrapper[4997]: I1205 07:43:15.638804 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cnlvf" Dec 05 07:43:15 crc kubenswrapper[4997]: I1205 07:43:15.694251 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cnlvf"] Dec 05 07:43:17 crc kubenswrapper[4997]: I1205 07:43:17.581825 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cnlvf" podUID="e31bc66d-fde2-4b7f-abcf-8c50aa39c943" containerName="registry-server" containerID="cri-o://10102b20931a8f4cce5fbec13c99260644e22ffbd3a1257983bde1bf14ee1408" gracePeriod=2 Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.102486 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cnlvf" Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.131754 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqcjk\" (UniqueName: \"kubernetes.io/projected/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-kube-api-access-sqcjk\") pod \"e31bc66d-fde2-4b7f-abcf-8c50aa39c943\" (UID: \"e31bc66d-fde2-4b7f-abcf-8c50aa39c943\") " Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.131864 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-utilities\") pod \"e31bc66d-fde2-4b7f-abcf-8c50aa39c943\" (UID: \"e31bc66d-fde2-4b7f-abcf-8c50aa39c943\") " Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.131978 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-catalog-content\") pod \"e31bc66d-fde2-4b7f-abcf-8c50aa39c943\" (UID: \"e31bc66d-fde2-4b7f-abcf-8c50aa39c943\") " Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.134829 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-utilities" (OuterVolumeSpecName: "utilities") pod "e31bc66d-fde2-4b7f-abcf-8c50aa39c943" (UID: "e31bc66d-fde2-4b7f-abcf-8c50aa39c943"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.140404 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-kube-api-access-sqcjk" (OuterVolumeSpecName: "kube-api-access-sqcjk") pod "e31bc66d-fde2-4b7f-abcf-8c50aa39c943" (UID: "e31bc66d-fde2-4b7f-abcf-8c50aa39c943"). InnerVolumeSpecName "kube-api-access-sqcjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.156030 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e31bc66d-fde2-4b7f-abcf-8c50aa39c943" (UID: "e31bc66d-fde2-4b7f-abcf-8c50aa39c943"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.234116 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqcjk\" (UniqueName: \"kubernetes.io/projected/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-kube-api-access-sqcjk\") on node \"crc\" DevicePath \"\"" Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.234158 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.234169 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e31bc66d-fde2-4b7f-abcf-8c50aa39c943-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.596923 4997 generic.go:334] "Generic (PLEG): container finished" podID="e31bc66d-fde2-4b7f-abcf-8c50aa39c943" containerID="10102b20931a8f4cce5fbec13c99260644e22ffbd3a1257983bde1bf14ee1408" exitCode=0 Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.597029 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cnlvf" event={"ID":"e31bc66d-fde2-4b7f-abcf-8c50aa39c943","Type":"ContainerDied","Data":"10102b20931a8f4cce5fbec13c99260644e22ffbd3a1257983bde1bf14ee1408"} Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.597049 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cnlvf" Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.599093 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cnlvf" event={"ID":"e31bc66d-fde2-4b7f-abcf-8c50aa39c943","Type":"ContainerDied","Data":"84265c221479174a3ad08c6949b43cce3083570fa5e2f6487dae344074629bee"} Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.599118 4997 scope.go:117] "RemoveContainer" containerID="10102b20931a8f4cce5fbec13c99260644e22ffbd3a1257983bde1bf14ee1408" Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.629188 4997 scope.go:117] "RemoveContainer" containerID="ba29c18d9529ac577babe41b39e59734d98f9f64a46a921850521a874df46a2c" Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.663104 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cnlvf"] Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.672603 4997 scope.go:117] "RemoveContainer" containerID="1df3bcb5111576815fc5645567632b7c7fab688dde00513f0c0bd69773ed4d16" Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.676536 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cnlvf"] Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.703549 4997 scope.go:117] "RemoveContainer" containerID="10102b20931a8f4cce5fbec13c99260644e22ffbd3a1257983bde1bf14ee1408" Dec 05 07:43:18 crc kubenswrapper[4997]: E1205 07:43:18.705165 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10102b20931a8f4cce5fbec13c99260644e22ffbd3a1257983bde1bf14ee1408\": container with ID starting with 10102b20931a8f4cce5fbec13c99260644e22ffbd3a1257983bde1bf14ee1408 not found: ID does not exist" containerID="10102b20931a8f4cce5fbec13c99260644e22ffbd3a1257983bde1bf14ee1408" Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.705248 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10102b20931a8f4cce5fbec13c99260644e22ffbd3a1257983bde1bf14ee1408"} err="failed to get container status \"10102b20931a8f4cce5fbec13c99260644e22ffbd3a1257983bde1bf14ee1408\": rpc error: code = NotFound desc = could not find container \"10102b20931a8f4cce5fbec13c99260644e22ffbd3a1257983bde1bf14ee1408\": container with ID starting with 10102b20931a8f4cce5fbec13c99260644e22ffbd3a1257983bde1bf14ee1408 not found: ID does not exist" Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.705299 4997 scope.go:117] "RemoveContainer" containerID="ba29c18d9529ac577babe41b39e59734d98f9f64a46a921850521a874df46a2c" Dec 05 07:43:18 crc kubenswrapper[4997]: E1205 07:43:18.705845 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba29c18d9529ac577babe41b39e59734d98f9f64a46a921850521a874df46a2c\": container with ID starting with ba29c18d9529ac577babe41b39e59734d98f9f64a46a921850521a874df46a2c not found: ID does not exist" containerID="ba29c18d9529ac577babe41b39e59734d98f9f64a46a921850521a874df46a2c" Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.705908 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba29c18d9529ac577babe41b39e59734d98f9f64a46a921850521a874df46a2c"} err="failed to get container status \"ba29c18d9529ac577babe41b39e59734d98f9f64a46a921850521a874df46a2c\": rpc error: code = NotFound desc = could not find container \"ba29c18d9529ac577babe41b39e59734d98f9f64a46a921850521a874df46a2c\": container with ID starting with ba29c18d9529ac577babe41b39e59734d98f9f64a46a921850521a874df46a2c not found: ID does not exist" Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.705948 4997 scope.go:117] "RemoveContainer" containerID="1df3bcb5111576815fc5645567632b7c7fab688dde00513f0c0bd69773ed4d16" Dec 05 07:43:18 crc kubenswrapper[4997]: E1205 07:43:18.706333 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1df3bcb5111576815fc5645567632b7c7fab688dde00513f0c0bd69773ed4d16\": container with ID starting with 1df3bcb5111576815fc5645567632b7c7fab688dde00513f0c0bd69773ed4d16 not found: ID does not exist" containerID="1df3bcb5111576815fc5645567632b7c7fab688dde00513f0c0bd69773ed4d16" Dec 05 07:43:18 crc kubenswrapper[4997]: I1205 07:43:18.706378 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1df3bcb5111576815fc5645567632b7c7fab688dde00513f0c0bd69773ed4d16"} err="failed to get container status \"1df3bcb5111576815fc5645567632b7c7fab688dde00513f0c0bd69773ed4d16\": rpc error: code = NotFound desc = could not find container \"1df3bcb5111576815fc5645567632b7c7fab688dde00513f0c0bd69773ed4d16\": container with ID starting with 1df3bcb5111576815fc5645567632b7c7fab688dde00513f0c0bd69773ed4d16 not found: ID does not exist" Dec 05 07:43:19 crc kubenswrapper[4997]: I1205 07:43:19.766480 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e31bc66d-fde2-4b7f-abcf-8c50aa39c943" path="/var/lib/kubelet/pods/e31bc66d-fde2-4b7f-abcf-8c50aa39c943/volumes" Dec 05 07:43:19 crc kubenswrapper[4997]: I1205 07:43:19.770257 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:43:19 crc kubenswrapper[4997]: I1205 07:43:19.770820 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:43:19 crc kubenswrapper[4997]: I1205 07:43:19.770913 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 07:43:19 crc kubenswrapper[4997]: I1205 07:43:19.771420 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f5170be035deef2d3f0f609623faa4e7d32aea6b730d5f3deee9491682e1ccbf"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 07:43:19 crc kubenswrapper[4997]: I1205 07:43:19.771515 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://f5170be035deef2d3f0f609623faa4e7d32aea6b730d5f3deee9491682e1ccbf" gracePeriod=600 Dec 05 07:43:20 crc kubenswrapper[4997]: I1205 07:43:20.625233 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="f5170be035deef2d3f0f609623faa4e7d32aea6b730d5f3deee9491682e1ccbf" exitCode=0 Dec 05 07:43:20 crc kubenswrapper[4997]: I1205 07:43:20.625329 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"f5170be035deef2d3f0f609623faa4e7d32aea6b730d5f3deee9491682e1ccbf"} Dec 05 07:43:20 crc kubenswrapper[4997]: I1205 07:43:20.625725 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260"} Dec 05 07:43:20 crc kubenswrapper[4997]: I1205 07:43:20.625759 4997 scope.go:117] "RemoveContainer" containerID="5b3067b6c286d3495dcc6dbadfe6c977ea0efcc9eaf327472fea95003d074899" Dec 05 07:44:39 crc kubenswrapper[4997]: I1205 07:44:39.372487 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7v687"] Dec 05 07:44:39 crc kubenswrapper[4997]: E1205 07:44:39.373572 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31bc66d-fde2-4b7f-abcf-8c50aa39c943" containerName="registry-server" Dec 05 07:44:39 crc kubenswrapper[4997]: I1205 07:44:39.373604 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31bc66d-fde2-4b7f-abcf-8c50aa39c943" containerName="registry-server" Dec 05 07:44:39 crc kubenswrapper[4997]: E1205 07:44:39.373657 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31bc66d-fde2-4b7f-abcf-8c50aa39c943" containerName="extract-utilities" Dec 05 07:44:39 crc kubenswrapper[4997]: I1205 07:44:39.373672 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31bc66d-fde2-4b7f-abcf-8c50aa39c943" containerName="extract-utilities" Dec 05 07:44:39 crc kubenswrapper[4997]: E1205 07:44:39.373720 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31bc66d-fde2-4b7f-abcf-8c50aa39c943" containerName="extract-content" Dec 05 07:44:39 crc kubenswrapper[4997]: I1205 07:44:39.373733 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31bc66d-fde2-4b7f-abcf-8c50aa39c943" containerName="extract-content" Dec 05 07:44:39 crc kubenswrapper[4997]: I1205 07:44:39.373991 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e31bc66d-fde2-4b7f-abcf-8c50aa39c943" containerName="registry-server" Dec 05 07:44:39 crc kubenswrapper[4997]: I1205 07:44:39.375896 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7v687" Dec 05 07:44:39 crc kubenswrapper[4997]: I1205 07:44:39.387475 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7v687"] Dec 05 07:44:39 crc kubenswrapper[4997]: I1205 07:44:39.510206 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-catalog-content\") pod \"certified-operators-7v687\" (UID: \"dc2a7615-7f50-41c2-bf0c-d94f6a61904a\") " pod="openshift-marketplace/certified-operators-7v687" Dec 05 07:44:39 crc kubenswrapper[4997]: I1205 07:44:39.510336 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxcpw\" (UniqueName: \"kubernetes.io/projected/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-kube-api-access-gxcpw\") pod \"certified-operators-7v687\" (UID: \"dc2a7615-7f50-41c2-bf0c-d94f6a61904a\") " pod="openshift-marketplace/certified-operators-7v687" Dec 05 07:44:39 crc kubenswrapper[4997]: I1205 07:44:39.510467 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-utilities\") pod \"certified-operators-7v687\" (UID: \"dc2a7615-7f50-41c2-bf0c-d94f6a61904a\") " pod="openshift-marketplace/certified-operators-7v687" Dec 05 07:44:39 crc kubenswrapper[4997]: I1205 07:44:39.612518 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxcpw\" (UniqueName: \"kubernetes.io/projected/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-kube-api-access-gxcpw\") pod \"certified-operators-7v687\" (UID: \"dc2a7615-7f50-41c2-bf0c-d94f6a61904a\") " pod="openshift-marketplace/certified-operators-7v687" Dec 05 07:44:39 crc kubenswrapper[4997]: I1205 07:44:39.612713 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-utilities\") pod \"certified-operators-7v687\" (UID: \"dc2a7615-7f50-41c2-bf0c-d94f6a61904a\") " pod="openshift-marketplace/certified-operators-7v687" Dec 05 07:44:39 crc kubenswrapper[4997]: I1205 07:44:39.612791 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-catalog-content\") pod \"certified-operators-7v687\" (UID: \"dc2a7615-7f50-41c2-bf0c-d94f6a61904a\") " pod="openshift-marketplace/certified-operators-7v687" Dec 05 07:44:39 crc kubenswrapper[4997]: I1205 07:44:39.613657 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-utilities\") pod \"certified-operators-7v687\" (UID: \"dc2a7615-7f50-41c2-bf0c-d94f6a61904a\") " pod="openshift-marketplace/certified-operators-7v687" Dec 05 07:44:39 crc kubenswrapper[4997]: I1205 07:44:39.613766 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-catalog-content\") pod \"certified-operators-7v687\" (UID: \"dc2a7615-7f50-41c2-bf0c-d94f6a61904a\") " pod="openshift-marketplace/certified-operators-7v687" Dec 05 07:44:39 crc kubenswrapper[4997]: I1205 07:44:39.638976 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxcpw\" (UniqueName: \"kubernetes.io/projected/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-kube-api-access-gxcpw\") pod \"certified-operators-7v687\" (UID: \"dc2a7615-7f50-41c2-bf0c-d94f6a61904a\") " pod="openshift-marketplace/certified-operators-7v687" Dec 05 07:44:39 crc kubenswrapper[4997]: I1205 07:44:39.728013 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7v687" Dec 05 07:44:40 crc kubenswrapper[4997]: I1205 07:44:40.005122 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7v687"] Dec 05 07:44:40 crc kubenswrapper[4997]: I1205 07:44:40.376097 4997 generic.go:334] "Generic (PLEG): container finished" podID="dc2a7615-7f50-41c2-bf0c-d94f6a61904a" containerID="48164d36a31f49b433223a8a8ee97c6515fb0386bdfac93c76d220a464b00b38" exitCode=0 Dec 05 07:44:40 crc kubenswrapper[4997]: I1205 07:44:40.376145 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7v687" event={"ID":"dc2a7615-7f50-41c2-bf0c-d94f6a61904a","Type":"ContainerDied","Data":"48164d36a31f49b433223a8a8ee97c6515fb0386bdfac93c76d220a464b00b38"} Dec 05 07:44:40 crc kubenswrapper[4997]: I1205 07:44:40.376174 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7v687" event={"ID":"dc2a7615-7f50-41c2-bf0c-d94f6a61904a","Type":"ContainerStarted","Data":"baeacb0ad2b01beb54c529f7e45dc1912c9008d17566e2c6cfc2e7d6d17af62a"} Dec 05 07:44:41 crc kubenswrapper[4997]: I1205 07:44:41.387318 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7v687" event={"ID":"dc2a7615-7f50-41c2-bf0c-d94f6a61904a","Type":"ContainerStarted","Data":"f85909f9251e736f66f4cd8fc4f3281439a59923101de3b7ee748185afe0d49f"} Dec 05 07:44:42 crc kubenswrapper[4997]: I1205 07:44:42.400774 4997 generic.go:334] "Generic (PLEG): container finished" podID="dc2a7615-7f50-41c2-bf0c-d94f6a61904a" containerID="f85909f9251e736f66f4cd8fc4f3281439a59923101de3b7ee748185afe0d49f" exitCode=0 Dec 05 07:44:42 crc kubenswrapper[4997]: I1205 07:44:42.400862 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7v687" event={"ID":"dc2a7615-7f50-41c2-bf0c-d94f6a61904a","Type":"ContainerDied","Data":"f85909f9251e736f66f4cd8fc4f3281439a59923101de3b7ee748185afe0d49f"} Dec 05 07:44:43 crc kubenswrapper[4997]: I1205 07:44:43.411993 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7v687" event={"ID":"dc2a7615-7f50-41c2-bf0c-d94f6a61904a","Type":"ContainerStarted","Data":"2889d1cef83690d093710a33eb1a4917340767a0a8ea0e9ebe5d960fe6e7ab5c"} Dec 05 07:44:43 crc kubenswrapper[4997]: I1205 07:44:43.439525 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7v687" podStartSLOduration=2.014107678 podStartE2EDuration="4.439502941s" podCreationTimestamp="2025-12-05 07:44:39 +0000 UTC" firstStartedPulling="2025-12-05 07:44:40.378408517 +0000 UTC m=+2980.907315798" lastFinishedPulling="2025-12-05 07:44:42.80380377 +0000 UTC m=+2983.332711061" observedRunningTime="2025-12-05 07:44:43.437016594 +0000 UTC m=+2983.965923935" watchObservedRunningTime="2025-12-05 07:44:43.439502941 +0000 UTC m=+2983.968410212" Dec 05 07:44:49 crc kubenswrapper[4997]: I1205 07:44:49.729092 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7v687" Dec 05 07:44:49 crc kubenswrapper[4997]: I1205 07:44:49.729849 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7v687" Dec 05 07:44:49 crc kubenswrapper[4997]: I1205 07:44:49.811036 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7v687" Dec 05 07:44:50 crc kubenswrapper[4997]: I1205 07:44:50.556186 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7v687" Dec 05 07:44:50 crc kubenswrapper[4997]: I1205 07:44:50.610643 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7v687"] Dec 05 07:44:52 crc kubenswrapper[4997]: I1205 07:44:52.515879 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7v687" podUID="dc2a7615-7f50-41c2-bf0c-d94f6a61904a" containerName="registry-server" containerID="cri-o://2889d1cef83690d093710a33eb1a4917340767a0a8ea0e9ebe5d960fe6e7ab5c" gracePeriod=2 Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.452741 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7v687" Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.524931 4997 generic.go:334] "Generic (PLEG): container finished" podID="dc2a7615-7f50-41c2-bf0c-d94f6a61904a" containerID="2889d1cef83690d093710a33eb1a4917340767a0a8ea0e9ebe5d960fe6e7ab5c" exitCode=0 Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.524970 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7v687" Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.524975 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7v687" event={"ID":"dc2a7615-7f50-41c2-bf0c-d94f6a61904a","Type":"ContainerDied","Data":"2889d1cef83690d093710a33eb1a4917340767a0a8ea0e9ebe5d960fe6e7ab5c"} Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.525003 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7v687" event={"ID":"dc2a7615-7f50-41c2-bf0c-d94f6a61904a","Type":"ContainerDied","Data":"baeacb0ad2b01beb54c529f7e45dc1912c9008d17566e2c6cfc2e7d6d17af62a"} Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.525019 4997 scope.go:117] "RemoveContainer" containerID="2889d1cef83690d093710a33eb1a4917340767a0a8ea0e9ebe5d960fe6e7ab5c" Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.543945 4997 scope.go:117] "RemoveContainer" containerID="f85909f9251e736f66f4cd8fc4f3281439a59923101de3b7ee748185afe0d49f" Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.547258 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxcpw\" (UniqueName: \"kubernetes.io/projected/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-kube-api-access-gxcpw\") pod \"dc2a7615-7f50-41c2-bf0c-d94f6a61904a\" (UID: \"dc2a7615-7f50-41c2-bf0c-d94f6a61904a\") " Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.547320 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-utilities\") pod \"dc2a7615-7f50-41c2-bf0c-d94f6a61904a\" (UID: \"dc2a7615-7f50-41c2-bf0c-d94f6a61904a\") " Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.547373 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-catalog-content\") pod \"dc2a7615-7f50-41c2-bf0c-d94f6a61904a\" (UID: \"dc2a7615-7f50-41c2-bf0c-d94f6a61904a\") " Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.548336 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-utilities" (OuterVolumeSpecName: "utilities") pod "dc2a7615-7f50-41c2-bf0c-d94f6a61904a" (UID: "dc2a7615-7f50-41c2-bf0c-d94f6a61904a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.552702 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-kube-api-access-gxcpw" (OuterVolumeSpecName: "kube-api-access-gxcpw") pod "dc2a7615-7f50-41c2-bf0c-d94f6a61904a" (UID: "dc2a7615-7f50-41c2-bf0c-d94f6a61904a"). InnerVolumeSpecName "kube-api-access-gxcpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.562854 4997 scope.go:117] "RemoveContainer" containerID="48164d36a31f49b433223a8a8ee97c6515fb0386bdfac93c76d220a464b00b38" Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.604231 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc2a7615-7f50-41c2-bf0c-d94f6a61904a" (UID: "dc2a7615-7f50-41c2-bf0c-d94f6a61904a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.610969 4997 scope.go:117] "RemoveContainer" containerID="2889d1cef83690d093710a33eb1a4917340767a0a8ea0e9ebe5d960fe6e7ab5c" Dec 05 07:44:53 crc kubenswrapper[4997]: E1205 07:44:53.611608 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2889d1cef83690d093710a33eb1a4917340767a0a8ea0e9ebe5d960fe6e7ab5c\": container with ID starting with 2889d1cef83690d093710a33eb1a4917340767a0a8ea0e9ebe5d960fe6e7ab5c not found: ID does not exist" containerID="2889d1cef83690d093710a33eb1a4917340767a0a8ea0e9ebe5d960fe6e7ab5c" Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.611900 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2889d1cef83690d093710a33eb1a4917340767a0a8ea0e9ebe5d960fe6e7ab5c"} err="failed to get container status \"2889d1cef83690d093710a33eb1a4917340767a0a8ea0e9ebe5d960fe6e7ab5c\": rpc error: code = NotFound desc = could not find container \"2889d1cef83690d093710a33eb1a4917340767a0a8ea0e9ebe5d960fe6e7ab5c\": container with ID starting with 2889d1cef83690d093710a33eb1a4917340767a0a8ea0e9ebe5d960fe6e7ab5c not found: ID does not exist" Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.611926 4997 scope.go:117] "RemoveContainer" containerID="f85909f9251e736f66f4cd8fc4f3281439a59923101de3b7ee748185afe0d49f" Dec 05 07:44:53 crc kubenswrapper[4997]: E1205 07:44:53.612539 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f85909f9251e736f66f4cd8fc4f3281439a59923101de3b7ee748185afe0d49f\": container with ID starting with f85909f9251e736f66f4cd8fc4f3281439a59923101de3b7ee748185afe0d49f not found: ID does not exist" containerID="f85909f9251e736f66f4cd8fc4f3281439a59923101de3b7ee748185afe0d49f" Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.612647 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f85909f9251e736f66f4cd8fc4f3281439a59923101de3b7ee748185afe0d49f"} err="failed to get container status \"f85909f9251e736f66f4cd8fc4f3281439a59923101de3b7ee748185afe0d49f\": rpc error: code = NotFound desc = could not find container \"f85909f9251e736f66f4cd8fc4f3281439a59923101de3b7ee748185afe0d49f\": container with ID starting with f85909f9251e736f66f4cd8fc4f3281439a59923101de3b7ee748185afe0d49f not found: ID does not exist" Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.612686 4997 scope.go:117] "RemoveContainer" containerID="48164d36a31f49b433223a8a8ee97c6515fb0386bdfac93c76d220a464b00b38" Dec 05 07:44:53 crc kubenswrapper[4997]: E1205 07:44:53.613054 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48164d36a31f49b433223a8a8ee97c6515fb0386bdfac93c76d220a464b00b38\": container with ID starting with 48164d36a31f49b433223a8a8ee97c6515fb0386bdfac93c76d220a464b00b38 not found: ID does not exist" containerID="48164d36a31f49b433223a8a8ee97c6515fb0386bdfac93c76d220a464b00b38" Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.613080 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48164d36a31f49b433223a8a8ee97c6515fb0386bdfac93c76d220a464b00b38"} err="failed to get container status \"48164d36a31f49b433223a8a8ee97c6515fb0386bdfac93c76d220a464b00b38\": rpc error: code = NotFound desc = could not find container \"48164d36a31f49b433223a8a8ee97c6515fb0386bdfac93c76d220a464b00b38\": container with ID starting with 48164d36a31f49b433223a8a8ee97c6515fb0386bdfac93c76d220a464b00b38 not found: ID does not exist" Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.649188 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxcpw\" (UniqueName: \"kubernetes.io/projected/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-kube-api-access-gxcpw\") on node \"crc\" DevicePath \"\"" Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.649433 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.649507 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc2a7615-7f50-41c2-bf0c-d94f6a61904a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:44:53 crc kubenswrapper[4997]: E1205 07:44:53.844318 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc2a7615_7f50_41c2_bf0c_d94f6a61904a.slice/crio-baeacb0ad2b01beb54c529f7e45dc1912c9008d17566e2c6cfc2e7d6d17af62a\": RecentStats: unable to find data in memory cache]" Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.858585 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7v687"] Dec 05 07:44:53 crc kubenswrapper[4997]: I1205 07:44:53.867551 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7v687"] Dec 05 07:44:55 crc kubenswrapper[4997]: I1205 07:44:55.763246 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc2a7615-7f50-41c2-bf0c-d94f6a61904a" path="/var/lib/kubelet/pods/dc2a7615-7f50-41c2-bf0c-d94f6a61904a/volumes" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.169398 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth"] Dec 05 07:45:00 crc kubenswrapper[4997]: E1205 07:45:00.170452 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc2a7615-7f50-41c2-bf0c-d94f6a61904a" containerName="extract-utilities" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.170473 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc2a7615-7f50-41c2-bf0c-d94f6a61904a" containerName="extract-utilities" Dec 05 07:45:00 crc kubenswrapper[4997]: E1205 07:45:00.170492 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc2a7615-7f50-41c2-bf0c-d94f6a61904a" containerName="registry-server" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.170501 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc2a7615-7f50-41c2-bf0c-d94f6a61904a" containerName="registry-server" Dec 05 07:45:00 crc kubenswrapper[4997]: E1205 07:45:00.170531 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc2a7615-7f50-41c2-bf0c-d94f6a61904a" containerName="extract-content" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.170539 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc2a7615-7f50-41c2-bf0c-d94f6a61904a" containerName="extract-content" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.170788 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc2a7615-7f50-41c2-bf0c-d94f6a61904a" containerName="registry-server" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.171499 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.174255 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.190107 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.198049 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth"] Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.349592 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/898582ab-9959-4f46-8fb8-f363db61cb7f-config-volume\") pod \"collect-profiles-29415345-s5dth\" (UID: \"898582ab-9959-4f46-8fb8-f363db61cb7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.349705 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/898582ab-9959-4f46-8fb8-f363db61cb7f-secret-volume\") pod \"collect-profiles-29415345-s5dth\" (UID: \"898582ab-9959-4f46-8fb8-f363db61cb7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.350136 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxk64\" (UniqueName: \"kubernetes.io/projected/898582ab-9959-4f46-8fb8-f363db61cb7f-kube-api-access-vxk64\") pod \"collect-profiles-29415345-s5dth\" (UID: \"898582ab-9959-4f46-8fb8-f363db61cb7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.452118 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/898582ab-9959-4f46-8fb8-f363db61cb7f-config-volume\") pod \"collect-profiles-29415345-s5dth\" (UID: \"898582ab-9959-4f46-8fb8-f363db61cb7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.452212 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/898582ab-9959-4f46-8fb8-f363db61cb7f-secret-volume\") pod \"collect-profiles-29415345-s5dth\" (UID: \"898582ab-9959-4f46-8fb8-f363db61cb7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.452389 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxk64\" (UniqueName: \"kubernetes.io/projected/898582ab-9959-4f46-8fb8-f363db61cb7f-kube-api-access-vxk64\") pod \"collect-profiles-29415345-s5dth\" (UID: \"898582ab-9959-4f46-8fb8-f363db61cb7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.453797 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/898582ab-9959-4f46-8fb8-f363db61cb7f-config-volume\") pod \"collect-profiles-29415345-s5dth\" (UID: \"898582ab-9959-4f46-8fb8-f363db61cb7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.459177 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/898582ab-9959-4f46-8fb8-f363db61cb7f-secret-volume\") pod \"collect-profiles-29415345-s5dth\" (UID: \"898582ab-9959-4f46-8fb8-f363db61cb7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.482385 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxk64\" (UniqueName: \"kubernetes.io/projected/898582ab-9959-4f46-8fb8-f363db61cb7f-kube-api-access-vxk64\") pod \"collect-profiles-29415345-s5dth\" (UID: \"898582ab-9959-4f46-8fb8-f363db61cb7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth" Dec 05 07:45:00 crc kubenswrapper[4997]: I1205 07:45:00.507430 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth" Dec 05 07:45:01 crc kubenswrapper[4997]: I1205 07:45:01.100444 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth"] Dec 05 07:45:01 crc kubenswrapper[4997]: I1205 07:45:01.635287 4997 generic.go:334] "Generic (PLEG): container finished" podID="898582ab-9959-4f46-8fb8-f363db61cb7f" containerID="c790c32fe98012838b02251f466cc90aa644532caee16c2661f4557eb3fee57c" exitCode=0 Dec 05 07:45:01 crc kubenswrapper[4997]: I1205 07:45:01.635343 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth" event={"ID":"898582ab-9959-4f46-8fb8-f363db61cb7f","Type":"ContainerDied","Data":"c790c32fe98012838b02251f466cc90aa644532caee16c2661f4557eb3fee57c"} Dec 05 07:45:01 crc kubenswrapper[4997]: I1205 07:45:01.635606 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth" event={"ID":"898582ab-9959-4f46-8fb8-f363db61cb7f","Type":"ContainerStarted","Data":"1e2f61a527ca992dc023bc437616b368590492d405e448429ed7380a1da2d8f6"} Dec 05 07:45:02 crc kubenswrapper[4997]: I1205 07:45:02.963357 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth" Dec 05 07:45:02 crc kubenswrapper[4997]: I1205 07:45:02.988539 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/898582ab-9959-4f46-8fb8-f363db61cb7f-config-volume\") pod \"898582ab-9959-4f46-8fb8-f363db61cb7f\" (UID: \"898582ab-9959-4f46-8fb8-f363db61cb7f\") " Dec 05 07:45:02 crc kubenswrapper[4997]: I1205 07:45:02.988921 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxk64\" (UniqueName: \"kubernetes.io/projected/898582ab-9959-4f46-8fb8-f363db61cb7f-kube-api-access-vxk64\") pod \"898582ab-9959-4f46-8fb8-f363db61cb7f\" (UID: \"898582ab-9959-4f46-8fb8-f363db61cb7f\") " Dec 05 07:45:02 crc kubenswrapper[4997]: I1205 07:45:02.988955 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/898582ab-9959-4f46-8fb8-f363db61cb7f-secret-volume\") pod \"898582ab-9959-4f46-8fb8-f363db61cb7f\" (UID: \"898582ab-9959-4f46-8fb8-f363db61cb7f\") " Dec 05 07:45:02 crc kubenswrapper[4997]: I1205 07:45:02.990412 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/898582ab-9959-4f46-8fb8-f363db61cb7f-config-volume" (OuterVolumeSpecName: "config-volume") pod "898582ab-9959-4f46-8fb8-f363db61cb7f" (UID: "898582ab-9959-4f46-8fb8-f363db61cb7f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 07:45:02 crc kubenswrapper[4997]: I1205 07:45:02.998295 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/898582ab-9959-4f46-8fb8-f363db61cb7f-kube-api-access-vxk64" (OuterVolumeSpecName: "kube-api-access-vxk64") pod "898582ab-9959-4f46-8fb8-f363db61cb7f" (UID: "898582ab-9959-4f46-8fb8-f363db61cb7f"). InnerVolumeSpecName "kube-api-access-vxk64". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:45:02 crc kubenswrapper[4997]: I1205 07:45:02.998295 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/898582ab-9959-4f46-8fb8-f363db61cb7f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "898582ab-9959-4f46-8fb8-f363db61cb7f" (UID: "898582ab-9959-4f46-8fb8-f363db61cb7f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 07:45:03 crc kubenswrapper[4997]: I1205 07:45:03.091398 4997 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/898582ab-9959-4f46-8fb8-f363db61cb7f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 07:45:03 crc kubenswrapper[4997]: I1205 07:45:03.091444 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxk64\" (UniqueName: \"kubernetes.io/projected/898582ab-9959-4f46-8fb8-f363db61cb7f-kube-api-access-vxk64\") on node \"crc\" DevicePath \"\"" Dec 05 07:45:03 crc kubenswrapper[4997]: I1205 07:45:03.091458 4997 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/898582ab-9959-4f46-8fb8-f363db61cb7f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 07:45:03 crc kubenswrapper[4997]: I1205 07:45:03.660366 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth" event={"ID":"898582ab-9959-4f46-8fb8-f363db61cb7f","Type":"ContainerDied","Data":"1e2f61a527ca992dc023bc437616b368590492d405e448429ed7380a1da2d8f6"} Dec 05 07:45:03 crc kubenswrapper[4997]: I1205 07:45:03.660418 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e2f61a527ca992dc023bc437616b368590492d405e448429ed7380a1da2d8f6" Dec 05 07:45:03 crc kubenswrapper[4997]: I1205 07:45:03.660482 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth" Dec 05 07:45:04 crc kubenswrapper[4997]: I1205 07:45:04.037052 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs"] Dec 05 07:45:04 crc kubenswrapper[4997]: I1205 07:45:04.041959 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415300-kvfgs"] Dec 05 07:45:05 crc kubenswrapper[4997]: I1205 07:45:05.775948 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85d4925b-c3c9-423d-8bf4-0294e312da5c" path="/var/lib/kubelet/pods/85d4925b-c3c9-423d-8bf4-0294e312da5c/volumes" Dec 05 07:45:11 crc kubenswrapper[4997]: I1205 07:45:11.161042 4997 scope.go:117] "RemoveContainer" containerID="cb4d3efb9e35206f1dade7c89b5f598b5e26bf62b3bbbac259ae49dbcef681e7" Dec 05 07:45:49 crc kubenswrapper[4997]: I1205 07:45:49.770794 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:45:49 crc kubenswrapper[4997]: I1205 07:45:49.771747 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:46:19 crc kubenswrapper[4997]: I1205 07:46:19.770299 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:46:19 crc kubenswrapper[4997]: I1205 07:46:19.770988 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:46:49 crc kubenswrapper[4997]: I1205 07:46:49.770000 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:46:49 crc kubenswrapper[4997]: I1205 07:46:49.770786 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:46:49 crc kubenswrapper[4997]: I1205 07:46:49.770844 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 07:46:49 crc kubenswrapper[4997]: I1205 07:46:49.771934 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 07:46:49 crc kubenswrapper[4997]: I1205 07:46:49.772155 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" gracePeriod=600 Dec 05 07:46:49 crc kubenswrapper[4997]: E1205 07:46:49.908111 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:46:50 crc kubenswrapper[4997]: I1205 07:46:50.604592 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" exitCode=0 Dec 05 07:46:50 crc kubenswrapper[4997]: I1205 07:46:50.604682 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260"} Dec 05 07:46:50 crc kubenswrapper[4997]: I1205 07:46:50.604767 4997 scope.go:117] "RemoveContainer" containerID="f5170be035deef2d3f0f609623faa4e7d32aea6b730d5f3deee9491682e1ccbf" Dec 05 07:46:50 crc kubenswrapper[4997]: I1205 07:46:50.606063 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:46:50 crc kubenswrapper[4997]: E1205 07:46:50.606900 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:47:05 crc kubenswrapper[4997]: I1205 07:47:05.749835 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:47:05 crc kubenswrapper[4997]: E1205 07:47:05.750826 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:47:16 crc kubenswrapper[4997]: I1205 07:47:16.750574 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:47:16 crc kubenswrapper[4997]: E1205 07:47:16.751308 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:47:27 crc kubenswrapper[4997]: I1205 07:47:27.752991 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:47:27 crc kubenswrapper[4997]: E1205 07:47:27.753848 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:47:38 crc kubenswrapper[4997]: I1205 07:47:38.749993 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:47:38 crc kubenswrapper[4997]: E1205 07:47:38.751189 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:47:52 crc kubenswrapper[4997]: I1205 07:47:52.749927 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:47:52 crc kubenswrapper[4997]: E1205 07:47:52.750817 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:48:03 crc kubenswrapper[4997]: I1205 07:48:03.750745 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:48:03 crc kubenswrapper[4997]: E1205 07:48:03.751920 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:48:09 crc kubenswrapper[4997]: I1205 07:48:09.413868 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hdqhc"] Dec 05 07:48:09 crc kubenswrapper[4997]: E1205 07:48:09.414907 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="898582ab-9959-4f46-8fb8-f363db61cb7f" containerName="collect-profiles" Dec 05 07:48:09 crc kubenswrapper[4997]: I1205 07:48:09.414984 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="898582ab-9959-4f46-8fb8-f363db61cb7f" containerName="collect-profiles" Dec 05 07:48:09 crc kubenswrapper[4997]: I1205 07:48:09.416702 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="898582ab-9959-4f46-8fb8-f363db61cb7f" containerName="collect-profiles" Dec 05 07:48:09 crc kubenswrapper[4997]: I1205 07:48:09.418210 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hdqhc" Dec 05 07:48:09 crc kubenswrapper[4997]: I1205 07:48:09.429442 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hdqhc"] Dec 05 07:48:09 crc kubenswrapper[4997]: I1205 07:48:09.603669 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk5d8\" (UniqueName: \"kubernetes.io/projected/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-kube-api-access-jk5d8\") pod \"redhat-operators-hdqhc\" (UID: \"2f40b7e7-af27-43bf-903b-f426e0cf2e1d\") " pod="openshift-marketplace/redhat-operators-hdqhc" Dec 05 07:48:09 crc kubenswrapper[4997]: I1205 07:48:09.603738 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-utilities\") pod \"redhat-operators-hdqhc\" (UID: \"2f40b7e7-af27-43bf-903b-f426e0cf2e1d\") " pod="openshift-marketplace/redhat-operators-hdqhc" Dec 05 07:48:09 crc kubenswrapper[4997]: I1205 07:48:09.603984 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-catalog-content\") pod \"redhat-operators-hdqhc\" (UID: \"2f40b7e7-af27-43bf-903b-f426e0cf2e1d\") " pod="openshift-marketplace/redhat-operators-hdqhc" Dec 05 07:48:09 crc kubenswrapper[4997]: I1205 07:48:09.705198 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-catalog-content\") pod \"redhat-operators-hdqhc\" (UID: \"2f40b7e7-af27-43bf-903b-f426e0cf2e1d\") " pod="openshift-marketplace/redhat-operators-hdqhc" Dec 05 07:48:09 crc kubenswrapper[4997]: I1205 07:48:09.705598 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk5d8\" (UniqueName: \"kubernetes.io/projected/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-kube-api-access-jk5d8\") pod \"redhat-operators-hdqhc\" (UID: \"2f40b7e7-af27-43bf-903b-f426e0cf2e1d\") " pod="openshift-marketplace/redhat-operators-hdqhc" Dec 05 07:48:09 crc kubenswrapper[4997]: I1205 07:48:09.705643 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-utilities\") pod \"redhat-operators-hdqhc\" (UID: \"2f40b7e7-af27-43bf-903b-f426e0cf2e1d\") " pod="openshift-marketplace/redhat-operators-hdqhc" Dec 05 07:48:09 crc kubenswrapper[4997]: I1205 07:48:09.706006 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-catalog-content\") pod \"redhat-operators-hdqhc\" (UID: \"2f40b7e7-af27-43bf-903b-f426e0cf2e1d\") " pod="openshift-marketplace/redhat-operators-hdqhc" Dec 05 07:48:09 crc kubenswrapper[4997]: I1205 07:48:09.706041 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-utilities\") pod \"redhat-operators-hdqhc\" (UID: \"2f40b7e7-af27-43bf-903b-f426e0cf2e1d\") " pod="openshift-marketplace/redhat-operators-hdqhc" Dec 05 07:48:09 crc kubenswrapper[4997]: I1205 07:48:09.729171 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk5d8\" (UniqueName: \"kubernetes.io/projected/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-kube-api-access-jk5d8\") pod \"redhat-operators-hdqhc\" (UID: \"2f40b7e7-af27-43bf-903b-f426e0cf2e1d\") " pod="openshift-marketplace/redhat-operators-hdqhc" Dec 05 07:48:09 crc kubenswrapper[4997]: I1205 07:48:09.743738 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hdqhc" Dec 05 07:48:10 crc kubenswrapper[4997]: I1205 07:48:10.196173 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hdqhc"] Dec 05 07:48:10 crc kubenswrapper[4997]: I1205 07:48:10.391113 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hdqhc" event={"ID":"2f40b7e7-af27-43bf-903b-f426e0cf2e1d","Type":"ContainerStarted","Data":"3fc3f18bccf7c3199ee98d8ba89fd2cc0d90332f1959594c19572a073b84f7fd"} Dec 05 07:48:10 crc kubenswrapper[4997]: I1205 07:48:10.391159 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hdqhc" event={"ID":"2f40b7e7-af27-43bf-903b-f426e0cf2e1d","Type":"ContainerStarted","Data":"99a07cc3a36a338eb0130d8a30b380543cd09892c2903fdaf543bcc5fb2b1616"} Dec 05 07:48:11 crc kubenswrapper[4997]: I1205 07:48:11.402323 4997 generic.go:334] "Generic (PLEG): container finished" podID="2f40b7e7-af27-43bf-903b-f426e0cf2e1d" containerID="3fc3f18bccf7c3199ee98d8ba89fd2cc0d90332f1959594c19572a073b84f7fd" exitCode=0 Dec 05 07:48:11 crc kubenswrapper[4997]: I1205 07:48:11.402409 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hdqhc" event={"ID":"2f40b7e7-af27-43bf-903b-f426e0cf2e1d","Type":"ContainerDied","Data":"3fc3f18bccf7c3199ee98d8ba89fd2cc0d90332f1959594c19572a073b84f7fd"} Dec 05 07:48:11 crc kubenswrapper[4997]: I1205 07:48:11.405805 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 07:48:12 crc kubenswrapper[4997]: I1205 07:48:12.412968 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hdqhc" event={"ID":"2f40b7e7-af27-43bf-903b-f426e0cf2e1d","Type":"ContainerStarted","Data":"0f99de4f9fc2cab103de26b1cfaeb0e57915151469ebab6b8ff996fe88a2872c"} Dec 05 07:48:13 crc kubenswrapper[4997]: I1205 07:48:13.423275 4997 generic.go:334] "Generic (PLEG): container finished" podID="2f40b7e7-af27-43bf-903b-f426e0cf2e1d" containerID="0f99de4f9fc2cab103de26b1cfaeb0e57915151469ebab6b8ff996fe88a2872c" exitCode=0 Dec 05 07:48:13 crc kubenswrapper[4997]: I1205 07:48:13.423393 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hdqhc" event={"ID":"2f40b7e7-af27-43bf-903b-f426e0cf2e1d","Type":"ContainerDied","Data":"0f99de4f9fc2cab103de26b1cfaeb0e57915151469ebab6b8ff996fe88a2872c"} Dec 05 07:48:14 crc kubenswrapper[4997]: I1205 07:48:14.439479 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hdqhc" event={"ID":"2f40b7e7-af27-43bf-903b-f426e0cf2e1d","Type":"ContainerStarted","Data":"17d236030d072b37fb7930cf0f312b207b745f6fb2c789e39b4442973424262b"} Dec 05 07:48:14 crc kubenswrapper[4997]: I1205 07:48:14.466545 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hdqhc" podStartSLOduration=3.083857765 podStartE2EDuration="5.466510867s" podCreationTimestamp="2025-12-05 07:48:09 +0000 UTC" firstStartedPulling="2025-12-05 07:48:11.405388576 +0000 UTC m=+3191.934295857" lastFinishedPulling="2025-12-05 07:48:13.788041698 +0000 UTC m=+3194.316948959" observedRunningTime="2025-12-05 07:48:14.463251379 +0000 UTC m=+3194.992158680" watchObservedRunningTime="2025-12-05 07:48:14.466510867 +0000 UTC m=+3194.995418128" Dec 05 07:48:18 crc kubenswrapper[4997]: I1205 07:48:18.750524 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:48:18 crc kubenswrapper[4997]: E1205 07:48:18.751247 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:48:19 crc kubenswrapper[4997]: I1205 07:48:19.744676 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hdqhc" Dec 05 07:48:19 crc kubenswrapper[4997]: I1205 07:48:19.744813 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hdqhc" Dec 05 07:48:20 crc kubenswrapper[4997]: I1205 07:48:20.806730 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hdqhc" podUID="2f40b7e7-af27-43bf-903b-f426e0cf2e1d" containerName="registry-server" probeResult="failure" output=< Dec 05 07:48:20 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 07:48:20 crc kubenswrapper[4997]: > Dec 05 07:48:29 crc kubenswrapper[4997]: I1205 07:48:29.802245 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hdqhc" Dec 05 07:48:29 crc kubenswrapper[4997]: I1205 07:48:29.874452 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hdqhc" Dec 05 07:48:31 crc kubenswrapper[4997]: I1205 07:48:31.749489 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:48:31 crc kubenswrapper[4997]: E1205 07:48:31.750297 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:48:33 crc kubenswrapper[4997]: I1205 07:48:33.262885 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hdqhc"] Dec 05 07:48:33 crc kubenswrapper[4997]: I1205 07:48:33.263652 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hdqhc" podUID="2f40b7e7-af27-43bf-903b-f426e0cf2e1d" containerName="registry-server" containerID="cri-o://17d236030d072b37fb7930cf0f312b207b745f6fb2c789e39b4442973424262b" gracePeriod=2 Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.458914 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hdqhc" Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.653886 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-catalog-content\") pod \"2f40b7e7-af27-43bf-903b-f426e0cf2e1d\" (UID: \"2f40b7e7-af27-43bf-903b-f426e0cf2e1d\") " Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.654074 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk5d8\" (UniqueName: \"kubernetes.io/projected/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-kube-api-access-jk5d8\") pod \"2f40b7e7-af27-43bf-903b-f426e0cf2e1d\" (UID: \"2f40b7e7-af27-43bf-903b-f426e0cf2e1d\") " Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.654221 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-utilities\") pod \"2f40b7e7-af27-43bf-903b-f426e0cf2e1d\" (UID: \"2f40b7e7-af27-43bf-903b-f426e0cf2e1d\") " Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.655504 4997 generic.go:334] "Generic (PLEG): container finished" podID="2f40b7e7-af27-43bf-903b-f426e0cf2e1d" containerID="17d236030d072b37fb7930cf0f312b207b745f6fb2c789e39b4442973424262b" exitCode=0 Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.655598 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hdqhc" event={"ID":"2f40b7e7-af27-43bf-903b-f426e0cf2e1d","Type":"ContainerDied","Data":"17d236030d072b37fb7930cf0f312b207b745f6fb2c789e39b4442973424262b"} Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.655706 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hdqhc" Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.655773 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hdqhc" event={"ID":"2f40b7e7-af27-43bf-903b-f426e0cf2e1d","Type":"ContainerDied","Data":"99a07cc3a36a338eb0130d8a30b380543cd09892c2903fdaf543bcc5fb2b1616"} Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.655865 4997 scope.go:117] "RemoveContainer" containerID="17d236030d072b37fb7930cf0f312b207b745f6fb2c789e39b4442973424262b" Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.658400 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-utilities" (OuterVolumeSpecName: "utilities") pod "2f40b7e7-af27-43bf-903b-f426e0cf2e1d" (UID: "2f40b7e7-af27-43bf-903b-f426e0cf2e1d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.665494 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-kube-api-access-jk5d8" (OuterVolumeSpecName: "kube-api-access-jk5d8") pod "2f40b7e7-af27-43bf-903b-f426e0cf2e1d" (UID: "2f40b7e7-af27-43bf-903b-f426e0cf2e1d"). InnerVolumeSpecName "kube-api-access-jk5d8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.729868 4997 scope.go:117] "RemoveContainer" containerID="0f99de4f9fc2cab103de26b1cfaeb0e57915151469ebab6b8ff996fe88a2872c" Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.755854 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.755903 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk5d8\" (UniqueName: \"kubernetes.io/projected/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-kube-api-access-jk5d8\") on node \"crc\" DevicePath \"\"" Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.761189 4997 scope.go:117] "RemoveContainer" containerID="3fc3f18bccf7c3199ee98d8ba89fd2cc0d90332f1959594c19572a073b84f7fd" Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.787921 4997 scope.go:117] "RemoveContainer" containerID="17d236030d072b37fb7930cf0f312b207b745f6fb2c789e39b4442973424262b" Dec 05 07:48:34 crc kubenswrapper[4997]: E1205 07:48:34.788541 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17d236030d072b37fb7930cf0f312b207b745f6fb2c789e39b4442973424262b\": container with ID starting with 17d236030d072b37fb7930cf0f312b207b745f6fb2c789e39b4442973424262b not found: ID does not exist" containerID="17d236030d072b37fb7930cf0f312b207b745f6fb2c789e39b4442973424262b" Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.788610 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17d236030d072b37fb7930cf0f312b207b745f6fb2c789e39b4442973424262b"} err="failed to get container status \"17d236030d072b37fb7930cf0f312b207b745f6fb2c789e39b4442973424262b\": rpc error: code = NotFound desc = could not find container \"17d236030d072b37fb7930cf0f312b207b745f6fb2c789e39b4442973424262b\": container with ID starting with 17d236030d072b37fb7930cf0f312b207b745f6fb2c789e39b4442973424262b not found: ID does not exist" Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.788707 4997 scope.go:117] "RemoveContainer" containerID="0f99de4f9fc2cab103de26b1cfaeb0e57915151469ebab6b8ff996fe88a2872c" Dec 05 07:48:34 crc kubenswrapper[4997]: E1205 07:48:34.789151 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f99de4f9fc2cab103de26b1cfaeb0e57915151469ebab6b8ff996fe88a2872c\": container with ID starting with 0f99de4f9fc2cab103de26b1cfaeb0e57915151469ebab6b8ff996fe88a2872c not found: ID does not exist" containerID="0f99de4f9fc2cab103de26b1cfaeb0e57915151469ebab6b8ff996fe88a2872c" Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.789200 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f99de4f9fc2cab103de26b1cfaeb0e57915151469ebab6b8ff996fe88a2872c"} err="failed to get container status \"0f99de4f9fc2cab103de26b1cfaeb0e57915151469ebab6b8ff996fe88a2872c\": rpc error: code = NotFound desc = could not find container \"0f99de4f9fc2cab103de26b1cfaeb0e57915151469ebab6b8ff996fe88a2872c\": container with ID starting with 0f99de4f9fc2cab103de26b1cfaeb0e57915151469ebab6b8ff996fe88a2872c not found: ID does not exist" Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.789232 4997 scope.go:117] "RemoveContainer" containerID="3fc3f18bccf7c3199ee98d8ba89fd2cc0d90332f1959594c19572a073b84f7fd" Dec 05 07:48:34 crc kubenswrapper[4997]: E1205 07:48:34.789560 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fc3f18bccf7c3199ee98d8ba89fd2cc0d90332f1959594c19572a073b84f7fd\": container with ID starting with 3fc3f18bccf7c3199ee98d8ba89fd2cc0d90332f1959594c19572a073b84f7fd not found: ID does not exist" containerID="3fc3f18bccf7c3199ee98d8ba89fd2cc0d90332f1959594c19572a073b84f7fd" Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.789721 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fc3f18bccf7c3199ee98d8ba89fd2cc0d90332f1959594c19572a073b84f7fd"} err="failed to get container status \"3fc3f18bccf7c3199ee98d8ba89fd2cc0d90332f1959594c19572a073b84f7fd\": rpc error: code = NotFound desc = could not find container \"3fc3f18bccf7c3199ee98d8ba89fd2cc0d90332f1959594c19572a073b84f7fd\": container with ID starting with 3fc3f18bccf7c3199ee98d8ba89fd2cc0d90332f1959594c19572a073b84f7fd not found: ID does not exist" Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.830712 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f40b7e7-af27-43bf-903b-f426e0cf2e1d" (UID: "2f40b7e7-af27-43bf-903b-f426e0cf2e1d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:48:34 crc kubenswrapper[4997]: I1205 07:48:34.858352 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f40b7e7-af27-43bf-903b-f426e0cf2e1d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:48:35 crc kubenswrapper[4997]: I1205 07:48:35.013812 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hdqhc"] Dec 05 07:48:35 crc kubenswrapper[4997]: I1205 07:48:35.024802 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hdqhc"] Dec 05 07:48:35 crc kubenswrapper[4997]: I1205 07:48:35.764564 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f40b7e7-af27-43bf-903b-f426e0cf2e1d" path="/var/lib/kubelet/pods/2f40b7e7-af27-43bf-903b-f426e0cf2e1d/volumes" Dec 05 07:48:44 crc kubenswrapper[4997]: I1205 07:48:44.750253 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:48:44 crc kubenswrapper[4997]: E1205 07:48:44.751906 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:48:58 crc kubenswrapper[4997]: I1205 07:48:58.749883 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:48:58 crc kubenswrapper[4997]: E1205 07:48:58.751070 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:49:10 crc kubenswrapper[4997]: I1205 07:49:10.749823 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:49:10 crc kubenswrapper[4997]: E1205 07:49:10.751399 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:49:23 crc kubenswrapper[4997]: I1205 07:49:23.749191 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:49:23 crc kubenswrapper[4997]: E1205 07:49:23.750359 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:49:38 crc kubenswrapper[4997]: I1205 07:49:38.748910 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:49:38 crc kubenswrapper[4997]: E1205 07:49:38.749891 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:49:51 crc kubenswrapper[4997]: I1205 07:49:51.749738 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:49:51 crc kubenswrapper[4997]: E1205 07:49:51.750661 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:50:06 crc kubenswrapper[4997]: I1205 07:50:06.749042 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:50:06 crc kubenswrapper[4997]: E1205 07:50:06.751154 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:50:20 crc kubenswrapper[4997]: I1205 07:50:20.749718 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:50:20 crc kubenswrapper[4997]: E1205 07:50:20.750431 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:50:33 crc kubenswrapper[4997]: I1205 07:50:33.799501 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:50:33 crc kubenswrapper[4997]: E1205 07:50:33.800943 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:50:44 crc kubenswrapper[4997]: I1205 07:50:44.749851 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:50:44 crc kubenswrapper[4997]: E1205 07:50:44.750812 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:50:56 crc kubenswrapper[4997]: I1205 07:50:56.749061 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:50:56 crc kubenswrapper[4997]: E1205 07:50:56.749745 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:51:11 crc kubenswrapper[4997]: I1205 07:51:11.751641 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:51:11 crc kubenswrapper[4997]: E1205 07:51:11.752386 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:51:25 crc kubenswrapper[4997]: I1205 07:51:25.749262 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:51:25 crc kubenswrapper[4997]: E1205 07:51:25.749883 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:51:39 crc kubenswrapper[4997]: I1205 07:51:39.761212 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:51:39 crc kubenswrapper[4997]: E1205 07:51:39.762366 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:51:50 crc kubenswrapper[4997]: I1205 07:51:50.750700 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:51:51 crc kubenswrapper[4997]: I1205 07:51:51.631690 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"75a9857f211838704681a7a4d52f9d31b2c917ba74ae7fa865ccbb5f50ba8e7f"} Dec 05 07:53:16 crc kubenswrapper[4997]: I1205 07:53:16.830281 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ppcmm"] Dec 05 07:53:16 crc kubenswrapper[4997]: E1205 07:53:16.831199 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f40b7e7-af27-43bf-903b-f426e0cf2e1d" containerName="extract-content" Dec 05 07:53:16 crc kubenswrapper[4997]: I1205 07:53:16.831221 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f40b7e7-af27-43bf-903b-f426e0cf2e1d" containerName="extract-content" Dec 05 07:53:16 crc kubenswrapper[4997]: E1205 07:53:16.831240 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f40b7e7-af27-43bf-903b-f426e0cf2e1d" containerName="extract-utilities" Dec 05 07:53:16 crc kubenswrapper[4997]: I1205 07:53:16.831249 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f40b7e7-af27-43bf-903b-f426e0cf2e1d" containerName="extract-utilities" Dec 05 07:53:16 crc kubenswrapper[4997]: E1205 07:53:16.831266 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f40b7e7-af27-43bf-903b-f426e0cf2e1d" containerName="registry-server" Dec 05 07:53:16 crc kubenswrapper[4997]: I1205 07:53:16.831274 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f40b7e7-af27-43bf-903b-f426e0cf2e1d" containerName="registry-server" Dec 05 07:53:16 crc kubenswrapper[4997]: I1205 07:53:16.831467 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f40b7e7-af27-43bf-903b-f426e0cf2e1d" containerName="registry-server" Dec 05 07:53:16 crc kubenswrapper[4997]: I1205 07:53:16.832820 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppcmm" Dec 05 07:53:16 crc kubenswrapper[4997]: I1205 07:53:16.849116 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h65r\" (UniqueName: \"kubernetes.io/projected/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-kube-api-access-7h65r\") pod \"community-operators-ppcmm\" (UID: \"3d9a3ae3-8d41-43e0-a1ed-a18d87778805\") " pod="openshift-marketplace/community-operators-ppcmm" Dec 05 07:53:16 crc kubenswrapper[4997]: I1205 07:53:16.849321 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-catalog-content\") pod \"community-operators-ppcmm\" (UID: \"3d9a3ae3-8d41-43e0-a1ed-a18d87778805\") " pod="openshift-marketplace/community-operators-ppcmm" Dec 05 07:53:16 crc kubenswrapper[4997]: I1205 07:53:16.849659 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-utilities\") pod \"community-operators-ppcmm\" (UID: \"3d9a3ae3-8d41-43e0-a1ed-a18d87778805\") " pod="openshift-marketplace/community-operators-ppcmm" Dec 05 07:53:16 crc kubenswrapper[4997]: I1205 07:53:16.863471 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ppcmm"] Dec 05 07:53:16 crc kubenswrapper[4997]: I1205 07:53:16.951117 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h65r\" (UniqueName: \"kubernetes.io/projected/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-kube-api-access-7h65r\") pod \"community-operators-ppcmm\" (UID: \"3d9a3ae3-8d41-43e0-a1ed-a18d87778805\") " pod="openshift-marketplace/community-operators-ppcmm" Dec 05 07:53:16 crc kubenswrapper[4997]: I1205 07:53:16.951175 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-catalog-content\") pod \"community-operators-ppcmm\" (UID: \"3d9a3ae3-8d41-43e0-a1ed-a18d87778805\") " pod="openshift-marketplace/community-operators-ppcmm" Dec 05 07:53:16 crc kubenswrapper[4997]: I1205 07:53:16.951243 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-utilities\") pod \"community-operators-ppcmm\" (UID: \"3d9a3ae3-8d41-43e0-a1ed-a18d87778805\") " pod="openshift-marketplace/community-operators-ppcmm" Dec 05 07:53:16 crc kubenswrapper[4997]: I1205 07:53:16.951936 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-utilities\") pod \"community-operators-ppcmm\" (UID: \"3d9a3ae3-8d41-43e0-a1ed-a18d87778805\") " pod="openshift-marketplace/community-operators-ppcmm" Dec 05 07:53:16 crc kubenswrapper[4997]: I1205 07:53:16.952010 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-catalog-content\") pod \"community-operators-ppcmm\" (UID: \"3d9a3ae3-8d41-43e0-a1ed-a18d87778805\") " pod="openshift-marketplace/community-operators-ppcmm" Dec 05 07:53:16 crc kubenswrapper[4997]: I1205 07:53:16.973907 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h65r\" (UniqueName: \"kubernetes.io/projected/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-kube-api-access-7h65r\") pod \"community-operators-ppcmm\" (UID: \"3d9a3ae3-8d41-43e0-a1ed-a18d87778805\") " pod="openshift-marketplace/community-operators-ppcmm" Dec 05 07:53:17 crc kubenswrapper[4997]: I1205 07:53:17.165593 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppcmm" Dec 05 07:53:17 crc kubenswrapper[4997]: I1205 07:53:17.473085 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ppcmm"] Dec 05 07:53:18 crc kubenswrapper[4997]: I1205 07:53:18.465157 4997 generic.go:334] "Generic (PLEG): container finished" podID="3d9a3ae3-8d41-43e0-a1ed-a18d87778805" containerID="a1826a40abd83ef816fefe0e068f5a31c8df69e1a8c54a6305946f47aca8557b" exitCode=0 Dec 05 07:53:18 crc kubenswrapper[4997]: I1205 07:53:18.465425 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppcmm" event={"ID":"3d9a3ae3-8d41-43e0-a1ed-a18d87778805","Type":"ContainerDied","Data":"a1826a40abd83ef816fefe0e068f5a31c8df69e1a8c54a6305946f47aca8557b"} Dec 05 07:53:18 crc kubenswrapper[4997]: I1205 07:53:18.467594 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppcmm" event={"ID":"3d9a3ae3-8d41-43e0-a1ed-a18d87778805","Type":"ContainerStarted","Data":"5f74f3546569bcba531635d9f723e3ebba85b1d0641d8fcc3f98ece4fb2b01d4"} Dec 05 07:53:18 crc kubenswrapper[4997]: I1205 07:53:18.467765 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 07:53:19 crc kubenswrapper[4997]: I1205 07:53:19.480328 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppcmm" event={"ID":"3d9a3ae3-8d41-43e0-a1ed-a18d87778805","Type":"ContainerStarted","Data":"32862f1a88522eaf61030cdf2f218668419027e6fce097d8f626e476033a23d3"} Dec 05 07:53:20 crc kubenswrapper[4997]: I1205 07:53:20.495170 4997 generic.go:334] "Generic (PLEG): container finished" podID="3d9a3ae3-8d41-43e0-a1ed-a18d87778805" containerID="32862f1a88522eaf61030cdf2f218668419027e6fce097d8f626e476033a23d3" exitCode=0 Dec 05 07:53:20 crc kubenswrapper[4997]: I1205 07:53:20.495768 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppcmm" event={"ID":"3d9a3ae3-8d41-43e0-a1ed-a18d87778805","Type":"ContainerDied","Data":"32862f1a88522eaf61030cdf2f218668419027e6fce097d8f626e476033a23d3"} Dec 05 07:53:21 crc kubenswrapper[4997]: I1205 07:53:21.510494 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppcmm" event={"ID":"3d9a3ae3-8d41-43e0-a1ed-a18d87778805","Type":"ContainerStarted","Data":"ba6d0b1bf67171ea0dc607ffe3ed46780d6ec71f461706927c5527311a348024"} Dec 05 07:53:21 crc kubenswrapper[4997]: I1205 07:53:21.543412 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ppcmm" podStartSLOduration=3.130732184 podStartE2EDuration="5.543353412s" podCreationTimestamp="2025-12-05 07:53:16 +0000 UTC" firstStartedPulling="2025-12-05 07:53:18.467351384 +0000 UTC m=+3498.996258655" lastFinishedPulling="2025-12-05 07:53:20.879972602 +0000 UTC m=+3501.408879883" observedRunningTime="2025-12-05 07:53:21.534900554 +0000 UTC m=+3502.063807875" watchObservedRunningTime="2025-12-05 07:53:21.543353412 +0000 UTC m=+3502.072260683" Dec 05 07:53:27 crc kubenswrapper[4997]: I1205 07:53:27.166374 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ppcmm" Dec 05 07:53:27 crc kubenswrapper[4997]: I1205 07:53:27.166822 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ppcmm" Dec 05 07:53:27 crc kubenswrapper[4997]: I1205 07:53:27.224145 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ppcmm" Dec 05 07:53:27 crc kubenswrapper[4997]: I1205 07:53:27.602056 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ppcmm" Dec 05 07:53:27 crc kubenswrapper[4997]: I1205 07:53:27.661943 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ppcmm"] Dec 05 07:53:29 crc kubenswrapper[4997]: I1205 07:53:29.584495 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ppcmm" podUID="3d9a3ae3-8d41-43e0-a1ed-a18d87778805" containerName="registry-server" containerID="cri-o://ba6d0b1bf67171ea0dc607ffe3ed46780d6ec71f461706927c5527311a348024" gracePeriod=2 Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.566098 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppcmm" Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.603032 4997 generic.go:334] "Generic (PLEG): container finished" podID="3d9a3ae3-8d41-43e0-a1ed-a18d87778805" containerID="ba6d0b1bf67171ea0dc607ffe3ed46780d6ec71f461706927c5527311a348024" exitCode=0 Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.603218 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppcmm" Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.603242 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppcmm" event={"ID":"3d9a3ae3-8d41-43e0-a1ed-a18d87778805","Type":"ContainerDied","Data":"ba6d0b1bf67171ea0dc607ffe3ed46780d6ec71f461706927c5527311a348024"} Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.605127 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppcmm" event={"ID":"3d9a3ae3-8d41-43e0-a1ed-a18d87778805","Type":"ContainerDied","Data":"5f74f3546569bcba531635d9f723e3ebba85b1d0641d8fcc3f98ece4fb2b01d4"} Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.605152 4997 scope.go:117] "RemoveContainer" containerID="ba6d0b1bf67171ea0dc607ffe3ed46780d6ec71f461706927c5527311a348024" Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.626976 4997 scope.go:117] "RemoveContainer" containerID="32862f1a88522eaf61030cdf2f218668419027e6fce097d8f626e476033a23d3" Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.649191 4997 scope.go:117] "RemoveContainer" containerID="a1826a40abd83ef816fefe0e068f5a31c8df69e1a8c54a6305946f47aca8557b" Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.666535 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-catalog-content\") pod \"3d9a3ae3-8d41-43e0-a1ed-a18d87778805\" (UID: \"3d9a3ae3-8d41-43e0-a1ed-a18d87778805\") " Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.666614 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h65r\" (UniqueName: \"kubernetes.io/projected/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-kube-api-access-7h65r\") pod \"3d9a3ae3-8d41-43e0-a1ed-a18d87778805\" (UID: \"3d9a3ae3-8d41-43e0-a1ed-a18d87778805\") " Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.666666 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-utilities\") pod \"3d9a3ae3-8d41-43e0-a1ed-a18d87778805\" (UID: \"3d9a3ae3-8d41-43e0-a1ed-a18d87778805\") " Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.667362 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-utilities" (OuterVolumeSpecName: "utilities") pod "3d9a3ae3-8d41-43e0-a1ed-a18d87778805" (UID: "3d9a3ae3-8d41-43e0-a1ed-a18d87778805"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.671602 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-kube-api-access-7h65r" (OuterVolumeSpecName: "kube-api-access-7h65r") pod "3d9a3ae3-8d41-43e0-a1ed-a18d87778805" (UID: "3d9a3ae3-8d41-43e0-a1ed-a18d87778805"). InnerVolumeSpecName "kube-api-access-7h65r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.673285 4997 scope.go:117] "RemoveContainer" containerID="ba6d0b1bf67171ea0dc607ffe3ed46780d6ec71f461706927c5527311a348024" Dec 05 07:53:30 crc kubenswrapper[4997]: E1205 07:53:30.673924 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba6d0b1bf67171ea0dc607ffe3ed46780d6ec71f461706927c5527311a348024\": container with ID starting with ba6d0b1bf67171ea0dc607ffe3ed46780d6ec71f461706927c5527311a348024 not found: ID does not exist" containerID="ba6d0b1bf67171ea0dc607ffe3ed46780d6ec71f461706927c5527311a348024" Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.673978 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba6d0b1bf67171ea0dc607ffe3ed46780d6ec71f461706927c5527311a348024"} err="failed to get container status \"ba6d0b1bf67171ea0dc607ffe3ed46780d6ec71f461706927c5527311a348024\": rpc error: code = NotFound desc = could not find container \"ba6d0b1bf67171ea0dc607ffe3ed46780d6ec71f461706927c5527311a348024\": container with ID starting with ba6d0b1bf67171ea0dc607ffe3ed46780d6ec71f461706927c5527311a348024 not found: ID does not exist" Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.674013 4997 scope.go:117] "RemoveContainer" containerID="32862f1a88522eaf61030cdf2f218668419027e6fce097d8f626e476033a23d3" Dec 05 07:53:30 crc kubenswrapper[4997]: E1205 07:53:30.674463 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32862f1a88522eaf61030cdf2f218668419027e6fce097d8f626e476033a23d3\": container with ID starting with 32862f1a88522eaf61030cdf2f218668419027e6fce097d8f626e476033a23d3 not found: ID does not exist" containerID="32862f1a88522eaf61030cdf2f218668419027e6fce097d8f626e476033a23d3" Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.674511 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32862f1a88522eaf61030cdf2f218668419027e6fce097d8f626e476033a23d3"} err="failed to get container status \"32862f1a88522eaf61030cdf2f218668419027e6fce097d8f626e476033a23d3\": rpc error: code = NotFound desc = could not find container \"32862f1a88522eaf61030cdf2f218668419027e6fce097d8f626e476033a23d3\": container with ID starting with 32862f1a88522eaf61030cdf2f218668419027e6fce097d8f626e476033a23d3 not found: ID does not exist" Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.674540 4997 scope.go:117] "RemoveContainer" containerID="a1826a40abd83ef816fefe0e068f5a31c8df69e1a8c54a6305946f47aca8557b" Dec 05 07:53:30 crc kubenswrapper[4997]: E1205 07:53:30.674899 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1826a40abd83ef816fefe0e068f5a31c8df69e1a8c54a6305946f47aca8557b\": container with ID starting with a1826a40abd83ef816fefe0e068f5a31c8df69e1a8c54a6305946f47aca8557b not found: ID does not exist" containerID="a1826a40abd83ef816fefe0e068f5a31c8df69e1a8c54a6305946f47aca8557b" Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.674943 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1826a40abd83ef816fefe0e068f5a31c8df69e1a8c54a6305946f47aca8557b"} err="failed to get container status \"a1826a40abd83ef816fefe0e068f5a31c8df69e1a8c54a6305946f47aca8557b\": rpc error: code = NotFound desc = could not find container \"a1826a40abd83ef816fefe0e068f5a31c8df69e1a8c54a6305946f47aca8557b\": container with ID starting with a1826a40abd83ef816fefe0e068f5a31c8df69e1a8c54a6305946f47aca8557b not found: ID does not exist" Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.727074 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d9a3ae3-8d41-43e0-a1ed-a18d87778805" (UID: "3d9a3ae3-8d41-43e0-a1ed-a18d87778805"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.768353 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.768396 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h65r\" (UniqueName: \"kubernetes.io/projected/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-kube-api-access-7h65r\") on node \"crc\" DevicePath \"\"" Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.768413 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9a3ae3-8d41-43e0-a1ed-a18d87778805-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.957695 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ppcmm"] Dec 05 07:53:30 crc kubenswrapper[4997]: I1205 07:53:30.969428 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ppcmm"] Dec 05 07:53:31 crc kubenswrapper[4997]: I1205 07:53:31.760739 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d9a3ae3-8d41-43e0-a1ed-a18d87778805" path="/var/lib/kubelet/pods/3d9a3ae3-8d41-43e0-a1ed-a18d87778805/volumes" Dec 05 07:54:03 crc kubenswrapper[4997]: I1205 07:54:03.637868 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hzzcx"] Dec 05 07:54:03 crc kubenswrapper[4997]: E1205 07:54:03.639316 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9a3ae3-8d41-43e0-a1ed-a18d87778805" containerName="extract-content" Dec 05 07:54:03 crc kubenswrapper[4997]: I1205 07:54:03.639341 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9a3ae3-8d41-43e0-a1ed-a18d87778805" containerName="extract-content" Dec 05 07:54:03 crc kubenswrapper[4997]: E1205 07:54:03.639372 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9a3ae3-8d41-43e0-a1ed-a18d87778805" containerName="registry-server" Dec 05 07:54:03 crc kubenswrapper[4997]: I1205 07:54:03.639387 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9a3ae3-8d41-43e0-a1ed-a18d87778805" containerName="registry-server" Dec 05 07:54:03 crc kubenswrapper[4997]: E1205 07:54:03.639422 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9a3ae3-8d41-43e0-a1ed-a18d87778805" containerName="extract-utilities" Dec 05 07:54:03 crc kubenswrapper[4997]: I1205 07:54:03.639436 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9a3ae3-8d41-43e0-a1ed-a18d87778805" containerName="extract-utilities" Dec 05 07:54:03 crc kubenswrapper[4997]: I1205 07:54:03.639968 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d9a3ae3-8d41-43e0-a1ed-a18d87778805" containerName="registry-server" Dec 05 07:54:03 crc kubenswrapper[4997]: I1205 07:54:03.645256 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hzzcx" Dec 05 07:54:03 crc kubenswrapper[4997]: I1205 07:54:03.663905 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hzzcx"] Dec 05 07:54:03 crc kubenswrapper[4997]: I1205 07:54:03.793989 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb55f5bd-56ab-4c9e-a285-781f75db09a5-catalog-content\") pod \"redhat-marketplace-hzzcx\" (UID: \"bb55f5bd-56ab-4c9e-a285-781f75db09a5\") " pod="openshift-marketplace/redhat-marketplace-hzzcx" Dec 05 07:54:03 crc kubenswrapper[4997]: I1205 07:54:03.794034 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhnbn\" (UniqueName: \"kubernetes.io/projected/bb55f5bd-56ab-4c9e-a285-781f75db09a5-kube-api-access-vhnbn\") pod \"redhat-marketplace-hzzcx\" (UID: \"bb55f5bd-56ab-4c9e-a285-781f75db09a5\") " pod="openshift-marketplace/redhat-marketplace-hzzcx" Dec 05 07:54:03 crc kubenswrapper[4997]: I1205 07:54:03.794061 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb55f5bd-56ab-4c9e-a285-781f75db09a5-utilities\") pod \"redhat-marketplace-hzzcx\" (UID: \"bb55f5bd-56ab-4c9e-a285-781f75db09a5\") " pod="openshift-marketplace/redhat-marketplace-hzzcx" Dec 05 07:54:03 crc kubenswrapper[4997]: I1205 07:54:03.894795 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhnbn\" (UniqueName: \"kubernetes.io/projected/bb55f5bd-56ab-4c9e-a285-781f75db09a5-kube-api-access-vhnbn\") pod \"redhat-marketplace-hzzcx\" (UID: \"bb55f5bd-56ab-4c9e-a285-781f75db09a5\") " pod="openshift-marketplace/redhat-marketplace-hzzcx" Dec 05 07:54:03 crc kubenswrapper[4997]: I1205 07:54:03.894852 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb55f5bd-56ab-4c9e-a285-781f75db09a5-utilities\") pod \"redhat-marketplace-hzzcx\" (UID: \"bb55f5bd-56ab-4c9e-a285-781f75db09a5\") " pod="openshift-marketplace/redhat-marketplace-hzzcx" Dec 05 07:54:03 crc kubenswrapper[4997]: I1205 07:54:03.894948 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb55f5bd-56ab-4c9e-a285-781f75db09a5-catalog-content\") pod \"redhat-marketplace-hzzcx\" (UID: \"bb55f5bd-56ab-4c9e-a285-781f75db09a5\") " pod="openshift-marketplace/redhat-marketplace-hzzcx" Dec 05 07:54:03 crc kubenswrapper[4997]: I1205 07:54:03.896022 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb55f5bd-56ab-4c9e-a285-781f75db09a5-utilities\") pod \"redhat-marketplace-hzzcx\" (UID: \"bb55f5bd-56ab-4c9e-a285-781f75db09a5\") " pod="openshift-marketplace/redhat-marketplace-hzzcx" Dec 05 07:54:03 crc kubenswrapper[4997]: I1205 07:54:03.896119 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb55f5bd-56ab-4c9e-a285-781f75db09a5-catalog-content\") pod \"redhat-marketplace-hzzcx\" (UID: \"bb55f5bd-56ab-4c9e-a285-781f75db09a5\") " pod="openshift-marketplace/redhat-marketplace-hzzcx" Dec 05 07:54:03 crc kubenswrapper[4997]: I1205 07:54:03.913535 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhnbn\" (UniqueName: \"kubernetes.io/projected/bb55f5bd-56ab-4c9e-a285-781f75db09a5-kube-api-access-vhnbn\") pod \"redhat-marketplace-hzzcx\" (UID: \"bb55f5bd-56ab-4c9e-a285-781f75db09a5\") " pod="openshift-marketplace/redhat-marketplace-hzzcx" Dec 05 07:54:03 crc kubenswrapper[4997]: I1205 07:54:03.974583 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hzzcx" Dec 05 07:54:04 crc kubenswrapper[4997]: I1205 07:54:04.200232 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hzzcx"] Dec 05 07:54:04 crc kubenswrapper[4997]: I1205 07:54:04.919328 4997 generic.go:334] "Generic (PLEG): container finished" podID="bb55f5bd-56ab-4c9e-a285-781f75db09a5" containerID="4a99cb6cb0563d9503f1f078a30a870f0064613f5850739e06e67e2e3b056cb1" exitCode=0 Dec 05 07:54:04 crc kubenswrapper[4997]: I1205 07:54:04.919393 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hzzcx" event={"ID":"bb55f5bd-56ab-4c9e-a285-781f75db09a5","Type":"ContainerDied","Data":"4a99cb6cb0563d9503f1f078a30a870f0064613f5850739e06e67e2e3b056cb1"} Dec 05 07:54:04 crc kubenswrapper[4997]: I1205 07:54:04.919689 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hzzcx" event={"ID":"bb55f5bd-56ab-4c9e-a285-781f75db09a5","Type":"ContainerStarted","Data":"0b887b2c8c089b627bbec98e22b9bdfd3966f75b4c60c1e6ca2adc88eae9d48f"} Dec 05 07:54:05 crc kubenswrapper[4997]: I1205 07:54:05.928824 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hzzcx" event={"ID":"bb55f5bd-56ab-4c9e-a285-781f75db09a5","Type":"ContainerStarted","Data":"af16636b685f05dc7e3261b1cc8b5ec3c1d92944ab3812475d6c35f823e73fea"} Dec 05 07:54:06 crc kubenswrapper[4997]: I1205 07:54:06.941431 4997 generic.go:334] "Generic (PLEG): container finished" podID="bb55f5bd-56ab-4c9e-a285-781f75db09a5" containerID="af16636b685f05dc7e3261b1cc8b5ec3c1d92944ab3812475d6c35f823e73fea" exitCode=0 Dec 05 07:54:06 crc kubenswrapper[4997]: I1205 07:54:06.941493 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hzzcx" event={"ID":"bb55f5bd-56ab-4c9e-a285-781f75db09a5","Type":"ContainerDied","Data":"af16636b685f05dc7e3261b1cc8b5ec3c1d92944ab3812475d6c35f823e73fea"} Dec 05 07:54:07 crc kubenswrapper[4997]: I1205 07:54:07.953128 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hzzcx" event={"ID":"bb55f5bd-56ab-4c9e-a285-781f75db09a5","Type":"ContainerStarted","Data":"30177e76297dd0d04a6287de92fc1b36501bb6387ebd593cf72b2771245a1bbb"} Dec 05 07:54:07 crc kubenswrapper[4997]: I1205 07:54:07.969421 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hzzcx" podStartSLOduration=2.497114176 podStartE2EDuration="4.969404491s" podCreationTimestamp="2025-12-05 07:54:03 +0000 UTC" firstStartedPulling="2025-12-05 07:54:04.920853493 +0000 UTC m=+3545.449760754" lastFinishedPulling="2025-12-05 07:54:07.393143778 +0000 UTC m=+3547.922051069" observedRunningTime="2025-12-05 07:54:07.968220829 +0000 UTC m=+3548.497128090" watchObservedRunningTime="2025-12-05 07:54:07.969404491 +0000 UTC m=+3548.498311752" Dec 05 07:54:13 crc kubenswrapper[4997]: I1205 07:54:13.975281 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hzzcx" Dec 05 07:54:13 crc kubenswrapper[4997]: I1205 07:54:13.975786 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hzzcx" Dec 05 07:54:14 crc kubenswrapper[4997]: I1205 07:54:14.053170 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hzzcx" Dec 05 07:54:14 crc kubenswrapper[4997]: I1205 07:54:14.101592 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hzzcx" Dec 05 07:54:14 crc kubenswrapper[4997]: I1205 07:54:14.293749 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hzzcx"] Dec 05 07:54:16 crc kubenswrapper[4997]: I1205 07:54:16.032768 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hzzcx" podUID="bb55f5bd-56ab-4c9e-a285-781f75db09a5" containerName="registry-server" containerID="cri-o://30177e76297dd0d04a6287de92fc1b36501bb6387ebd593cf72b2771245a1bbb" gracePeriod=2 Dec 05 07:54:16 crc kubenswrapper[4997]: E1205 07:54:16.360453 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb55f5bd_56ab_4c9e_a285_781f75db09a5.slice/crio-conmon-30177e76297dd0d04a6287de92fc1b36501bb6387ebd593cf72b2771245a1bbb.scope\": RecentStats: unable to find data in memory cache]" Dec 05 07:54:16 crc kubenswrapper[4997]: I1205 07:54:16.961001 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hzzcx" Dec 05 07:54:16 crc kubenswrapper[4997]: I1205 07:54:16.989796 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb55f5bd-56ab-4c9e-a285-781f75db09a5-utilities\") pod \"bb55f5bd-56ab-4c9e-a285-781f75db09a5\" (UID: \"bb55f5bd-56ab-4c9e-a285-781f75db09a5\") " Dec 05 07:54:16 crc kubenswrapper[4997]: I1205 07:54:16.989909 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb55f5bd-56ab-4c9e-a285-781f75db09a5-catalog-content\") pod \"bb55f5bd-56ab-4c9e-a285-781f75db09a5\" (UID: \"bb55f5bd-56ab-4c9e-a285-781f75db09a5\") " Dec 05 07:54:16 crc kubenswrapper[4997]: I1205 07:54:16.989938 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhnbn\" (UniqueName: \"kubernetes.io/projected/bb55f5bd-56ab-4c9e-a285-781f75db09a5-kube-api-access-vhnbn\") pod \"bb55f5bd-56ab-4c9e-a285-781f75db09a5\" (UID: \"bb55f5bd-56ab-4c9e-a285-781f75db09a5\") " Dec 05 07:54:16 crc kubenswrapper[4997]: I1205 07:54:16.993243 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb55f5bd-56ab-4c9e-a285-781f75db09a5-utilities" (OuterVolumeSpecName: "utilities") pod "bb55f5bd-56ab-4c9e-a285-781f75db09a5" (UID: "bb55f5bd-56ab-4c9e-a285-781f75db09a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:54:16 crc kubenswrapper[4997]: I1205 07:54:16.998824 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb55f5bd-56ab-4c9e-a285-781f75db09a5-kube-api-access-vhnbn" (OuterVolumeSpecName: "kube-api-access-vhnbn") pod "bb55f5bd-56ab-4c9e-a285-781f75db09a5" (UID: "bb55f5bd-56ab-4c9e-a285-781f75db09a5"). InnerVolumeSpecName "kube-api-access-vhnbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.009355 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb55f5bd-56ab-4c9e-a285-781f75db09a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb55f5bd-56ab-4c9e-a285-781f75db09a5" (UID: "bb55f5bd-56ab-4c9e-a285-781f75db09a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.041658 4997 generic.go:334] "Generic (PLEG): container finished" podID="bb55f5bd-56ab-4c9e-a285-781f75db09a5" containerID="30177e76297dd0d04a6287de92fc1b36501bb6387ebd593cf72b2771245a1bbb" exitCode=0 Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.041696 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hzzcx" event={"ID":"bb55f5bd-56ab-4c9e-a285-781f75db09a5","Type":"ContainerDied","Data":"30177e76297dd0d04a6287de92fc1b36501bb6387ebd593cf72b2771245a1bbb"} Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.041715 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hzzcx" Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.041733 4997 scope.go:117] "RemoveContainer" containerID="30177e76297dd0d04a6287de92fc1b36501bb6387ebd593cf72b2771245a1bbb" Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.041722 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hzzcx" event={"ID":"bb55f5bd-56ab-4c9e-a285-781f75db09a5","Type":"ContainerDied","Data":"0b887b2c8c089b627bbec98e22b9bdfd3966f75b4c60c1e6ca2adc88eae9d48f"} Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.070904 4997 scope.go:117] "RemoveContainer" containerID="af16636b685f05dc7e3261b1cc8b5ec3c1d92944ab3812475d6c35f823e73fea" Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.080197 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hzzcx"] Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.090561 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hzzcx"] Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.092039 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb55f5bd-56ab-4c9e-a285-781f75db09a5-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.092075 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb55f5bd-56ab-4c9e-a285-781f75db09a5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.092091 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhnbn\" (UniqueName: \"kubernetes.io/projected/bb55f5bd-56ab-4c9e-a285-781f75db09a5-kube-api-access-vhnbn\") on node \"crc\" DevicePath \"\"" Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.105927 4997 scope.go:117] "RemoveContainer" containerID="4a99cb6cb0563d9503f1f078a30a870f0064613f5850739e06e67e2e3b056cb1" Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.128218 4997 scope.go:117] "RemoveContainer" containerID="30177e76297dd0d04a6287de92fc1b36501bb6387ebd593cf72b2771245a1bbb" Dec 05 07:54:17 crc kubenswrapper[4997]: E1205 07:54:17.128747 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30177e76297dd0d04a6287de92fc1b36501bb6387ebd593cf72b2771245a1bbb\": container with ID starting with 30177e76297dd0d04a6287de92fc1b36501bb6387ebd593cf72b2771245a1bbb not found: ID does not exist" containerID="30177e76297dd0d04a6287de92fc1b36501bb6387ebd593cf72b2771245a1bbb" Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.128811 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30177e76297dd0d04a6287de92fc1b36501bb6387ebd593cf72b2771245a1bbb"} err="failed to get container status \"30177e76297dd0d04a6287de92fc1b36501bb6387ebd593cf72b2771245a1bbb\": rpc error: code = NotFound desc = could not find container \"30177e76297dd0d04a6287de92fc1b36501bb6387ebd593cf72b2771245a1bbb\": container with ID starting with 30177e76297dd0d04a6287de92fc1b36501bb6387ebd593cf72b2771245a1bbb not found: ID does not exist" Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.128851 4997 scope.go:117] "RemoveContainer" containerID="af16636b685f05dc7e3261b1cc8b5ec3c1d92944ab3812475d6c35f823e73fea" Dec 05 07:54:17 crc kubenswrapper[4997]: E1205 07:54:17.129231 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af16636b685f05dc7e3261b1cc8b5ec3c1d92944ab3812475d6c35f823e73fea\": container with ID starting with af16636b685f05dc7e3261b1cc8b5ec3c1d92944ab3812475d6c35f823e73fea not found: ID does not exist" containerID="af16636b685f05dc7e3261b1cc8b5ec3c1d92944ab3812475d6c35f823e73fea" Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.129287 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af16636b685f05dc7e3261b1cc8b5ec3c1d92944ab3812475d6c35f823e73fea"} err="failed to get container status \"af16636b685f05dc7e3261b1cc8b5ec3c1d92944ab3812475d6c35f823e73fea\": rpc error: code = NotFound desc = could not find container \"af16636b685f05dc7e3261b1cc8b5ec3c1d92944ab3812475d6c35f823e73fea\": container with ID starting with af16636b685f05dc7e3261b1cc8b5ec3c1d92944ab3812475d6c35f823e73fea not found: ID does not exist" Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.129321 4997 scope.go:117] "RemoveContainer" containerID="4a99cb6cb0563d9503f1f078a30a870f0064613f5850739e06e67e2e3b056cb1" Dec 05 07:54:17 crc kubenswrapper[4997]: E1205 07:54:17.129714 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a99cb6cb0563d9503f1f078a30a870f0064613f5850739e06e67e2e3b056cb1\": container with ID starting with 4a99cb6cb0563d9503f1f078a30a870f0064613f5850739e06e67e2e3b056cb1 not found: ID does not exist" containerID="4a99cb6cb0563d9503f1f078a30a870f0064613f5850739e06e67e2e3b056cb1" Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.129746 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a99cb6cb0563d9503f1f078a30a870f0064613f5850739e06e67e2e3b056cb1"} err="failed to get container status \"4a99cb6cb0563d9503f1f078a30a870f0064613f5850739e06e67e2e3b056cb1\": rpc error: code = NotFound desc = could not find container \"4a99cb6cb0563d9503f1f078a30a870f0064613f5850739e06e67e2e3b056cb1\": container with ID starting with 4a99cb6cb0563d9503f1f078a30a870f0064613f5850739e06e67e2e3b056cb1 not found: ID does not exist" Dec 05 07:54:17 crc kubenswrapper[4997]: I1205 07:54:17.763556 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb55f5bd-56ab-4c9e-a285-781f75db09a5" path="/var/lib/kubelet/pods/bb55f5bd-56ab-4c9e-a285-781f75db09a5/volumes" Dec 05 07:54:19 crc kubenswrapper[4997]: I1205 07:54:19.770352 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:54:19 crc kubenswrapper[4997]: I1205 07:54:19.771118 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:54:42 crc kubenswrapper[4997]: I1205 07:54:42.593805 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bj74g"] Dec 05 07:54:42 crc kubenswrapper[4997]: E1205 07:54:42.594475 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb55f5bd-56ab-4c9e-a285-781f75db09a5" containerName="registry-server" Dec 05 07:54:42 crc kubenswrapper[4997]: I1205 07:54:42.594485 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb55f5bd-56ab-4c9e-a285-781f75db09a5" containerName="registry-server" Dec 05 07:54:42 crc kubenswrapper[4997]: E1205 07:54:42.594514 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb55f5bd-56ab-4c9e-a285-781f75db09a5" containerName="extract-utilities" Dec 05 07:54:42 crc kubenswrapper[4997]: I1205 07:54:42.594523 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb55f5bd-56ab-4c9e-a285-781f75db09a5" containerName="extract-utilities" Dec 05 07:54:42 crc kubenswrapper[4997]: E1205 07:54:42.594530 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb55f5bd-56ab-4c9e-a285-781f75db09a5" containerName="extract-content" Dec 05 07:54:42 crc kubenswrapper[4997]: I1205 07:54:42.594536 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb55f5bd-56ab-4c9e-a285-781f75db09a5" containerName="extract-content" Dec 05 07:54:42 crc kubenswrapper[4997]: I1205 07:54:42.594678 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb55f5bd-56ab-4c9e-a285-781f75db09a5" containerName="registry-server" Dec 05 07:54:42 crc kubenswrapper[4997]: I1205 07:54:42.595654 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bj74g" Dec 05 07:54:42 crc kubenswrapper[4997]: I1205 07:54:42.612475 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bj74g"] Dec 05 07:54:42 crc kubenswrapper[4997]: I1205 07:54:42.707823 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e022bd86-466c-4393-8605-041774e8a636-catalog-content\") pod \"certified-operators-bj74g\" (UID: \"e022bd86-466c-4393-8605-041774e8a636\") " pod="openshift-marketplace/certified-operators-bj74g" Dec 05 07:54:42 crc kubenswrapper[4997]: I1205 07:54:42.707902 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rgrx\" (UniqueName: \"kubernetes.io/projected/e022bd86-466c-4393-8605-041774e8a636-kube-api-access-8rgrx\") pod \"certified-operators-bj74g\" (UID: \"e022bd86-466c-4393-8605-041774e8a636\") " pod="openshift-marketplace/certified-operators-bj74g" Dec 05 07:54:42 crc kubenswrapper[4997]: I1205 07:54:42.707941 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e022bd86-466c-4393-8605-041774e8a636-utilities\") pod \"certified-operators-bj74g\" (UID: \"e022bd86-466c-4393-8605-041774e8a636\") " pod="openshift-marketplace/certified-operators-bj74g" Dec 05 07:54:42 crc kubenswrapper[4997]: I1205 07:54:42.809448 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rgrx\" (UniqueName: \"kubernetes.io/projected/e022bd86-466c-4393-8605-041774e8a636-kube-api-access-8rgrx\") pod \"certified-operators-bj74g\" (UID: \"e022bd86-466c-4393-8605-041774e8a636\") " pod="openshift-marketplace/certified-operators-bj74g" Dec 05 07:54:42 crc kubenswrapper[4997]: I1205 07:54:42.809592 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e022bd86-466c-4393-8605-041774e8a636-utilities\") pod \"certified-operators-bj74g\" (UID: \"e022bd86-466c-4393-8605-041774e8a636\") " pod="openshift-marketplace/certified-operators-bj74g" Dec 05 07:54:42 crc kubenswrapper[4997]: I1205 07:54:42.809720 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e022bd86-466c-4393-8605-041774e8a636-catalog-content\") pod \"certified-operators-bj74g\" (UID: \"e022bd86-466c-4393-8605-041774e8a636\") " pod="openshift-marketplace/certified-operators-bj74g" Dec 05 07:54:42 crc kubenswrapper[4997]: I1205 07:54:42.810226 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e022bd86-466c-4393-8605-041774e8a636-utilities\") pod \"certified-operators-bj74g\" (UID: \"e022bd86-466c-4393-8605-041774e8a636\") " pod="openshift-marketplace/certified-operators-bj74g" Dec 05 07:54:42 crc kubenswrapper[4997]: I1205 07:54:42.810310 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e022bd86-466c-4393-8605-041774e8a636-catalog-content\") pod \"certified-operators-bj74g\" (UID: \"e022bd86-466c-4393-8605-041774e8a636\") " pod="openshift-marketplace/certified-operators-bj74g" Dec 05 07:54:42 crc kubenswrapper[4997]: I1205 07:54:42.829162 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rgrx\" (UniqueName: \"kubernetes.io/projected/e022bd86-466c-4393-8605-041774e8a636-kube-api-access-8rgrx\") pod \"certified-operators-bj74g\" (UID: \"e022bd86-466c-4393-8605-041774e8a636\") " pod="openshift-marketplace/certified-operators-bj74g" Dec 05 07:54:42 crc kubenswrapper[4997]: I1205 07:54:42.922916 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bj74g" Dec 05 07:54:43 crc kubenswrapper[4997]: I1205 07:54:43.276430 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bj74g"] Dec 05 07:54:44 crc kubenswrapper[4997]: I1205 07:54:44.283684 4997 generic.go:334] "Generic (PLEG): container finished" podID="e022bd86-466c-4393-8605-041774e8a636" containerID="ff5be2573614d1819bfb6085c7f5bfc14702d0ed5a9bffe97c917bd7dc70de44" exitCode=0 Dec 05 07:54:44 crc kubenswrapper[4997]: I1205 07:54:44.283777 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bj74g" event={"ID":"e022bd86-466c-4393-8605-041774e8a636","Type":"ContainerDied","Data":"ff5be2573614d1819bfb6085c7f5bfc14702d0ed5a9bffe97c917bd7dc70de44"} Dec 05 07:54:44 crc kubenswrapper[4997]: I1205 07:54:44.284150 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bj74g" event={"ID":"e022bd86-466c-4393-8605-041774e8a636","Type":"ContainerStarted","Data":"bbbe412d4cce3f2e2d18a1853aa58b9b581b2b1fdd50c40fc372d40a62f5bbf7"} Dec 05 07:54:45 crc kubenswrapper[4997]: I1205 07:54:45.294163 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bj74g" event={"ID":"e022bd86-466c-4393-8605-041774e8a636","Type":"ContainerStarted","Data":"2b8437e2b21b2d67247ac10712864f9c82a7c786fcedeb71591696fda5d64d93"} Dec 05 07:54:46 crc kubenswrapper[4997]: I1205 07:54:46.312147 4997 generic.go:334] "Generic (PLEG): container finished" podID="e022bd86-466c-4393-8605-041774e8a636" containerID="2b8437e2b21b2d67247ac10712864f9c82a7c786fcedeb71591696fda5d64d93" exitCode=0 Dec 05 07:54:46 crc kubenswrapper[4997]: I1205 07:54:46.312245 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bj74g" event={"ID":"e022bd86-466c-4393-8605-041774e8a636","Type":"ContainerDied","Data":"2b8437e2b21b2d67247ac10712864f9c82a7c786fcedeb71591696fda5d64d93"} Dec 05 07:54:47 crc kubenswrapper[4997]: I1205 07:54:47.325147 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bj74g" event={"ID":"e022bd86-466c-4393-8605-041774e8a636","Type":"ContainerStarted","Data":"02ed0e1da46752db14022288ce1cac62ee169e9eee6eb07269d8dbad938af0c0"} Dec 05 07:54:47 crc kubenswrapper[4997]: I1205 07:54:47.354368 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bj74g" podStartSLOduration=2.919958436 podStartE2EDuration="5.354345907s" podCreationTimestamp="2025-12-05 07:54:42 +0000 UTC" firstStartedPulling="2025-12-05 07:54:44.28581182 +0000 UTC m=+3584.814719081" lastFinishedPulling="2025-12-05 07:54:46.720199251 +0000 UTC m=+3587.249106552" observedRunningTime="2025-12-05 07:54:47.346951858 +0000 UTC m=+3587.875859189" watchObservedRunningTime="2025-12-05 07:54:47.354345907 +0000 UTC m=+3587.883253178" Dec 05 07:54:49 crc kubenswrapper[4997]: I1205 07:54:49.770185 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:54:49 crc kubenswrapper[4997]: I1205 07:54:49.770666 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:54:52 crc kubenswrapper[4997]: I1205 07:54:52.923697 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bj74g" Dec 05 07:54:52 crc kubenswrapper[4997]: I1205 07:54:52.924095 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bj74g" Dec 05 07:54:53 crc kubenswrapper[4997]: I1205 07:54:53.000801 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bj74g" Dec 05 07:54:53 crc kubenswrapper[4997]: I1205 07:54:53.452265 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bj74g" Dec 05 07:54:53 crc kubenswrapper[4997]: I1205 07:54:53.514115 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bj74g"] Dec 05 07:54:55 crc kubenswrapper[4997]: I1205 07:54:55.405502 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bj74g" podUID="e022bd86-466c-4393-8605-041774e8a636" containerName="registry-server" containerID="cri-o://02ed0e1da46752db14022288ce1cac62ee169e9eee6eb07269d8dbad938af0c0" gracePeriod=2 Dec 05 07:54:55 crc kubenswrapper[4997]: I1205 07:54:55.934567 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bj74g" Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.059778 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rgrx\" (UniqueName: \"kubernetes.io/projected/e022bd86-466c-4393-8605-041774e8a636-kube-api-access-8rgrx\") pod \"e022bd86-466c-4393-8605-041774e8a636\" (UID: \"e022bd86-466c-4393-8605-041774e8a636\") " Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.059874 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e022bd86-466c-4393-8605-041774e8a636-utilities\") pod \"e022bd86-466c-4393-8605-041774e8a636\" (UID: \"e022bd86-466c-4393-8605-041774e8a636\") " Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.059976 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e022bd86-466c-4393-8605-041774e8a636-catalog-content\") pod \"e022bd86-466c-4393-8605-041774e8a636\" (UID: \"e022bd86-466c-4393-8605-041774e8a636\") " Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.061671 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e022bd86-466c-4393-8605-041774e8a636-utilities" (OuterVolumeSpecName: "utilities") pod "e022bd86-466c-4393-8605-041774e8a636" (UID: "e022bd86-466c-4393-8605-041774e8a636"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.067932 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e022bd86-466c-4393-8605-041774e8a636-kube-api-access-8rgrx" (OuterVolumeSpecName: "kube-api-access-8rgrx") pod "e022bd86-466c-4393-8605-041774e8a636" (UID: "e022bd86-466c-4393-8605-041774e8a636"). InnerVolumeSpecName "kube-api-access-8rgrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.106716 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e022bd86-466c-4393-8605-041774e8a636-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e022bd86-466c-4393-8605-041774e8a636" (UID: "e022bd86-466c-4393-8605-041774e8a636"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.162338 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e022bd86-466c-4393-8605-041774e8a636-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.162722 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e022bd86-466c-4393-8605-041774e8a636-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.162880 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rgrx\" (UniqueName: \"kubernetes.io/projected/e022bd86-466c-4393-8605-041774e8a636-kube-api-access-8rgrx\") on node \"crc\" DevicePath \"\"" Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.416373 4997 generic.go:334] "Generic (PLEG): container finished" podID="e022bd86-466c-4393-8605-041774e8a636" containerID="02ed0e1da46752db14022288ce1cac62ee169e9eee6eb07269d8dbad938af0c0" exitCode=0 Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.416446 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bj74g" Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.416441 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bj74g" event={"ID":"e022bd86-466c-4393-8605-041774e8a636","Type":"ContainerDied","Data":"02ed0e1da46752db14022288ce1cac62ee169e9eee6eb07269d8dbad938af0c0"} Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.416589 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bj74g" event={"ID":"e022bd86-466c-4393-8605-041774e8a636","Type":"ContainerDied","Data":"bbbe412d4cce3f2e2d18a1853aa58b9b581b2b1fdd50c40fc372d40a62f5bbf7"} Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.416648 4997 scope.go:117] "RemoveContainer" containerID="02ed0e1da46752db14022288ce1cac62ee169e9eee6eb07269d8dbad938af0c0" Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.454811 4997 scope.go:117] "RemoveContainer" containerID="2b8437e2b21b2d67247ac10712864f9c82a7c786fcedeb71591696fda5d64d93" Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.458046 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bj74g"] Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.468025 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bj74g"] Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.478166 4997 scope.go:117] "RemoveContainer" containerID="ff5be2573614d1819bfb6085c7f5bfc14702d0ed5a9bffe97c917bd7dc70de44" Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.502769 4997 scope.go:117] "RemoveContainer" containerID="02ed0e1da46752db14022288ce1cac62ee169e9eee6eb07269d8dbad938af0c0" Dec 05 07:54:56 crc kubenswrapper[4997]: E1205 07:54:56.503337 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02ed0e1da46752db14022288ce1cac62ee169e9eee6eb07269d8dbad938af0c0\": container with ID starting with 02ed0e1da46752db14022288ce1cac62ee169e9eee6eb07269d8dbad938af0c0 not found: ID does not exist" containerID="02ed0e1da46752db14022288ce1cac62ee169e9eee6eb07269d8dbad938af0c0" Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.503427 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02ed0e1da46752db14022288ce1cac62ee169e9eee6eb07269d8dbad938af0c0"} err="failed to get container status \"02ed0e1da46752db14022288ce1cac62ee169e9eee6eb07269d8dbad938af0c0\": rpc error: code = NotFound desc = could not find container \"02ed0e1da46752db14022288ce1cac62ee169e9eee6eb07269d8dbad938af0c0\": container with ID starting with 02ed0e1da46752db14022288ce1cac62ee169e9eee6eb07269d8dbad938af0c0 not found: ID does not exist" Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.503459 4997 scope.go:117] "RemoveContainer" containerID="2b8437e2b21b2d67247ac10712864f9c82a7c786fcedeb71591696fda5d64d93" Dec 05 07:54:56 crc kubenswrapper[4997]: E1205 07:54:56.503972 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b8437e2b21b2d67247ac10712864f9c82a7c786fcedeb71591696fda5d64d93\": container with ID starting with 2b8437e2b21b2d67247ac10712864f9c82a7c786fcedeb71591696fda5d64d93 not found: ID does not exist" containerID="2b8437e2b21b2d67247ac10712864f9c82a7c786fcedeb71591696fda5d64d93" Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.504022 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b8437e2b21b2d67247ac10712864f9c82a7c786fcedeb71591696fda5d64d93"} err="failed to get container status \"2b8437e2b21b2d67247ac10712864f9c82a7c786fcedeb71591696fda5d64d93\": rpc error: code = NotFound desc = could not find container \"2b8437e2b21b2d67247ac10712864f9c82a7c786fcedeb71591696fda5d64d93\": container with ID starting with 2b8437e2b21b2d67247ac10712864f9c82a7c786fcedeb71591696fda5d64d93 not found: ID does not exist" Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.504056 4997 scope.go:117] "RemoveContainer" containerID="ff5be2573614d1819bfb6085c7f5bfc14702d0ed5a9bffe97c917bd7dc70de44" Dec 05 07:54:56 crc kubenswrapper[4997]: E1205 07:54:56.504422 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff5be2573614d1819bfb6085c7f5bfc14702d0ed5a9bffe97c917bd7dc70de44\": container with ID starting with ff5be2573614d1819bfb6085c7f5bfc14702d0ed5a9bffe97c917bd7dc70de44 not found: ID does not exist" containerID="ff5be2573614d1819bfb6085c7f5bfc14702d0ed5a9bffe97c917bd7dc70de44" Dec 05 07:54:56 crc kubenswrapper[4997]: I1205 07:54:56.504456 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff5be2573614d1819bfb6085c7f5bfc14702d0ed5a9bffe97c917bd7dc70de44"} err="failed to get container status \"ff5be2573614d1819bfb6085c7f5bfc14702d0ed5a9bffe97c917bd7dc70de44\": rpc error: code = NotFound desc = could not find container \"ff5be2573614d1819bfb6085c7f5bfc14702d0ed5a9bffe97c917bd7dc70de44\": container with ID starting with ff5be2573614d1819bfb6085c7f5bfc14702d0ed5a9bffe97c917bd7dc70de44 not found: ID does not exist" Dec 05 07:54:57 crc kubenswrapper[4997]: I1205 07:54:57.766367 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e022bd86-466c-4393-8605-041774e8a636" path="/var/lib/kubelet/pods/e022bd86-466c-4393-8605-041774e8a636/volumes" Dec 05 07:55:19 crc kubenswrapper[4997]: I1205 07:55:19.770079 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:55:19 crc kubenswrapper[4997]: I1205 07:55:19.771004 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:55:19 crc kubenswrapper[4997]: I1205 07:55:19.771077 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 07:55:19 crc kubenswrapper[4997]: I1205 07:55:19.772121 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"75a9857f211838704681a7a4d52f9d31b2c917ba74ae7fa865ccbb5f50ba8e7f"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 07:55:19 crc kubenswrapper[4997]: I1205 07:55:19.772258 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://75a9857f211838704681a7a4d52f9d31b2c917ba74ae7fa865ccbb5f50ba8e7f" gracePeriod=600 Dec 05 07:55:20 crc kubenswrapper[4997]: I1205 07:55:20.636052 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="75a9857f211838704681a7a4d52f9d31b2c917ba74ae7fa865ccbb5f50ba8e7f" exitCode=0 Dec 05 07:55:20 crc kubenswrapper[4997]: I1205 07:55:20.636138 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"75a9857f211838704681a7a4d52f9d31b2c917ba74ae7fa865ccbb5f50ba8e7f"} Dec 05 07:55:20 crc kubenswrapper[4997]: I1205 07:55:20.636497 4997 scope.go:117] "RemoveContainer" containerID="5b6ddcb2a17d085fdd1fec2f83178d138d106936cee2efc6b0dee14ff9814260" Dec 05 07:55:21 crc kubenswrapper[4997]: I1205 07:55:21.651132 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571"} Dec 05 07:57:49 crc kubenswrapper[4997]: I1205 07:57:49.770743 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:57:49 crc kubenswrapper[4997]: I1205 07:57:49.771548 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:58:19 crc kubenswrapper[4997]: I1205 07:58:19.770347 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:58:19 crc kubenswrapper[4997]: I1205 07:58:19.771133 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:58:20 crc kubenswrapper[4997]: I1205 07:58:20.979595 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s9rsz"] Dec 05 07:58:20 crc kubenswrapper[4997]: E1205 07:58:20.980377 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e022bd86-466c-4393-8605-041774e8a636" containerName="extract-utilities" Dec 05 07:58:20 crc kubenswrapper[4997]: I1205 07:58:20.980398 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e022bd86-466c-4393-8605-041774e8a636" containerName="extract-utilities" Dec 05 07:58:20 crc kubenswrapper[4997]: E1205 07:58:20.980430 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e022bd86-466c-4393-8605-041774e8a636" containerName="registry-server" Dec 05 07:58:20 crc kubenswrapper[4997]: I1205 07:58:20.980438 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e022bd86-466c-4393-8605-041774e8a636" containerName="registry-server" Dec 05 07:58:20 crc kubenswrapper[4997]: E1205 07:58:20.980454 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e022bd86-466c-4393-8605-041774e8a636" containerName="extract-content" Dec 05 07:58:20 crc kubenswrapper[4997]: I1205 07:58:20.980462 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e022bd86-466c-4393-8605-041774e8a636" containerName="extract-content" Dec 05 07:58:20 crc kubenswrapper[4997]: I1205 07:58:20.980768 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e022bd86-466c-4393-8605-041774e8a636" containerName="registry-server" Dec 05 07:58:20 crc kubenswrapper[4997]: I1205 07:58:20.982130 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s9rsz" Dec 05 07:58:21 crc kubenswrapper[4997]: I1205 07:58:21.007180 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s9rsz"] Dec 05 07:58:21 crc kubenswrapper[4997]: I1205 07:58:21.083364 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43836eb0-fabf-48c6-a296-a89b2cb962e5-utilities\") pod \"redhat-operators-s9rsz\" (UID: \"43836eb0-fabf-48c6-a296-a89b2cb962e5\") " pod="openshift-marketplace/redhat-operators-s9rsz" Dec 05 07:58:21 crc kubenswrapper[4997]: I1205 07:58:21.083837 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43836eb0-fabf-48c6-a296-a89b2cb962e5-catalog-content\") pod \"redhat-operators-s9rsz\" (UID: \"43836eb0-fabf-48c6-a296-a89b2cb962e5\") " pod="openshift-marketplace/redhat-operators-s9rsz" Dec 05 07:58:21 crc kubenswrapper[4997]: I1205 07:58:21.083928 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhk7r\" (UniqueName: \"kubernetes.io/projected/43836eb0-fabf-48c6-a296-a89b2cb962e5-kube-api-access-hhk7r\") pod \"redhat-operators-s9rsz\" (UID: \"43836eb0-fabf-48c6-a296-a89b2cb962e5\") " pod="openshift-marketplace/redhat-operators-s9rsz" Dec 05 07:58:21 crc kubenswrapper[4997]: I1205 07:58:21.185517 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43836eb0-fabf-48c6-a296-a89b2cb962e5-utilities\") pod \"redhat-operators-s9rsz\" (UID: \"43836eb0-fabf-48c6-a296-a89b2cb962e5\") " pod="openshift-marketplace/redhat-operators-s9rsz" Dec 05 07:58:21 crc kubenswrapper[4997]: I1205 07:58:21.185660 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43836eb0-fabf-48c6-a296-a89b2cb962e5-catalog-content\") pod \"redhat-operators-s9rsz\" (UID: \"43836eb0-fabf-48c6-a296-a89b2cb962e5\") " pod="openshift-marketplace/redhat-operators-s9rsz" Dec 05 07:58:21 crc kubenswrapper[4997]: I1205 07:58:21.185694 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhk7r\" (UniqueName: \"kubernetes.io/projected/43836eb0-fabf-48c6-a296-a89b2cb962e5-kube-api-access-hhk7r\") pod \"redhat-operators-s9rsz\" (UID: \"43836eb0-fabf-48c6-a296-a89b2cb962e5\") " pod="openshift-marketplace/redhat-operators-s9rsz" Dec 05 07:58:21 crc kubenswrapper[4997]: I1205 07:58:21.186333 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43836eb0-fabf-48c6-a296-a89b2cb962e5-utilities\") pod \"redhat-operators-s9rsz\" (UID: \"43836eb0-fabf-48c6-a296-a89b2cb962e5\") " pod="openshift-marketplace/redhat-operators-s9rsz" Dec 05 07:58:21 crc kubenswrapper[4997]: I1205 07:58:21.186428 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43836eb0-fabf-48c6-a296-a89b2cb962e5-catalog-content\") pod \"redhat-operators-s9rsz\" (UID: \"43836eb0-fabf-48c6-a296-a89b2cb962e5\") " pod="openshift-marketplace/redhat-operators-s9rsz" Dec 05 07:58:21 crc kubenswrapper[4997]: I1205 07:58:21.211798 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhk7r\" (UniqueName: \"kubernetes.io/projected/43836eb0-fabf-48c6-a296-a89b2cb962e5-kube-api-access-hhk7r\") pod \"redhat-operators-s9rsz\" (UID: \"43836eb0-fabf-48c6-a296-a89b2cb962e5\") " pod="openshift-marketplace/redhat-operators-s9rsz" Dec 05 07:58:21 crc kubenswrapper[4997]: I1205 07:58:21.316980 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s9rsz" Dec 05 07:58:21 crc kubenswrapper[4997]: I1205 07:58:21.565915 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s9rsz"] Dec 05 07:58:21 crc kubenswrapper[4997]: E1205 07:58:21.872534 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43836eb0_fabf_48c6_a296_a89b2cb962e5.slice/crio-conmon-d49729597699f01e04ae5f4a0848ada6de19cd37073c5c55492d427f787a4324.scope\": RecentStats: unable to find data in memory cache]" Dec 05 07:58:22 crc kubenswrapper[4997]: I1205 07:58:22.432147 4997 generic.go:334] "Generic (PLEG): container finished" podID="43836eb0-fabf-48c6-a296-a89b2cb962e5" containerID="d49729597699f01e04ae5f4a0848ada6de19cd37073c5c55492d427f787a4324" exitCode=0 Dec 05 07:58:22 crc kubenswrapper[4997]: I1205 07:58:22.432395 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9rsz" event={"ID":"43836eb0-fabf-48c6-a296-a89b2cb962e5","Type":"ContainerDied","Data":"d49729597699f01e04ae5f4a0848ada6de19cd37073c5c55492d427f787a4324"} Dec 05 07:58:22 crc kubenswrapper[4997]: I1205 07:58:22.432420 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9rsz" event={"ID":"43836eb0-fabf-48c6-a296-a89b2cb962e5","Type":"ContainerStarted","Data":"88bfe8e087daea3aab197107b646409401f1b714a425fe7d8a3cabce8212eb3e"} Dec 05 07:58:22 crc kubenswrapper[4997]: I1205 07:58:22.434320 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 07:58:23 crc kubenswrapper[4997]: I1205 07:58:23.441951 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9rsz" event={"ID":"43836eb0-fabf-48c6-a296-a89b2cb962e5","Type":"ContainerStarted","Data":"41066dbc54ec328b275ea03e67820778a08227520405413c240102651aa00a2a"} Dec 05 07:58:24 crc kubenswrapper[4997]: I1205 07:58:24.452135 4997 generic.go:334] "Generic (PLEG): container finished" podID="43836eb0-fabf-48c6-a296-a89b2cb962e5" containerID="41066dbc54ec328b275ea03e67820778a08227520405413c240102651aa00a2a" exitCode=0 Dec 05 07:58:24 crc kubenswrapper[4997]: I1205 07:58:24.452202 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9rsz" event={"ID":"43836eb0-fabf-48c6-a296-a89b2cb962e5","Type":"ContainerDied","Data":"41066dbc54ec328b275ea03e67820778a08227520405413c240102651aa00a2a"} Dec 05 07:58:25 crc kubenswrapper[4997]: I1205 07:58:25.463897 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9rsz" event={"ID":"43836eb0-fabf-48c6-a296-a89b2cb962e5","Type":"ContainerStarted","Data":"8ed02cb34355a5c4cdd8f521f695b8681c9a4780c905c11d87158b56fac31212"} Dec 05 07:58:25 crc kubenswrapper[4997]: I1205 07:58:25.484750 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s9rsz" podStartSLOduration=3.055383203 podStartE2EDuration="5.484728428s" podCreationTimestamp="2025-12-05 07:58:20 +0000 UTC" firstStartedPulling="2025-12-05 07:58:22.434080312 +0000 UTC m=+3802.962987573" lastFinishedPulling="2025-12-05 07:58:24.863425527 +0000 UTC m=+3805.392332798" observedRunningTime="2025-12-05 07:58:25.482882877 +0000 UTC m=+3806.011790168" watchObservedRunningTime="2025-12-05 07:58:25.484728428 +0000 UTC m=+3806.013635719" Dec 05 07:58:31 crc kubenswrapper[4997]: I1205 07:58:31.317894 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s9rsz" Dec 05 07:58:31 crc kubenswrapper[4997]: I1205 07:58:31.318651 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s9rsz" Dec 05 07:58:31 crc kubenswrapper[4997]: I1205 07:58:31.640540 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s9rsz" Dec 05 07:58:31 crc kubenswrapper[4997]: I1205 07:58:31.683338 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s9rsz" Dec 05 07:58:31 crc kubenswrapper[4997]: I1205 07:58:31.877968 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s9rsz"] Dec 05 07:58:33 crc kubenswrapper[4997]: I1205 07:58:33.550453 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s9rsz" podUID="43836eb0-fabf-48c6-a296-a89b2cb962e5" containerName="registry-server" containerID="cri-o://8ed02cb34355a5c4cdd8f521f695b8681c9a4780c905c11d87158b56fac31212" gracePeriod=2 Dec 05 07:58:35 crc kubenswrapper[4997]: I1205 07:58:35.572451 4997 generic.go:334] "Generic (PLEG): container finished" podID="43836eb0-fabf-48c6-a296-a89b2cb962e5" containerID="8ed02cb34355a5c4cdd8f521f695b8681c9a4780c905c11d87158b56fac31212" exitCode=0 Dec 05 07:58:35 crc kubenswrapper[4997]: I1205 07:58:35.572578 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9rsz" event={"ID":"43836eb0-fabf-48c6-a296-a89b2cb962e5","Type":"ContainerDied","Data":"8ed02cb34355a5c4cdd8f521f695b8681c9a4780c905c11d87158b56fac31212"} Dec 05 07:58:35 crc kubenswrapper[4997]: I1205 07:58:35.752195 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s9rsz" Dec 05 07:58:35 crc kubenswrapper[4997]: I1205 07:58:35.895299 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43836eb0-fabf-48c6-a296-a89b2cb962e5-utilities\") pod \"43836eb0-fabf-48c6-a296-a89b2cb962e5\" (UID: \"43836eb0-fabf-48c6-a296-a89b2cb962e5\") " Dec 05 07:58:35 crc kubenswrapper[4997]: I1205 07:58:35.895431 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43836eb0-fabf-48c6-a296-a89b2cb962e5-catalog-content\") pod \"43836eb0-fabf-48c6-a296-a89b2cb962e5\" (UID: \"43836eb0-fabf-48c6-a296-a89b2cb962e5\") " Dec 05 07:58:35 crc kubenswrapper[4997]: I1205 07:58:35.895558 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhk7r\" (UniqueName: \"kubernetes.io/projected/43836eb0-fabf-48c6-a296-a89b2cb962e5-kube-api-access-hhk7r\") pod \"43836eb0-fabf-48c6-a296-a89b2cb962e5\" (UID: \"43836eb0-fabf-48c6-a296-a89b2cb962e5\") " Dec 05 07:58:35 crc kubenswrapper[4997]: I1205 07:58:35.896671 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43836eb0-fabf-48c6-a296-a89b2cb962e5-utilities" (OuterVolumeSpecName: "utilities") pod "43836eb0-fabf-48c6-a296-a89b2cb962e5" (UID: "43836eb0-fabf-48c6-a296-a89b2cb962e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:58:35 crc kubenswrapper[4997]: I1205 07:58:35.901411 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43836eb0-fabf-48c6-a296-a89b2cb962e5-kube-api-access-hhk7r" (OuterVolumeSpecName: "kube-api-access-hhk7r") pod "43836eb0-fabf-48c6-a296-a89b2cb962e5" (UID: "43836eb0-fabf-48c6-a296-a89b2cb962e5"). InnerVolumeSpecName "kube-api-access-hhk7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 07:58:35 crc kubenswrapper[4997]: I1205 07:58:35.997664 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhk7r\" (UniqueName: \"kubernetes.io/projected/43836eb0-fabf-48c6-a296-a89b2cb962e5-kube-api-access-hhk7r\") on node \"crc\" DevicePath \"\"" Dec 05 07:58:35 crc kubenswrapper[4997]: I1205 07:58:35.997701 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43836eb0-fabf-48c6-a296-a89b2cb962e5-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 07:58:36 crc kubenswrapper[4997]: I1205 07:58:36.053473 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43836eb0-fabf-48c6-a296-a89b2cb962e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "43836eb0-fabf-48c6-a296-a89b2cb962e5" (UID: "43836eb0-fabf-48c6-a296-a89b2cb962e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 07:58:36 crc kubenswrapper[4997]: I1205 07:58:36.100048 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43836eb0-fabf-48c6-a296-a89b2cb962e5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 07:58:36 crc kubenswrapper[4997]: I1205 07:58:36.584199 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9rsz" event={"ID":"43836eb0-fabf-48c6-a296-a89b2cb962e5","Type":"ContainerDied","Data":"88bfe8e087daea3aab197107b646409401f1b714a425fe7d8a3cabce8212eb3e"} Dec 05 07:58:36 crc kubenswrapper[4997]: I1205 07:58:36.584261 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s9rsz" Dec 05 07:58:36 crc kubenswrapper[4997]: I1205 07:58:36.584287 4997 scope.go:117] "RemoveContainer" containerID="8ed02cb34355a5c4cdd8f521f695b8681c9a4780c905c11d87158b56fac31212" Dec 05 07:58:36 crc kubenswrapper[4997]: I1205 07:58:36.606337 4997 scope.go:117] "RemoveContainer" containerID="41066dbc54ec328b275ea03e67820778a08227520405413c240102651aa00a2a" Dec 05 07:58:36 crc kubenswrapper[4997]: I1205 07:58:36.624681 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s9rsz"] Dec 05 07:58:36 crc kubenswrapper[4997]: I1205 07:58:36.633189 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s9rsz"] Dec 05 07:58:36 crc kubenswrapper[4997]: I1205 07:58:36.655963 4997 scope.go:117] "RemoveContainer" containerID="d49729597699f01e04ae5f4a0848ada6de19cd37073c5c55492d427f787a4324" Dec 05 07:58:37 crc kubenswrapper[4997]: I1205 07:58:37.758796 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43836eb0-fabf-48c6-a296-a89b2cb962e5" path="/var/lib/kubelet/pods/43836eb0-fabf-48c6-a296-a89b2cb962e5/volumes" Dec 05 07:58:49 crc kubenswrapper[4997]: I1205 07:58:49.769836 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 07:58:49 crc kubenswrapper[4997]: I1205 07:58:49.770557 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 07:58:49 crc kubenswrapper[4997]: I1205 07:58:49.770647 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 07:58:49 crc kubenswrapper[4997]: I1205 07:58:49.771859 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 07:58:49 crc kubenswrapper[4997]: I1205 07:58:49.771975 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" gracePeriod=600 Dec 05 07:58:49 crc kubenswrapper[4997]: E1205 07:58:49.922967 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:58:50 crc kubenswrapper[4997]: I1205 07:58:50.731965 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" exitCode=0 Dec 05 07:58:50 crc kubenswrapper[4997]: I1205 07:58:50.732030 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571"} Dec 05 07:58:50 crc kubenswrapper[4997]: I1205 07:58:50.732839 4997 scope.go:117] "RemoveContainer" containerID="75a9857f211838704681a7a4d52f9d31b2c917ba74ae7fa865ccbb5f50ba8e7f" Dec 05 07:58:50 crc kubenswrapper[4997]: I1205 07:58:50.733352 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 07:58:50 crc kubenswrapper[4997]: E1205 07:58:50.733715 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:59:02 crc kubenswrapper[4997]: I1205 07:59:02.749140 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 07:59:02 crc kubenswrapper[4997]: E1205 07:59:02.750269 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:59:15 crc kubenswrapper[4997]: I1205 07:59:15.750142 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 07:59:15 crc kubenswrapper[4997]: E1205 07:59:15.751326 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:59:28 crc kubenswrapper[4997]: I1205 07:59:28.750009 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 07:59:28 crc kubenswrapper[4997]: E1205 07:59:28.751356 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:59:40 crc kubenswrapper[4997]: I1205 07:59:40.750219 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 07:59:40 crc kubenswrapper[4997]: E1205 07:59:40.751278 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 07:59:51 crc kubenswrapper[4997]: I1205 07:59:51.750034 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 07:59:51 crc kubenswrapper[4997]: E1205 07:59:51.751008 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.188985 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx"] Dec 05 08:00:00 crc kubenswrapper[4997]: E1205 08:00:00.189800 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43836eb0-fabf-48c6-a296-a89b2cb962e5" containerName="extract-content" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.189814 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="43836eb0-fabf-48c6-a296-a89b2cb962e5" containerName="extract-content" Dec 05 08:00:00 crc kubenswrapper[4997]: E1205 08:00:00.189826 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43836eb0-fabf-48c6-a296-a89b2cb962e5" containerName="registry-server" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.189832 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="43836eb0-fabf-48c6-a296-a89b2cb962e5" containerName="registry-server" Dec 05 08:00:00 crc kubenswrapper[4997]: E1205 08:00:00.189841 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43836eb0-fabf-48c6-a296-a89b2cb962e5" containerName="extract-utilities" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.189848 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="43836eb0-fabf-48c6-a296-a89b2cb962e5" containerName="extract-utilities" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.190020 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="43836eb0-fabf-48c6-a296-a89b2cb962e5" containerName="registry-server" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.190596 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.193014 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.194038 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.203144 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx"] Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.318294 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljzf7\" (UniqueName: \"kubernetes.io/projected/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-kube-api-access-ljzf7\") pod \"collect-profiles-29415360-l9xdx\" (UID: \"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.318347 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-secret-volume\") pod \"collect-profiles-29415360-l9xdx\" (UID: \"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.318407 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-config-volume\") pod \"collect-profiles-29415360-l9xdx\" (UID: \"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.419736 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljzf7\" (UniqueName: \"kubernetes.io/projected/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-kube-api-access-ljzf7\") pod \"collect-profiles-29415360-l9xdx\" (UID: \"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.419808 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-secret-volume\") pod \"collect-profiles-29415360-l9xdx\" (UID: \"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.419872 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-config-volume\") pod \"collect-profiles-29415360-l9xdx\" (UID: \"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.420949 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-config-volume\") pod \"collect-profiles-29415360-l9xdx\" (UID: \"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.427725 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-secret-volume\") pod \"collect-profiles-29415360-l9xdx\" (UID: \"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.450108 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljzf7\" (UniqueName: \"kubernetes.io/projected/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-kube-api-access-ljzf7\") pod \"collect-profiles-29415360-l9xdx\" (UID: \"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.511398 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx" Dec 05 08:00:00 crc kubenswrapper[4997]: I1205 08:00:00.731818 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx"] Dec 05 08:00:01 crc kubenswrapper[4997]: I1205 08:00:01.365023 4997 generic.go:334] "Generic (PLEG): container finished" podID="3e6e989f-8977-4f2f-aff8-e9ca326b8fdf" containerID="4f0dc440d0ee08775c9046a60c51145b81e29935ac0f67b442dcccae9eedd3f6" exitCode=0 Dec 05 08:00:01 crc kubenswrapper[4997]: I1205 08:00:01.365070 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx" event={"ID":"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf","Type":"ContainerDied","Data":"4f0dc440d0ee08775c9046a60c51145b81e29935ac0f67b442dcccae9eedd3f6"} Dec 05 08:00:01 crc kubenswrapper[4997]: I1205 08:00:01.365100 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx" event={"ID":"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf","Type":"ContainerStarted","Data":"2048bd1ef9c52633d4a75ba530f120c0418e8cb486d7d8d488e7c7818aa20e6f"} Dec 05 08:00:02 crc kubenswrapper[4997]: I1205 08:00:02.688293 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx" Dec 05 08:00:02 crc kubenswrapper[4997]: I1205 08:00:02.754187 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-config-volume\") pod \"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf\" (UID: \"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf\") " Dec 05 08:00:02 crc kubenswrapper[4997]: I1205 08:00:02.754289 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljzf7\" (UniqueName: \"kubernetes.io/projected/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-kube-api-access-ljzf7\") pod \"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf\" (UID: \"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf\") " Dec 05 08:00:02 crc kubenswrapper[4997]: I1205 08:00:02.754331 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-secret-volume\") pod \"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf\" (UID: \"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf\") " Dec 05 08:00:02 crc kubenswrapper[4997]: I1205 08:00:02.755036 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-config-volume" (OuterVolumeSpecName: "config-volume") pod "3e6e989f-8977-4f2f-aff8-e9ca326b8fdf" (UID: "3e6e989f-8977-4f2f-aff8-e9ca326b8fdf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:00:02 crc kubenswrapper[4997]: I1205 08:00:02.760511 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-kube-api-access-ljzf7" (OuterVolumeSpecName: "kube-api-access-ljzf7") pod "3e6e989f-8977-4f2f-aff8-e9ca326b8fdf" (UID: "3e6e989f-8977-4f2f-aff8-e9ca326b8fdf"). InnerVolumeSpecName "kube-api-access-ljzf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:00:02 crc kubenswrapper[4997]: I1205 08:00:02.761205 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3e6e989f-8977-4f2f-aff8-e9ca326b8fdf" (UID: "3e6e989f-8977-4f2f-aff8-e9ca326b8fdf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:00:02 crc kubenswrapper[4997]: I1205 08:00:02.856265 4997 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 08:00:02 crc kubenswrapper[4997]: I1205 08:00:02.856304 4997 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 08:00:02 crc kubenswrapper[4997]: I1205 08:00:02.856315 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljzf7\" (UniqueName: \"kubernetes.io/projected/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf-kube-api-access-ljzf7\") on node \"crc\" DevicePath \"\"" Dec 05 08:00:03 crc kubenswrapper[4997]: I1205 08:00:03.385343 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx" event={"ID":"3e6e989f-8977-4f2f-aff8-e9ca326b8fdf","Type":"ContainerDied","Data":"2048bd1ef9c52633d4a75ba530f120c0418e8cb486d7d8d488e7c7818aa20e6f"} Dec 05 08:00:03 crc kubenswrapper[4997]: I1205 08:00:03.385396 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2048bd1ef9c52633d4a75ba530f120c0418e8cb486d7d8d488e7c7818aa20e6f" Dec 05 08:00:03 crc kubenswrapper[4997]: I1205 08:00:03.385744 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx" Dec 05 08:00:03 crc kubenswrapper[4997]: I1205 08:00:03.763251 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm"] Dec 05 08:00:03 crc kubenswrapper[4997]: I1205 08:00:03.768631 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415315-m2gdm"] Dec 05 08:00:05 crc kubenswrapper[4997]: I1205 08:00:05.767574 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb045b24-fb27-4f7e-af8c-e1206fad74cd" path="/var/lib/kubelet/pods/cb045b24-fb27-4f7e-af8c-e1206fad74cd/volumes" Dec 05 08:00:06 crc kubenswrapper[4997]: I1205 08:00:06.749566 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:00:06 crc kubenswrapper[4997]: E1205 08:00:06.750178 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:00:11 crc kubenswrapper[4997]: I1205 08:00:11.560514 4997 scope.go:117] "RemoveContainer" containerID="55c6a36d35ea7edcb8f82a46e34f0555480e13f08ce0ed00ebf4d91e920a6a4e" Dec 05 08:00:18 crc kubenswrapper[4997]: I1205 08:00:18.749543 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:00:18 crc kubenswrapper[4997]: E1205 08:00:18.750384 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:00:32 crc kubenswrapper[4997]: I1205 08:00:32.749247 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:00:32 crc kubenswrapper[4997]: E1205 08:00:32.750044 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:00:45 crc kubenswrapper[4997]: I1205 08:00:45.750066 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:00:45 crc kubenswrapper[4997]: E1205 08:00:45.751214 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:00:58 crc kubenswrapper[4997]: I1205 08:00:58.749845 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:00:58 crc kubenswrapper[4997]: E1205 08:00:58.751039 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:01:10 crc kubenswrapper[4997]: I1205 08:01:10.749462 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:01:10 crc kubenswrapper[4997]: E1205 08:01:10.750360 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:01:25 crc kubenswrapper[4997]: I1205 08:01:25.749403 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:01:25 crc kubenswrapper[4997]: E1205 08:01:25.750310 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:01:40 crc kubenswrapper[4997]: I1205 08:01:40.749229 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:01:40 crc kubenswrapper[4997]: E1205 08:01:40.750298 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:01:51 crc kubenswrapper[4997]: I1205 08:01:51.749487 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:01:51 crc kubenswrapper[4997]: E1205 08:01:51.750803 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:02:06 crc kubenswrapper[4997]: I1205 08:02:06.749650 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:02:06 crc kubenswrapper[4997]: E1205 08:02:06.750359 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:02:17 crc kubenswrapper[4997]: I1205 08:02:17.749577 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:02:17 crc kubenswrapper[4997]: E1205 08:02:17.750607 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:02:29 crc kubenswrapper[4997]: I1205 08:02:29.757361 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:02:29 crc kubenswrapper[4997]: E1205 08:02:29.758234 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:02:41 crc kubenswrapper[4997]: I1205 08:02:41.750001 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:02:41 crc kubenswrapper[4997]: E1205 08:02:41.751162 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:02:56 crc kubenswrapper[4997]: I1205 08:02:56.748873 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:02:56 crc kubenswrapper[4997]: E1205 08:02:56.749801 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:03:08 crc kubenswrapper[4997]: I1205 08:03:08.749917 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:03:08 crc kubenswrapper[4997]: E1205 08:03:08.753151 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:03:23 crc kubenswrapper[4997]: I1205 08:03:23.750012 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:03:23 crc kubenswrapper[4997]: E1205 08:03:23.753997 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:03:35 crc kubenswrapper[4997]: I1205 08:03:35.749410 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:03:35 crc kubenswrapper[4997]: E1205 08:03:35.750192 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:03:46 crc kubenswrapper[4997]: I1205 08:03:46.749981 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:03:46 crc kubenswrapper[4997]: E1205 08:03:46.751176 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:03:52 crc kubenswrapper[4997]: I1205 08:03:52.537866 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l8vhf"] Dec 05 08:03:52 crc kubenswrapper[4997]: E1205 08:03:52.539379 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e6e989f-8977-4f2f-aff8-e9ca326b8fdf" containerName="collect-profiles" Dec 05 08:03:52 crc kubenswrapper[4997]: I1205 08:03:52.539430 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e6e989f-8977-4f2f-aff8-e9ca326b8fdf" containerName="collect-profiles" Dec 05 08:03:52 crc kubenswrapper[4997]: I1205 08:03:52.539852 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e6e989f-8977-4f2f-aff8-e9ca326b8fdf" containerName="collect-profiles" Dec 05 08:03:52 crc kubenswrapper[4997]: I1205 08:03:52.542412 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l8vhf" Dec 05 08:03:52 crc kubenswrapper[4997]: I1205 08:03:52.551344 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l8vhf"] Dec 05 08:03:52 crc kubenswrapper[4997]: I1205 08:03:52.685769 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx6w7\" (UniqueName: \"kubernetes.io/projected/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-kube-api-access-cx6w7\") pod \"community-operators-l8vhf\" (UID: \"aa47b60f-94ca-4053-8dfb-fe7c775b22d6\") " pod="openshift-marketplace/community-operators-l8vhf" Dec 05 08:03:52 crc kubenswrapper[4997]: I1205 08:03:52.686166 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-catalog-content\") pod \"community-operators-l8vhf\" (UID: \"aa47b60f-94ca-4053-8dfb-fe7c775b22d6\") " pod="openshift-marketplace/community-operators-l8vhf" Dec 05 08:03:52 crc kubenswrapper[4997]: I1205 08:03:52.686856 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-utilities\") pod \"community-operators-l8vhf\" (UID: \"aa47b60f-94ca-4053-8dfb-fe7c775b22d6\") " pod="openshift-marketplace/community-operators-l8vhf" Dec 05 08:03:52 crc kubenswrapper[4997]: I1205 08:03:52.789222 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-catalog-content\") pod \"community-operators-l8vhf\" (UID: \"aa47b60f-94ca-4053-8dfb-fe7c775b22d6\") " pod="openshift-marketplace/community-operators-l8vhf" Dec 05 08:03:52 crc kubenswrapper[4997]: I1205 08:03:52.789357 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-utilities\") pod \"community-operators-l8vhf\" (UID: \"aa47b60f-94ca-4053-8dfb-fe7c775b22d6\") " pod="openshift-marketplace/community-operators-l8vhf" Dec 05 08:03:52 crc kubenswrapper[4997]: I1205 08:03:52.789480 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx6w7\" (UniqueName: \"kubernetes.io/projected/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-kube-api-access-cx6w7\") pod \"community-operators-l8vhf\" (UID: \"aa47b60f-94ca-4053-8dfb-fe7c775b22d6\") " pod="openshift-marketplace/community-operators-l8vhf" Dec 05 08:03:52 crc kubenswrapper[4997]: I1205 08:03:52.790576 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-catalog-content\") pod \"community-operators-l8vhf\" (UID: \"aa47b60f-94ca-4053-8dfb-fe7c775b22d6\") " pod="openshift-marketplace/community-operators-l8vhf" Dec 05 08:03:52 crc kubenswrapper[4997]: I1205 08:03:52.790927 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-utilities\") pod \"community-operators-l8vhf\" (UID: \"aa47b60f-94ca-4053-8dfb-fe7c775b22d6\") " pod="openshift-marketplace/community-operators-l8vhf" Dec 05 08:03:52 crc kubenswrapper[4997]: I1205 08:03:52.822856 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx6w7\" (UniqueName: \"kubernetes.io/projected/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-kube-api-access-cx6w7\") pod \"community-operators-l8vhf\" (UID: \"aa47b60f-94ca-4053-8dfb-fe7c775b22d6\") " pod="openshift-marketplace/community-operators-l8vhf" Dec 05 08:03:52 crc kubenswrapper[4997]: I1205 08:03:52.876710 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l8vhf" Dec 05 08:03:53 crc kubenswrapper[4997]: I1205 08:03:53.173830 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l8vhf"] Dec 05 08:03:53 crc kubenswrapper[4997]: I1205 08:03:53.770568 4997 generic.go:334] "Generic (PLEG): container finished" podID="aa47b60f-94ca-4053-8dfb-fe7c775b22d6" containerID="1afaedbc21d4cd1df4a4968479a904403266e2ec32e10843224e80fde009a02d" exitCode=0 Dec 05 08:03:53 crc kubenswrapper[4997]: I1205 08:03:53.770668 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8vhf" event={"ID":"aa47b60f-94ca-4053-8dfb-fe7c775b22d6","Type":"ContainerDied","Data":"1afaedbc21d4cd1df4a4968479a904403266e2ec32e10843224e80fde009a02d"} Dec 05 08:03:53 crc kubenswrapper[4997]: I1205 08:03:53.770992 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8vhf" event={"ID":"aa47b60f-94ca-4053-8dfb-fe7c775b22d6","Type":"ContainerStarted","Data":"c3c08b44d98ffb0ac293ba0b0228e63ad1bc7862870c724fa3be88ea027268b0"} Dec 05 08:03:53 crc kubenswrapper[4997]: I1205 08:03:53.772875 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 08:03:54 crc kubenswrapper[4997]: I1205 08:03:54.790049 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8vhf" event={"ID":"aa47b60f-94ca-4053-8dfb-fe7c775b22d6","Type":"ContainerStarted","Data":"a4e33123ddce990cf5a694e79fcb73390933cb5ea15755de1f94b26987b3adef"} Dec 05 08:03:55 crc kubenswrapper[4997]: I1205 08:03:55.805889 4997 generic.go:334] "Generic (PLEG): container finished" podID="aa47b60f-94ca-4053-8dfb-fe7c775b22d6" containerID="a4e33123ddce990cf5a694e79fcb73390933cb5ea15755de1f94b26987b3adef" exitCode=0 Dec 05 08:03:55 crc kubenswrapper[4997]: I1205 08:03:55.805963 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8vhf" event={"ID":"aa47b60f-94ca-4053-8dfb-fe7c775b22d6","Type":"ContainerDied","Data":"a4e33123ddce990cf5a694e79fcb73390933cb5ea15755de1f94b26987b3adef"} Dec 05 08:03:56 crc kubenswrapper[4997]: I1205 08:03:56.818173 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8vhf" event={"ID":"aa47b60f-94ca-4053-8dfb-fe7c775b22d6","Type":"ContainerStarted","Data":"e4766878bb79df3f647204ad9dbd9c5bad952029deabaf9254177732bf1b98c1"} Dec 05 08:03:56 crc kubenswrapper[4997]: I1205 08:03:56.851587 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l8vhf" podStartSLOduration=2.393883261 podStartE2EDuration="4.851419817s" podCreationTimestamp="2025-12-05 08:03:52 +0000 UTC" firstStartedPulling="2025-12-05 08:03:53.772521813 +0000 UTC m=+4134.301429084" lastFinishedPulling="2025-12-05 08:03:56.230058379 +0000 UTC m=+4136.758965640" observedRunningTime="2025-12-05 08:03:56.843309987 +0000 UTC m=+4137.372217328" watchObservedRunningTime="2025-12-05 08:03:56.851419817 +0000 UTC m=+4137.380327108" Dec 05 08:03:59 crc kubenswrapper[4997]: I1205 08:03:59.766155 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:04:00 crc kubenswrapper[4997]: I1205 08:04:00.861152 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"e65f871a5d8f1b6b491a6cfe1489bad2714fbbf443695e3bcd99bcb1a0d11097"} Dec 05 08:04:02 crc kubenswrapper[4997]: I1205 08:04:02.877711 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l8vhf" Dec 05 08:04:02 crc kubenswrapper[4997]: I1205 08:04:02.878217 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l8vhf" Dec 05 08:04:02 crc kubenswrapper[4997]: I1205 08:04:02.954919 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l8vhf" Dec 05 08:04:03 crc kubenswrapper[4997]: I1205 08:04:03.964658 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l8vhf" Dec 05 08:04:04 crc kubenswrapper[4997]: I1205 08:04:04.044065 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l8vhf"] Dec 05 08:04:05 crc kubenswrapper[4997]: I1205 08:04:05.902836 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l8vhf" podUID="aa47b60f-94ca-4053-8dfb-fe7c775b22d6" containerName="registry-server" containerID="cri-o://e4766878bb79df3f647204ad9dbd9c5bad952029deabaf9254177732bf1b98c1" gracePeriod=2 Dec 05 08:04:06 crc kubenswrapper[4997]: I1205 08:04:06.896229 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l8vhf" Dec 05 08:04:06 crc kubenswrapper[4997]: I1205 08:04:06.914448 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-catalog-content\") pod \"aa47b60f-94ca-4053-8dfb-fe7c775b22d6\" (UID: \"aa47b60f-94ca-4053-8dfb-fe7c775b22d6\") " Dec 05 08:04:06 crc kubenswrapper[4997]: I1205 08:04:06.914515 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cx6w7\" (UniqueName: \"kubernetes.io/projected/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-kube-api-access-cx6w7\") pod \"aa47b60f-94ca-4053-8dfb-fe7c775b22d6\" (UID: \"aa47b60f-94ca-4053-8dfb-fe7c775b22d6\") " Dec 05 08:04:06 crc kubenswrapper[4997]: I1205 08:04:06.914569 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-utilities\") pod \"aa47b60f-94ca-4053-8dfb-fe7c775b22d6\" (UID: \"aa47b60f-94ca-4053-8dfb-fe7c775b22d6\") " Dec 05 08:04:06 crc kubenswrapper[4997]: I1205 08:04:06.915795 4997 generic.go:334] "Generic (PLEG): container finished" podID="aa47b60f-94ca-4053-8dfb-fe7c775b22d6" containerID="e4766878bb79df3f647204ad9dbd9c5bad952029deabaf9254177732bf1b98c1" exitCode=0 Dec 05 08:04:06 crc kubenswrapper[4997]: I1205 08:04:06.915916 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l8vhf" Dec 05 08:04:06 crc kubenswrapper[4997]: I1205 08:04:06.916007 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8vhf" event={"ID":"aa47b60f-94ca-4053-8dfb-fe7c775b22d6","Type":"ContainerDied","Data":"e4766878bb79df3f647204ad9dbd9c5bad952029deabaf9254177732bf1b98c1"} Dec 05 08:04:06 crc kubenswrapper[4997]: I1205 08:04:06.916092 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8vhf" event={"ID":"aa47b60f-94ca-4053-8dfb-fe7c775b22d6","Type":"ContainerDied","Data":"c3c08b44d98ffb0ac293ba0b0228e63ad1bc7862870c724fa3be88ea027268b0"} Dec 05 08:04:06 crc kubenswrapper[4997]: I1205 08:04:06.916120 4997 scope.go:117] "RemoveContainer" containerID="e4766878bb79df3f647204ad9dbd9c5bad952029deabaf9254177732bf1b98c1" Dec 05 08:04:06 crc kubenswrapper[4997]: I1205 08:04:06.916034 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-utilities" (OuterVolumeSpecName: "utilities") pod "aa47b60f-94ca-4053-8dfb-fe7c775b22d6" (UID: "aa47b60f-94ca-4053-8dfb-fe7c775b22d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:04:06 crc kubenswrapper[4997]: I1205 08:04:06.929424 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-kube-api-access-cx6w7" (OuterVolumeSpecName: "kube-api-access-cx6w7") pod "aa47b60f-94ca-4053-8dfb-fe7c775b22d6" (UID: "aa47b60f-94ca-4053-8dfb-fe7c775b22d6"). InnerVolumeSpecName "kube-api-access-cx6w7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:04:06 crc kubenswrapper[4997]: I1205 08:04:06.967841 4997 scope.go:117] "RemoveContainer" containerID="a4e33123ddce990cf5a694e79fcb73390933cb5ea15755de1f94b26987b3adef" Dec 05 08:04:06 crc kubenswrapper[4997]: I1205 08:04:06.995519 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa47b60f-94ca-4053-8dfb-fe7c775b22d6" (UID: "aa47b60f-94ca-4053-8dfb-fe7c775b22d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:04:07 crc kubenswrapper[4997]: I1205 08:04:07.001851 4997 scope.go:117] "RemoveContainer" containerID="1afaedbc21d4cd1df4a4968479a904403266e2ec32e10843224e80fde009a02d" Dec 05 08:04:07 crc kubenswrapper[4997]: I1205 08:04:07.015817 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:04:07 crc kubenswrapper[4997]: I1205 08:04:07.015842 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cx6w7\" (UniqueName: \"kubernetes.io/projected/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-kube-api-access-cx6w7\") on node \"crc\" DevicePath \"\"" Dec 05 08:04:07 crc kubenswrapper[4997]: I1205 08:04:07.015876 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa47b60f-94ca-4053-8dfb-fe7c775b22d6-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:04:07 crc kubenswrapper[4997]: I1205 08:04:07.060848 4997 scope.go:117] "RemoveContainer" containerID="e4766878bb79df3f647204ad9dbd9c5bad952029deabaf9254177732bf1b98c1" Dec 05 08:04:07 crc kubenswrapper[4997]: E1205 08:04:07.062034 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4766878bb79df3f647204ad9dbd9c5bad952029deabaf9254177732bf1b98c1\": container with ID starting with e4766878bb79df3f647204ad9dbd9c5bad952029deabaf9254177732bf1b98c1 not found: ID does not exist" containerID="e4766878bb79df3f647204ad9dbd9c5bad952029deabaf9254177732bf1b98c1" Dec 05 08:04:07 crc kubenswrapper[4997]: I1205 08:04:07.062077 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4766878bb79df3f647204ad9dbd9c5bad952029deabaf9254177732bf1b98c1"} err="failed to get container status \"e4766878bb79df3f647204ad9dbd9c5bad952029deabaf9254177732bf1b98c1\": rpc error: code = NotFound desc = could not find container \"e4766878bb79df3f647204ad9dbd9c5bad952029deabaf9254177732bf1b98c1\": container with ID starting with e4766878bb79df3f647204ad9dbd9c5bad952029deabaf9254177732bf1b98c1 not found: ID does not exist" Dec 05 08:04:07 crc kubenswrapper[4997]: I1205 08:04:07.062135 4997 scope.go:117] "RemoveContainer" containerID="a4e33123ddce990cf5a694e79fcb73390933cb5ea15755de1f94b26987b3adef" Dec 05 08:04:07 crc kubenswrapper[4997]: E1205 08:04:07.062795 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4e33123ddce990cf5a694e79fcb73390933cb5ea15755de1f94b26987b3adef\": container with ID starting with a4e33123ddce990cf5a694e79fcb73390933cb5ea15755de1f94b26987b3adef not found: ID does not exist" containerID="a4e33123ddce990cf5a694e79fcb73390933cb5ea15755de1f94b26987b3adef" Dec 05 08:04:07 crc kubenswrapper[4997]: I1205 08:04:07.062830 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4e33123ddce990cf5a694e79fcb73390933cb5ea15755de1f94b26987b3adef"} err="failed to get container status \"a4e33123ddce990cf5a694e79fcb73390933cb5ea15755de1f94b26987b3adef\": rpc error: code = NotFound desc = could not find container \"a4e33123ddce990cf5a694e79fcb73390933cb5ea15755de1f94b26987b3adef\": container with ID starting with a4e33123ddce990cf5a694e79fcb73390933cb5ea15755de1f94b26987b3adef not found: ID does not exist" Dec 05 08:04:07 crc kubenswrapper[4997]: I1205 08:04:07.062852 4997 scope.go:117] "RemoveContainer" containerID="1afaedbc21d4cd1df4a4968479a904403266e2ec32e10843224e80fde009a02d" Dec 05 08:04:07 crc kubenswrapper[4997]: E1205 08:04:07.063114 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1afaedbc21d4cd1df4a4968479a904403266e2ec32e10843224e80fde009a02d\": container with ID starting with 1afaedbc21d4cd1df4a4968479a904403266e2ec32e10843224e80fde009a02d not found: ID does not exist" containerID="1afaedbc21d4cd1df4a4968479a904403266e2ec32e10843224e80fde009a02d" Dec 05 08:04:07 crc kubenswrapper[4997]: I1205 08:04:07.063161 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1afaedbc21d4cd1df4a4968479a904403266e2ec32e10843224e80fde009a02d"} err="failed to get container status \"1afaedbc21d4cd1df4a4968479a904403266e2ec32e10843224e80fde009a02d\": rpc error: code = NotFound desc = could not find container \"1afaedbc21d4cd1df4a4968479a904403266e2ec32e10843224e80fde009a02d\": container with ID starting with 1afaedbc21d4cd1df4a4968479a904403266e2ec32e10843224e80fde009a02d not found: ID does not exist" Dec 05 08:04:07 crc kubenswrapper[4997]: I1205 08:04:07.277493 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l8vhf"] Dec 05 08:04:07 crc kubenswrapper[4997]: I1205 08:04:07.290125 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l8vhf"] Dec 05 08:04:07 crc kubenswrapper[4997]: I1205 08:04:07.762267 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa47b60f-94ca-4053-8dfb-fe7c775b22d6" path="/var/lib/kubelet/pods/aa47b60f-94ca-4053-8dfb-fe7c775b22d6/volumes" Dec 05 08:04:09 crc kubenswrapper[4997]: I1205 08:04:09.444064 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nq8s6"] Dec 05 08:04:09 crc kubenswrapper[4997]: E1205 08:04:09.444482 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa47b60f-94ca-4053-8dfb-fe7c775b22d6" containerName="extract-content" Dec 05 08:04:09 crc kubenswrapper[4997]: I1205 08:04:09.444502 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa47b60f-94ca-4053-8dfb-fe7c775b22d6" containerName="extract-content" Dec 05 08:04:09 crc kubenswrapper[4997]: E1205 08:04:09.444520 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa47b60f-94ca-4053-8dfb-fe7c775b22d6" containerName="registry-server" Dec 05 08:04:09 crc kubenswrapper[4997]: I1205 08:04:09.444530 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa47b60f-94ca-4053-8dfb-fe7c775b22d6" containerName="registry-server" Dec 05 08:04:09 crc kubenswrapper[4997]: E1205 08:04:09.444569 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa47b60f-94ca-4053-8dfb-fe7c775b22d6" containerName="extract-utilities" Dec 05 08:04:09 crc kubenswrapper[4997]: I1205 08:04:09.444581 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa47b60f-94ca-4053-8dfb-fe7c775b22d6" containerName="extract-utilities" Dec 05 08:04:09 crc kubenswrapper[4997]: I1205 08:04:09.444830 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa47b60f-94ca-4053-8dfb-fe7c775b22d6" containerName="registry-server" Dec 05 08:04:09 crc kubenswrapper[4997]: I1205 08:04:09.446809 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nq8s6" Dec 05 08:04:09 crc kubenswrapper[4997]: I1205 08:04:09.453543 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhl96\" (UniqueName: \"kubernetes.io/projected/822d7260-aa6d-49c1-a8a6-dd4716f174e2-kube-api-access-vhl96\") pod \"redhat-marketplace-nq8s6\" (UID: \"822d7260-aa6d-49c1-a8a6-dd4716f174e2\") " pod="openshift-marketplace/redhat-marketplace-nq8s6" Dec 05 08:04:09 crc kubenswrapper[4997]: I1205 08:04:09.453855 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/822d7260-aa6d-49c1-a8a6-dd4716f174e2-catalog-content\") pod \"redhat-marketplace-nq8s6\" (UID: \"822d7260-aa6d-49c1-a8a6-dd4716f174e2\") " pod="openshift-marketplace/redhat-marketplace-nq8s6" Dec 05 08:04:09 crc kubenswrapper[4997]: I1205 08:04:09.453890 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/822d7260-aa6d-49c1-a8a6-dd4716f174e2-utilities\") pod \"redhat-marketplace-nq8s6\" (UID: \"822d7260-aa6d-49c1-a8a6-dd4716f174e2\") " pod="openshift-marketplace/redhat-marketplace-nq8s6" Dec 05 08:04:09 crc kubenswrapper[4997]: I1205 08:04:09.469777 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nq8s6"] Dec 05 08:04:09 crc kubenswrapper[4997]: I1205 08:04:09.555310 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhl96\" (UniqueName: \"kubernetes.io/projected/822d7260-aa6d-49c1-a8a6-dd4716f174e2-kube-api-access-vhl96\") pod \"redhat-marketplace-nq8s6\" (UID: \"822d7260-aa6d-49c1-a8a6-dd4716f174e2\") " pod="openshift-marketplace/redhat-marketplace-nq8s6" Dec 05 08:04:09 crc kubenswrapper[4997]: I1205 08:04:09.555527 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/822d7260-aa6d-49c1-a8a6-dd4716f174e2-catalog-content\") pod \"redhat-marketplace-nq8s6\" (UID: \"822d7260-aa6d-49c1-a8a6-dd4716f174e2\") " pod="openshift-marketplace/redhat-marketplace-nq8s6" Dec 05 08:04:09 crc kubenswrapper[4997]: I1205 08:04:09.555597 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/822d7260-aa6d-49c1-a8a6-dd4716f174e2-utilities\") pod \"redhat-marketplace-nq8s6\" (UID: \"822d7260-aa6d-49c1-a8a6-dd4716f174e2\") " pod="openshift-marketplace/redhat-marketplace-nq8s6" Dec 05 08:04:09 crc kubenswrapper[4997]: I1205 08:04:09.556094 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/822d7260-aa6d-49c1-a8a6-dd4716f174e2-utilities\") pod \"redhat-marketplace-nq8s6\" (UID: \"822d7260-aa6d-49c1-a8a6-dd4716f174e2\") " pod="openshift-marketplace/redhat-marketplace-nq8s6" Dec 05 08:04:09 crc kubenswrapper[4997]: I1205 08:04:09.556179 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/822d7260-aa6d-49c1-a8a6-dd4716f174e2-catalog-content\") pod \"redhat-marketplace-nq8s6\" (UID: \"822d7260-aa6d-49c1-a8a6-dd4716f174e2\") " pod="openshift-marketplace/redhat-marketplace-nq8s6" Dec 05 08:04:09 crc kubenswrapper[4997]: I1205 08:04:09.582872 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhl96\" (UniqueName: \"kubernetes.io/projected/822d7260-aa6d-49c1-a8a6-dd4716f174e2-kube-api-access-vhl96\") pod \"redhat-marketplace-nq8s6\" (UID: \"822d7260-aa6d-49c1-a8a6-dd4716f174e2\") " pod="openshift-marketplace/redhat-marketplace-nq8s6" Dec 05 08:04:09 crc kubenswrapper[4997]: I1205 08:04:09.772631 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nq8s6" Dec 05 08:04:10 crc kubenswrapper[4997]: W1205 08:04:10.230452 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod822d7260_aa6d_49c1_a8a6_dd4716f174e2.slice/crio-04fc9005d8e9b86f5598b5f178548ef215947fbce010d3470728de6c3e68e8fd WatchSource:0}: Error finding container 04fc9005d8e9b86f5598b5f178548ef215947fbce010d3470728de6c3e68e8fd: Status 404 returned error can't find the container with id 04fc9005d8e9b86f5598b5f178548ef215947fbce010d3470728de6c3e68e8fd Dec 05 08:04:10 crc kubenswrapper[4997]: I1205 08:04:10.231429 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nq8s6"] Dec 05 08:04:10 crc kubenswrapper[4997]: I1205 08:04:10.961027 4997 generic.go:334] "Generic (PLEG): container finished" podID="822d7260-aa6d-49c1-a8a6-dd4716f174e2" containerID="119c07f65354402c18ba18b9a29066c0901e3d8bde9d1ab5f55ff536bbf08db3" exitCode=0 Dec 05 08:04:10 crc kubenswrapper[4997]: I1205 08:04:10.961178 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nq8s6" event={"ID":"822d7260-aa6d-49c1-a8a6-dd4716f174e2","Type":"ContainerDied","Data":"119c07f65354402c18ba18b9a29066c0901e3d8bde9d1ab5f55ff536bbf08db3"} Dec 05 08:04:10 crc kubenswrapper[4997]: I1205 08:04:10.961526 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nq8s6" event={"ID":"822d7260-aa6d-49c1-a8a6-dd4716f174e2","Type":"ContainerStarted","Data":"04fc9005d8e9b86f5598b5f178548ef215947fbce010d3470728de6c3e68e8fd"} Dec 05 08:04:12 crc kubenswrapper[4997]: I1205 08:04:12.987202 4997 generic.go:334] "Generic (PLEG): container finished" podID="822d7260-aa6d-49c1-a8a6-dd4716f174e2" containerID="67c6acd686c1e82f437d82e0ad4e2d6419336435d7fc66b5ee9aac90a1583991" exitCode=0 Dec 05 08:04:12 crc kubenswrapper[4997]: I1205 08:04:12.987289 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nq8s6" event={"ID":"822d7260-aa6d-49c1-a8a6-dd4716f174e2","Type":"ContainerDied","Data":"67c6acd686c1e82f437d82e0ad4e2d6419336435d7fc66b5ee9aac90a1583991"} Dec 05 08:04:13 crc kubenswrapper[4997]: I1205 08:04:13.997995 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nq8s6" event={"ID":"822d7260-aa6d-49c1-a8a6-dd4716f174e2","Type":"ContainerStarted","Data":"51ddd4ced978720284ce8b0f1624d3efcc5dc49c7afcc917d2a309a590e1a540"} Dec 05 08:04:14 crc kubenswrapper[4997]: I1205 08:04:14.027551 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nq8s6" podStartSLOduration=2.5742150219999997 podStartE2EDuration="5.027534875s" podCreationTimestamp="2025-12-05 08:04:09 +0000 UTC" firstStartedPulling="2025-12-05 08:04:10.963101502 +0000 UTC m=+4151.492008803" lastFinishedPulling="2025-12-05 08:04:13.416421385 +0000 UTC m=+4153.945328656" observedRunningTime="2025-12-05 08:04:14.023123315 +0000 UTC m=+4154.552030576" watchObservedRunningTime="2025-12-05 08:04:14.027534875 +0000 UTC m=+4154.556442136" Dec 05 08:04:19 crc kubenswrapper[4997]: I1205 08:04:19.773756 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nq8s6" Dec 05 08:04:19 crc kubenswrapper[4997]: I1205 08:04:19.774368 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nq8s6" Dec 05 08:04:19 crc kubenswrapper[4997]: I1205 08:04:19.860740 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nq8s6" Dec 05 08:04:20 crc kubenswrapper[4997]: I1205 08:04:20.108820 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nq8s6" Dec 05 08:04:23 crc kubenswrapper[4997]: I1205 08:04:23.543206 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nq8s6"] Dec 05 08:04:23 crc kubenswrapper[4997]: I1205 08:04:23.543989 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nq8s6" podUID="822d7260-aa6d-49c1-a8a6-dd4716f174e2" containerName="registry-server" containerID="cri-o://51ddd4ced978720284ce8b0f1624d3efcc5dc49c7afcc917d2a309a590e1a540" gracePeriod=2 Dec 05 08:04:24 crc kubenswrapper[4997]: I1205 08:04:24.098633 4997 generic.go:334] "Generic (PLEG): container finished" podID="822d7260-aa6d-49c1-a8a6-dd4716f174e2" containerID="51ddd4ced978720284ce8b0f1624d3efcc5dc49c7afcc917d2a309a590e1a540" exitCode=0 Dec 05 08:04:24 crc kubenswrapper[4997]: I1205 08:04:24.098682 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nq8s6" event={"ID":"822d7260-aa6d-49c1-a8a6-dd4716f174e2","Type":"ContainerDied","Data":"51ddd4ced978720284ce8b0f1624d3efcc5dc49c7afcc917d2a309a590e1a540"} Dec 05 08:04:24 crc kubenswrapper[4997]: I1205 08:04:24.525851 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nq8s6" Dec 05 08:04:24 crc kubenswrapper[4997]: I1205 08:04:24.704660 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/822d7260-aa6d-49c1-a8a6-dd4716f174e2-catalog-content\") pod \"822d7260-aa6d-49c1-a8a6-dd4716f174e2\" (UID: \"822d7260-aa6d-49c1-a8a6-dd4716f174e2\") " Dec 05 08:04:24 crc kubenswrapper[4997]: I1205 08:04:24.704741 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhl96\" (UniqueName: \"kubernetes.io/projected/822d7260-aa6d-49c1-a8a6-dd4716f174e2-kube-api-access-vhl96\") pod \"822d7260-aa6d-49c1-a8a6-dd4716f174e2\" (UID: \"822d7260-aa6d-49c1-a8a6-dd4716f174e2\") " Dec 05 08:04:24 crc kubenswrapper[4997]: I1205 08:04:24.704897 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/822d7260-aa6d-49c1-a8a6-dd4716f174e2-utilities\") pod \"822d7260-aa6d-49c1-a8a6-dd4716f174e2\" (UID: \"822d7260-aa6d-49c1-a8a6-dd4716f174e2\") " Dec 05 08:04:24 crc kubenswrapper[4997]: I1205 08:04:24.706878 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/822d7260-aa6d-49c1-a8a6-dd4716f174e2-utilities" (OuterVolumeSpecName: "utilities") pod "822d7260-aa6d-49c1-a8a6-dd4716f174e2" (UID: "822d7260-aa6d-49c1-a8a6-dd4716f174e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:04:24 crc kubenswrapper[4997]: I1205 08:04:24.714603 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/822d7260-aa6d-49c1-a8a6-dd4716f174e2-kube-api-access-vhl96" (OuterVolumeSpecName: "kube-api-access-vhl96") pod "822d7260-aa6d-49c1-a8a6-dd4716f174e2" (UID: "822d7260-aa6d-49c1-a8a6-dd4716f174e2"). InnerVolumeSpecName "kube-api-access-vhl96". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:04:24 crc kubenswrapper[4997]: I1205 08:04:24.731937 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/822d7260-aa6d-49c1-a8a6-dd4716f174e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "822d7260-aa6d-49c1-a8a6-dd4716f174e2" (UID: "822d7260-aa6d-49c1-a8a6-dd4716f174e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:04:24 crc kubenswrapper[4997]: I1205 08:04:24.806964 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/822d7260-aa6d-49c1-a8a6-dd4716f174e2-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:04:24 crc kubenswrapper[4997]: I1205 08:04:24.807002 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/822d7260-aa6d-49c1-a8a6-dd4716f174e2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:04:24 crc kubenswrapper[4997]: I1205 08:04:24.807019 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhl96\" (UniqueName: \"kubernetes.io/projected/822d7260-aa6d-49c1-a8a6-dd4716f174e2-kube-api-access-vhl96\") on node \"crc\" DevicePath \"\"" Dec 05 08:04:25 crc kubenswrapper[4997]: I1205 08:04:25.111121 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nq8s6" event={"ID":"822d7260-aa6d-49c1-a8a6-dd4716f174e2","Type":"ContainerDied","Data":"04fc9005d8e9b86f5598b5f178548ef215947fbce010d3470728de6c3e68e8fd"} Dec 05 08:04:25 crc kubenswrapper[4997]: I1205 08:04:25.111190 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nq8s6" Dec 05 08:04:25 crc kubenswrapper[4997]: I1205 08:04:25.111203 4997 scope.go:117] "RemoveContainer" containerID="51ddd4ced978720284ce8b0f1624d3efcc5dc49c7afcc917d2a309a590e1a540" Dec 05 08:04:25 crc kubenswrapper[4997]: I1205 08:04:25.141745 4997 scope.go:117] "RemoveContainer" containerID="67c6acd686c1e82f437d82e0ad4e2d6419336435d7fc66b5ee9aac90a1583991" Dec 05 08:04:25 crc kubenswrapper[4997]: I1205 08:04:25.152755 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nq8s6"] Dec 05 08:04:25 crc kubenswrapper[4997]: I1205 08:04:25.163072 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nq8s6"] Dec 05 08:04:25 crc kubenswrapper[4997]: I1205 08:04:25.174293 4997 scope.go:117] "RemoveContainer" containerID="119c07f65354402c18ba18b9a29066c0901e3d8bde9d1ab5f55ff536bbf08db3" Dec 05 08:04:25 crc kubenswrapper[4997]: I1205 08:04:25.766212 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="822d7260-aa6d-49c1-a8a6-dd4716f174e2" path="/var/lib/kubelet/pods/822d7260-aa6d-49c1-a8a6-dd4716f174e2/volumes" Dec 05 08:05:35 crc kubenswrapper[4997]: I1205 08:05:35.531754 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m6ffp"] Dec 05 08:05:35 crc kubenswrapper[4997]: E1205 08:05:35.533046 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="822d7260-aa6d-49c1-a8a6-dd4716f174e2" containerName="extract-utilities" Dec 05 08:05:35 crc kubenswrapper[4997]: I1205 08:05:35.533090 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="822d7260-aa6d-49c1-a8a6-dd4716f174e2" containerName="extract-utilities" Dec 05 08:05:35 crc kubenswrapper[4997]: E1205 08:05:35.533163 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="822d7260-aa6d-49c1-a8a6-dd4716f174e2" containerName="registry-server" Dec 05 08:05:35 crc kubenswrapper[4997]: I1205 08:05:35.533180 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="822d7260-aa6d-49c1-a8a6-dd4716f174e2" containerName="registry-server" Dec 05 08:05:35 crc kubenswrapper[4997]: E1205 08:05:35.533217 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="822d7260-aa6d-49c1-a8a6-dd4716f174e2" containerName="extract-content" Dec 05 08:05:35 crc kubenswrapper[4997]: I1205 08:05:35.533235 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="822d7260-aa6d-49c1-a8a6-dd4716f174e2" containerName="extract-content" Dec 05 08:05:35 crc kubenswrapper[4997]: I1205 08:05:35.533589 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="822d7260-aa6d-49c1-a8a6-dd4716f174e2" containerName="registry-server" Dec 05 08:05:35 crc kubenswrapper[4997]: I1205 08:05:35.538352 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m6ffp" Dec 05 08:05:35 crc kubenswrapper[4997]: I1205 08:05:35.549494 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m6ffp"] Dec 05 08:05:35 crc kubenswrapper[4997]: I1205 08:05:35.663949 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dsml\" (UniqueName: \"kubernetes.io/projected/84626aa9-20fb-4679-89d1-f400faa7a236-kube-api-access-9dsml\") pod \"certified-operators-m6ffp\" (UID: \"84626aa9-20fb-4679-89d1-f400faa7a236\") " pod="openshift-marketplace/certified-operators-m6ffp" Dec 05 08:05:35 crc kubenswrapper[4997]: I1205 08:05:35.664044 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84626aa9-20fb-4679-89d1-f400faa7a236-utilities\") pod \"certified-operators-m6ffp\" (UID: \"84626aa9-20fb-4679-89d1-f400faa7a236\") " pod="openshift-marketplace/certified-operators-m6ffp" Dec 05 08:05:35 crc kubenswrapper[4997]: I1205 08:05:35.664133 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84626aa9-20fb-4679-89d1-f400faa7a236-catalog-content\") pod \"certified-operators-m6ffp\" (UID: \"84626aa9-20fb-4679-89d1-f400faa7a236\") " pod="openshift-marketplace/certified-operators-m6ffp" Dec 05 08:05:35 crc kubenswrapper[4997]: I1205 08:05:35.765797 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84626aa9-20fb-4679-89d1-f400faa7a236-catalog-content\") pod \"certified-operators-m6ffp\" (UID: \"84626aa9-20fb-4679-89d1-f400faa7a236\") " pod="openshift-marketplace/certified-operators-m6ffp" Dec 05 08:05:35 crc kubenswrapper[4997]: I1205 08:05:35.765895 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dsml\" (UniqueName: \"kubernetes.io/projected/84626aa9-20fb-4679-89d1-f400faa7a236-kube-api-access-9dsml\") pod \"certified-operators-m6ffp\" (UID: \"84626aa9-20fb-4679-89d1-f400faa7a236\") " pod="openshift-marketplace/certified-operators-m6ffp" Dec 05 08:05:35 crc kubenswrapper[4997]: I1205 08:05:35.765940 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84626aa9-20fb-4679-89d1-f400faa7a236-utilities\") pod \"certified-operators-m6ffp\" (UID: \"84626aa9-20fb-4679-89d1-f400faa7a236\") " pod="openshift-marketplace/certified-operators-m6ffp" Dec 05 08:05:35 crc kubenswrapper[4997]: I1205 08:05:35.766416 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84626aa9-20fb-4679-89d1-f400faa7a236-catalog-content\") pod \"certified-operators-m6ffp\" (UID: \"84626aa9-20fb-4679-89d1-f400faa7a236\") " pod="openshift-marketplace/certified-operators-m6ffp" Dec 05 08:05:35 crc kubenswrapper[4997]: I1205 08:05:35.766448 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84626aa9-20fb-4679-89d1-f400faa7a236-utilities\") pod \"certified-operators-m6ffp\" (UID: \"84626aa9-20fb-4679-89d1-f400faa7a236\") " pod="openshift-marketplace/certified-operators-m6ffp" Dec 05 08:05:35 crc kubenswrapper[4997]: I1205 08:05:35.786900 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dsml\" (UniqueName: \"kubernetes.io/projected/84626aa9-20fb-4679-89d1-f400faa7a236-kube-api-access-9dsml\") pod \"certified-operators-m6ffp\" (UID: \"84626aa9-20fb-4679-89d1-f400faa7a236\") " pod="openshift-marketplace/certified-operators-m6ffp" Dec 05 08:05:35 crc kubenswrapper[4997]: I1205 08:05:35.876787 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m6ffp" Dec 05 08:05:36 crc kubenswrapper[4997]: I1205 08:05:36.192015 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m6ffp"] Dec 05 08:05:36 crc kubenswrapper[4997]: I1205 08:05:36.840988 4997 generic.go:334] "Generic (PLEG): container finished" podID="84626aa9-20fb-4679-89d1-f400faa7a236" containerID="eb8fb0ab8300cec9062fae01b1b99bba98338536526d711744657f1fa7ca4dac" exitCode=0 Dec 05 08:05:36 crc kubenswrapper[4997]: I1205 08:05:36.841306 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6ffp" event={"ID":"84626aa9-20fb-4679-89d1-f400faa7a236","Type":"ContainerDied","Data":"eb8fb0ab8300cec9062fae01b1b99bba98338536526d711744657f1fa7ca4dac"} Dec 05 08:05:36 crc kubenswrapper[4997]: I1205 08:05:36.841342 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6ffp" event={"ID":"84626aa9-20fb-4679-89d1-f400faa7a236","Type":"ContainerStarted","Data":"fa93d2bcb6293eed9b6469671687eb84e109e890e60adf3e65b85884809c3e8c"} Dec 05 08:05:37 crc kubenswrapper[4997]: I1205 08:05:37.852219 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6ffp" event={"ID":"84626aa9-20fb-4679-89d1-f400faa7a236","Type":"ContainerStarted","Data":"f2f0cb18808fd47bd90cf781f11d739266e90426b5a711ab9659f9b803f42473"} Dec 05 08:05:38 crc kubenswrapper[4997]: I1205 08:05:38.865186 4997 generic.go:334] "Generic (PLEG): container finished" podID="84626aa9-20fb-4679-89d1-f400faa7a236" containerID="f2f0cb18808fd47bd90cf781f11d739266e90426b5a711ab9659f9b803f42473" exitCode=0 Dec 05 08:05:38 crc kubenswrapper[4997]: I1205 08:05:38.865542 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6ffp" event={"ID":"84626aa9-20fb-4679-89d1-f400faa7a236","Type":"ContainerDied","Data":"f2f0cb18808fd47bd90cf781f11d739266e90426b5a711ab9659f9b803f42473"} Dec 05 08:05:39 crc kubenswrapper[4997]: I1205 08:05:39.881044 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6ffp" event={"ID":"84626aa9-20fb-4679-89d1-f400faa7a236","Type":"ContainerStarted","Data":"5d678af20ab53d253d966a074f430d698bddc8d7340387f7cf3796ab5395cc74"} Dec 05 08:05:39 crc kubenswrapper[4997]: I1205 08:05:39.907890 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m6ffp" podStartSLOduration=2.370465313 podStartE2EDuration="4.907865144s" podCreationTimestamp="2025-12-05 08:05:35 +0000 UTC" firstStartedPulling="2025-12-05 08:05:36.842878427 +0000 UTC m=+4237.371785708" lastFinishedPulling="2025-12-05 08:05:39.380278258 +0000 UTC m=+4239.909185539" observedRunningTime="2025-12-05 08:05:39.903906386 +0000 UTC m=+4240.432813737" watchObservedRunningTime="2025-12-05 08:05:39.907865144 +0000 UTC m=+4240.436772445" Dec 05 08:05:45 crc kubenswrapper[4997]: I1205 08:05:45.876973 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m6ffp" Dec 05 08:05:45 crc kubenswrapper[4997]: I1205 08:05:45.877592 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m6ffp" Dec 05 08:05:45 crc kubenswrapper[4997]: I1205 08:05:45.932529 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m6ffp" Dec 05 08:05:46 crc kubenswrapper[4997]: I1205 08:05:46.005068 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m6ffp" Dec 05 08:05:46 crc kubenswrapper[4997]: I1205 08:05:46.501733 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m6ffp"] Dec 05 08:05:47 crc kubenswrapper[4997]: I1205 08:05:47.944923 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m6ffp" podUID="84626aa9-20fb-4679-89d1-f400faa7a236" containerName="registry-server" containerID="cri-o://5d678af20ab53d253d966a074f430d698bddc8d7340387f7cf3796ab5395cc74" gracePeriod=2 Dec 05 08:05:48 crc kubenswrapper[4997]: I1205 08:05:48.836495 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m6ffp" Dec 05 08:05:48 crc kubenswrapper[4997]: I1205 08:05:48.954725 4997 generic.go:334] "Generic (PLEG): container finished" podID="84626aa9-20fb-4679-89d1-f400faa7a236" containerID="5d678af20ab53d253d966a074f430d698bddc8d7340387f7cf3796ab5395cc74" exitCode=0 Dec 05 08:05:48 crc kubenswrapper[4997]: I1205 08:05:48.954790 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6ffp" event={"ID":"84626aa9-20fb-4679-89d1-f400faa7a236","Type":"ContainerDied","Data":"5d678af20ab53d253d966a074f430d698bddc8d7340387f7cf3796ab5395cc74"} Dec 05 08:05:48 crc kubenswrapper[4997]: I1205 08:05:48.955153 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6ffp" event={"ID":"84626aa9-20fb-4679-89d1-f400faa7a236","Type":"ContainerDied","Data":"fa93d2bcb6293eed9b6469671687eb84e109e890e60adf3e65b85884809c3e8c"} Dec 05 08:05:48 crc kubenswrapper[4997]: I1205 08:05:48.955186 4997 scope.go:117] "RemoveContainer" containerID="5d678af20ab53d253d966a074f430d698bddc8d7340387f7cf3796ab5395cc74" Dec 05 08:05:48 crc kubenswrapper[4997]: I1205 08:05:48.954842 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m6ffp" Dec 05 08:05:48 crc kubenswrapper[4997]: I1205 08:05:48.968984 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dsml\" (UniqueName: \"kubernetes.io/projected/84626aa9-20fb-4679-89d1-f400faa7a236-kube-api-access-9dsml\") pod \"84626aa9-20fb-4679-89d1-f400faa7a236\" (UID: \"84626aa9-20fb-4679-89d1-f400faa7a236\") " Dec 05 08:05:48 crc kubenswrapper[4997]: I1205 08:05:48.969138 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84626aa9-20fb-4679-89d1-f400faa7a236-catalog-content\") pod \"84626aa9-20fb-4679-89d1-f400faa7a236\" (UID: \"84626aa9-20fb-4679-89d1-f400faa7a236\") " Dec 05 08:05:48 crc kubenswrapper[4997]: I1205 08:05:48.969193 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84626aa9-20fb-4679-89d1-f400faa7a236-utilities\") pod \"84626aa9-20fb-4679-89d1-f400faa7a236\" (UID: \"84626aa9-20fb-4679-89d1-f400faa7a236\") " Dec 05 08:05:48 crc kubenswrapper[4997]: I1205 08:05:48.970109 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84626aa9-20fb-4679-89d1-f400faa7a236-utilities" (OuterVolumeSpecName: "utilities") pod "84626aa9-20fb-4679-89d1-f400faa7a236" (UID: "84626aa9-20fb-4679-89d1-f400faa7a236"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:05:48 crc kubenswrapper[4997]: I1205 08:05:48.975977 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84626aa9-20fb-4679-89d1-f400faa7a236-kube-api-access-9dsml" (OuterVolumeSpecName: "kube-api-access-9dsml") pod "84626aa9-20fb-4679-89d1-f400faa7a236" (UID: "84626aa9-20fb-4679-89d1-f400faa7a236"). InnerVolumeSpecName "kube-api-access-9dsml". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:05:48 crc kubenswrapper[4997]: I1205 08:05:48.978761 4997 scope.go:117] "RemoveContainer" containerID="f2f0cb18808fd47bd90cf781f11d739266e90426b5a711ab9659f9b803f42473" Dec 05 08:05:49 crc kubenswrapper[4997]: I1205 08:05:49.032550 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84626aa9-20fb-4679-89d1-f400faa7a236-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84626aa9-20fb-4679-89d1-f400faa7a236" (UID: "84626aa9-20fb-4679-89d1-f400faa7a236"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:05:49 crc kubenswrapper[4997]: I1205 08:05:49.070941 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84626aa9-20fb-4679-89d1-f400faa7a236-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:05:49 crc kubenswrapper[4997]: I1205 08:05:49.071010 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dsml\" (UniqueName: \"kubernetes.io/projected/84626aa9-20fb-4679-89d1-f400faa7a236-kube-api-access-9dsml\") on node \"crc\" DevicePath \"\"" Dec 05 08:05:49 crc kubenswrapper[4997]: I1205 08:05:49.071027 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84626aa9-20fb-4679-89d1-f400faa7a236-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:05:49 crc kubenswrapper[4997]: I1205 08:05:49.290033 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m6ffp"] Dec 05 08:05:49 crc kubenswrapper[4997]: I1205 08:05:49.297963 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m6ffp"] Dec 05 08:05:49 crc kubenswrapper[4997]: I1205 08:05:49.413570 4997 scope.go:117] "RemoveContainer" containerID="eb8fb0ab8300cec9062fae01b1b99bba98338536526d711744657f1fa7ca4dac" Dec 05 08:05:49 crc kubenswrapper[4997]: I1205 08:05:49.466189 4997 scope.go:117] "RemoveContainer" containerID="5d678af20ab53d253d966a074f430d698bddc8d7340387f7cf3796ab5395cc74" Dec 05 08:05:49 crc kubenswrapper[4997]: E1205 08:05:49.466812 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d678af20ab53d253d966a074f430d698bddc8d7340387f7cf3796ab5395cc74\": container with ID starting with 5d678af20ab53d253d966a074f430d698bddc8d7340387f7cf3796ab5395cc74 not found: ID does not exist" containerID="5d678af20ab53d253d966a074f430d698bddc8d7340387f7cf3796ab5395cc74" Dec 05 08:05:49 crc kubenswrapper[4997]: I1205 08:05:49.466865 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d678af20ab53d253d966a074f430d698bddc8d7340387f7cf3796ab5395cc74"} err="failed to get container status \"5d678af20ab53d253d966a074f430d698bddc8d7340387f7cf3796ab5395cc74\": rpc error: code = NotFound desc = could not find container \"5d678af20ab53d253d966a074f430d698bddc8d7340387f7cf3796ab5395cc74\": container with ID starting with 5d678af20ab53d253d966a074f430d698bddc8d7340387f7cf3796ab5395cc74 not found: ID does not exist" Dec 05 08:05:49 crc kubenswrapper[4997]: I1205 08:05:49.466899 4997 scope.go:117] "RemoveContainer" containerID="f2f0cb18808fd47bd90cf781f11d739266e90426b5a711ab9659f9b803f42473" Dec 05 08:05:49 crc kubenswrapper[4997]: E1205 08:05:49.467356 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2f0cb18808fd47bd90cf781f11d739266e90426b5a711ab9659f9b803f42473\": container with ID starting with f2f0cb18808fd47bd90cf781f11d739266e90426b5a711ab9659f9b803f42473 not found: ID does not exist" containerID="f2f0cb18808fd47bd90cf781f11d739266e90426b5a711ab9659f9b803f42473" Dec 05 08:05:49 crc kubenswrapper[4997]: I1205 08:05:49.467410 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2f0cb18808fd47bd90cf781f11d739266e90426b5a711ab9659f9b803f42473"} err="failed to get container status \"f2f0cb18808fd47bd90cf781f11d739266e90426b5a711ab9659f9b803f42473\": rpc error: code = NotFound desc = could not find container \"f2f0cb18808fd47bd90cf781f11d739266e90426b5a711ab9659f9b803f42473\": container with ID starting with f2f0cb18808fd47bd90cf781f11d739266e90426b5a711ab9659f9b803f42473 not found: ID does not exist" Dec 05 08:05:49 crc kubenswrapper[4997]: I1205 08:05:49.467445 4997 scope.go:117] "RemoveContainer" containerID="eb8fb0ab8300cec9062fae01b1b99bba98338536526d711744657f1fa7ca4dac" Dec 05 08:05:49 crc kubenswrapper[4997]: E1205 08:05:49.468027 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb8fb0ab8300cec9062fae01b1b99bba98338536526d711744657f1fa7ca4dac\": container with ID starting with eb8fb0ab8300cec9062fae01b1b99bba98338536526d711744657f1fa7ca4dac not found: ID does not exist" containerID="eb8fb0ab8300cec9062fae01b1b99bba98338536526d711744657f1fa7ca4dac" Dec 05 08:05:49 crc kubenswrapper[4997]: I1205 08:05:49.468048 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb8fb0ab8300cec9062fae01b1b99bba98338536526d711744657f1fa7ca4dac"} err="failed to get container status \"eb8fb0ab8300cec9062fae01b1b99bba98338536526d711744657f1fa7ca4dac\": rpc error: code = NotFound desc = could not find container \"eb8fb0ab8300cec9062fae01b1b99bba98338536526d711744657f1fa7ca4dac\": container with ID starting with eb8fb0ab8300cec9062fae01b1b99bba98338536526d711744657f1fa7ca4dac not found: ID does not exist" Dec 05 08:05:49 crc kubenswrapper[4997]: I1205 08:05:49.768349 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84626aa9-20fb-4679-89d1-f400faa7a236" path="/var/lib/kubelet/pods/84626aa9-20fb-4679-89d1-f400faa7a236/volumes" Dec 05 08:06:19 crc kubenswrapper[4997]: I1205 08:06:19.770271 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:06:19 crc kubenswrapper[4997]: I1205 08:06:19.770728 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:06:49 crc kubenswrapper[4997]: I1205 08:06:49.769724 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:06:49 crc kubenswrapper[4997]: I1205 08:06:49.770472 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:07:19 crc kubenswrapper[4997]: I1205 08:07:19.770567 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:07:19 crc kubenswrapper[4997]: I1205 08:07:19.771330 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:07:19 crc kubenswrapper[4997]: I1205 08:07:19.771419 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 08:07:19 crc kubenswrapper[4997]: I1205 08:07:19.772639 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e65f871a5d8f1b6b491a6cfe1489bad2714fbbf443695e3bcd99bcb1a0d11097"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:07:19 crc kubenswrapper[4997]: I1205 08:07:19.772745 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://e65f871a5d8f1b6b491a6cfe1489bad2714fbbf443695e3bcd99bcb1a0d11097" gracePeriod=600 Dec 05 08:07:20 crc kubenswrapper[4997]: I1205 08:07:20.769431 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="e65f871a5d8f1b6b491a6cfe1489bad2714fbbf443695e3bcd99bcb1a0d11097" exitCode=0 Dec 05 08:07:20 crc kubenswrapper[4997]: I1205 08:07:20.769479 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"e65f871a5d8f1b6b491a6cfe1489bad2714fbbf443695e3bcd99bcb1a0d11097"} Dec 05 08:07:20 crc kubenswrapper[4997]: I1205 08:07:20.770161 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a"} Dec 05 08:07:20 crc kubenswrapper[4997]: I1205 08:07:20.770184 4997 scope.go:117] "RemoveContainer" containerID="70d5815bd0d29f85237165bc278582daffd9b473f7c6f09a3afdafaedebfa571" Dec 05 08:09:14 crc kubenswrapper[4997]: I1205 08:09:14.968227 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d5rwp"] Dec 05 08:09:14 crc kubenswrapper[4997]: E1205 08:09:14.969102 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84626aa9-20fb-4679-89d1-f400faa7a236" containerName="extract-utilities" Dec 05 08:09:14 crc kubenswrapper[4997]: I1205 08:09:14.969128 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="84626aa9-20fb-4679-89d1-f400faa7a236" containerName="extract-utilities" Dec 05 08:09:14 crc kubenswrapper[4997]: E1205 08:09:14.969148 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84626aa9-20fb-4679-89d1-f400faa7a236" containerName="registry-server" Dec 05 08:09:14 crc kubenswrapper[4997]: I1205 08:09:14.969157 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="84626aa9-20fb-4679-89d1-f400faa7a236" containerName="registry-server" Dec 05 08:09:14 crc kubenswrapper[4997]: E1205 08:09:14.969171 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84626aa9-20fb-4679-89d1-f400faa7a236" containerName="extract-content" Dec 05 08:09:14 crc kubenswrapper[4997]: I1205 08:09:14.969180 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="84626aa9-20fb-4679-89d1-f400faa7a236" containerName="extract-content" Dec 05 08:09:14 crc kubenswrapper[4997]: I1205 08:09:14.969359 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="84626aa9-20fb-4679-89d1-f400faa7a236" containerName="registry-server" Dec 05 08:09:14 crc kubenswrapper[4997]: I1205 08:09:14.970600 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d5rwp" Dec 05 08:09:14 crc kubenswrapper[4997]: I1205 08:09:14.977223 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d5rwp"] Dec 05 08:09:15 crc kubenswrapper[4997]: I1205 08:09:15.054743 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37f162c9-ecf4-489c-8f20-fe32ffdff906-catalog-content\") pod \"redhat-operators-d5rwp\" (UID: \"37f162c9-ecf4-489c-8f20-fe32ffdff906\") " pod="openshift-marketplace/redhat-operators-d5rwp" Dec 05 08:09:15 crc kubenswrapper[4997]: I1205 08:09:15.055052 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37f162c9-ecf4-489c-8f20-fe32ffdff906-utilities\") pod \"redhat-operators-d5rwp\" (UID: \"37f162c9-ecf4-489c-8f20-fe32ffdff906\") " pod="openshift-marketplace/redhat-operators-d5rwp" Dec 05 08:09:15 crc kubenswrapper[4997]: I1205 08:09:15.055134 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwmxt\" (UniqueName: \"kubernetes.io/projected/37f162c9-ecf4-489c-8f20-fe32ffdff906-kube-api-access-cwmxt\") pod \"redhat-operators-d5rwp\" (UID: \"37f162c9-ecf4-489c-8f20-fe32ffdff906\") " pod="openshift-marketplace/redhat-operators-d5rwp" Dec 05 08:09:15 crc kubenswrapper[4997]: I1205 08:09:15.156193 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37f162c9-ecf4-489c-8f20-fe32ffdff906-catalog-content\") pod \"redhat-operators-d5rwp\" (UID: \"37f162c9-ecf4-489c-8f20-fe32ffdff906\") " pod="openshift-marketplace/redhat-operators-d5rwp" Dec 05 08:09:15 crc kubenswrapper[4997]: I1205 08:09:15.156299 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37f162c9-ecf4-489c-8f20-fe32ffdff906-utilities\") pod \"redhat-operators-d5rwp\" (UID: \"37f162c9-ecf4-489c-8f20-fe32ffdff906\") " pod="openshift-marketplace/redhat-operators-d5rwp" Dec 05 08:09:15 crc kubenswrapper[4997]: I1205 08:09:15.156358 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwmxt\" (UniqueName: \"kubernetes.io/projected/37f162c9-ecf4-489c-8f20-fe32ffdff906-kube-api-access-cwmxt\") pod \"redhat-operators-d5rwp\" (UID: \"37f162c9-ecf4-489c-8f20-fe32ffdff906\") " pod="openshift-marketplace/redhat-operators-d5rwp" Dec 05 08:09:15 crc kubenswrapper[4997]: I1205 08:09:15.156813 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37f162c9-ecf4-489c-8f20-fe32ffdff906-catalog-content\") pod \"redhat-operators-d5rwp\" (UID: \"37f162c9-ecf4-489c-8f20-fe32ffdff906\") " pod="openshift-marketplace/redhat-operators-d5rwp" Dec 05 08:09:15 crc kubenswrapper[4997]: I1205 08:09:15.156902 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37f162c9-ecf4-489c-8f20-fe32ffdff906-utilities\") pod \"redhat-operators-d5rwp\" (UID: \"37f162c9-ecf4-489c-8f20-fe32ffdff906\") " pod="openshift-marketplace/redhat-operators-d5rwp" Dec 05 08:09:15 crc kubenswrapper[4997]: I1205 08:09:15.178416 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwmxt\" (UniqueName: \"kubernetes.io/projected/37f162c9-ecf4-489c-8f20-fe32ffdff906-kube-api-access-cwmxt\") pod \"redhat-operators-d5rwp\" (UID: \"37f162c9-ecf4-489c-8f20-fe32ffdff906\") " pod="openshift-marketplace/redhat-operators-d5rwp" Dec 05 08:09:15 crc kubenswrapper[4997]: I1205 08:09:15.292888 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d5rwp" Dec 05 08:09:15 crc kubenswrapper[4997]: I1205 08:09:15.514529 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d5rwp"] Dec 05 08:09:15 crc kubenswrapper[4997]: I1205 08:09:15.827969 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d5rwp" event={"ID":"37f162c9-ecf4-489c-8f20-fe32ffdff906","Type":"ContainerStarted","Data":"b7b78e37e5315029a75c33ff894615cec4dd774689d63de8c30d38be36ac72c3"} Dec 05 08:09:16 crc kubenswrapper[4997]: I1205 08:09:16.839021 4997 generic.go:334] "Generic (PLEG): container finished" podID="37f162c9-ecf4-489c-8f20-fe32ffdff906" containerID="24f313bc12a33950bc34659e4d80ae4d34a564a2fa1eba665b0cc68edc386bc1" exitCode=0 Dec 05 08:09:16 crc kubenswrapper[4997]: I1205 08:09:16.839082 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d5rwp" event={"ID":"37f162c9-ecf4-489c-8f20-fe32ffdff906","Type":"ContainerDied","Data":"24f313bc12a33950bc34659e4d80ae4d34a564a2fa1eba665b0cc68edc386bc1"} Dec 05 08:09:16 crc kubenswrapper[4997]: I1205 08:09:16.841542 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 08:09:17 crc kubenswrapper[4997]: I1205 08:09:17.849220 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d5rwp" event={"ID":"37f162c9-ecf4-489c-8f20-fe32ffdff906","Type":"ContainerStarted","Data":"409593ce7b9f757911c9dcb11e1cd2378584f7231b062a74f6038bb293641b86"} Dec 05 08:09:18 crc kubenswrapper[4997]: I1205 08:09:18.864850 4997 generic.go:334] "Generic (PLEG): container finished" podID="37f162c9-ecf4-489c-8f20-fe32ffdff906" containerID="409593ce7b9f757911c9dcb11e1cd2378584f7231b062a74f6038bb293641b86" exitCode=0 Dec 05 08:09:18 crc kubenswrapper[4997]: I1205 08:09:18.865311 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d5rwp" event={"ID":"37f162c9-ecf4-489c-8f20-fe32ffdff906","Type":"ContainerDied","Data":"409593ce7b9f757911c9dcb11e1cd2378584f7231b062a74f6038bb293641b86"} Dec 05 08:09:19 crc kubenswrapper[4997]: I1205 08:09:19.881515 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d5rwp" event={"ID":"37f162c9-ecf4-489c-8f20-fe32ffdff906","Type":"ContainerStarted","Data":"16d77564c1d411cbbf77d3f58307a6a1efd44c74703500244c9a856265336a8d"} Dec 05 08:09:19 crc kubenswrapper[4997]: I1205 08:09:19.910543 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d5rwp" podStartSLOduration=3.407242987 podStartE2EDuration="5.910520487s" podCreationTimestamp="2025-12-05 08:09:14 +0000 UTC" firstStartedPulling="2025-12-05 08:09:16.841021891 +0000 UTC m=+4457.369929192" lastFinishedPulling="2025-12-05 08:09:19.344299421 +0000 UTC m=+4459.873206692" observedRunningTime="2025-12-05 08:09:19.899987883 +0000 UTC m=+4460.428895144" watchObservedRunningTime="2025-12-05 08:09:19.910520487 +0000 UTC m=+4460.439427758" Dec 05 08:09:25 crc kubenswrapper[4997]: I1205 08:09:25.293237 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d5rwp" Dec 05 08:09:25 crc kubenswrapper[4997]: I1205 08:09:25.293756 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d5rwp" Dec 05 08:09:26 crc kubenswrapper[4997]: I1205 08:09:26.344352 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d5rwp" podUID="37f162c9-ecf4-489c-8f20-fe32ffdff906" containerName="registry-server" probeResult="failure" output=< Dec 05 08:09:26 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 08:09:26 crc kubenswrapper[4997]: > Dec 05 08:09:35 crc kubenswrapper[4997]: I1205 08:09:35.346576 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d5rwp" Dec 05 08:09:35 crc kubenswrapper[4997]: I1205 08:09:35.412438 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d5rwp" Dec 05 08:09:35 crc kubenswrapper[4997]: I1205 08:09:35.591210 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d5rwp"] Dec 05 08:09:37 crc kubenswrapper[4997]: I1205 08:09:37.034763 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d5rwp" podUID="37f162c9-ecf4-489c-8f20-fe32ffdff906" containerName="registry-server" containerID="cri-o://16d77564c1d411cbbf77d3f58307a6a1efd44c74703500244c9a856265336a8d" gracePeriod=2 Dec 05 08:09:37 crc kubenswrapper[4997]: I1205 08:09:37.446951 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d5rwp" Dec 05 08:09:37 crc kubenswrapper[4997]: I1205 08:09:37.618463 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37f162c9-ecf4-489c-8f20-fe32ffdff906-catalog-content\") pod \"37f162c9-ecf4-489c-8f20-fe32ffdff906\" (UID: \"37f162c9-ecf4-489c-8f20-fe32ffdff906\") " Dec 05 08:09:37 crc kubenswrapper[4997]: I1205 08:09:37.618600 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwmxt\" (UniqueName: \"kubernetes.io/projected/37f162c9-ecf4-489c-8f20-fe32ffdff906-kube-api-access-cwmxt\") pod \"37f162c9-ecf4-489c-8f20-fe32ffdff906\" (UID: \"37f162c9-ecf4-489c-8f20-fe32ffdff906\") " Dec 05 08:09:37 crc kubenswrapper[4997]: I1205 08:09:37.618770 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37f162c9-ecf4-489c-8f20-fe32ffdff906-utilities\") pod \"37f162c9-ecf4-489c-8f20-fe32ffdff906\" (UID: \"37f162c9-ecf4-489c-8f20-fe32ffdff906\") " Dec 05 08:09:37 crc kubenswrapper[4997]: I1205 08:09:37.620474 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37f162c9-ecf4-489c-8f20-fe32ffdff906-utilities" (OuterVolumeSpecName: "utilities") pod "37f162c9-ecf4-489c-8f20-fe32ffdff906" (UID: "37f162c9-ecf4-489c-8f20-fe32ffdff906"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:09:37 crc kubenswrapper[4997]: I1205 08:09:37.635215 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37f162c9-ecf4-489c-8f20-fe32ffdff906-kube-api-access-cwmxt" (OuterVolumeSpecName: "kube-api-access-cwmxt") pod "37f162c9-ecf4-489c-8f20-fe32ffdff906" (UID: "37f162c9-ecf4-489c-8f20-fe32ffdff906"). InnerVolumeSpecName "kube-api-access-cwmxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:09:37 crc kubenswrapper[4997]: I1205 08:09:37.720704 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwmxt\" (UniqueName: \"kubernetes.io/projected/37f162c9-ecf4-489c-8f20-fe32ffdff906-kube-api-access-cwmxt\") on node \"crc\" DevicePath \"\"" Dec 05 08:09:37 crc kubenswrapper[4997]: I1205 08:09:37.720757 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37f162c9-ecf4-489c-8f20-fe32ffdff906-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:09:37 crc kubenswrapper[4997]: I1205 08:09:37.756272 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37f162c9-ecf4-489c-8f20-fe32ffdff906-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37f162c9-ecf4-489c-8f20-fe32ffdff906" (UID: "37f162c9-ecf4-489c-8f20-fe32ffdff906"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:09:37 crc kubenswrapper[4997]: I1205 08:09:37.822494 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37f162c9-ecf4-489c-8f20-fe32ffdff906-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:09:38 crc kubenswrapper[4997]: I1205 08:09:38.046575 4997 generic.go:334] "Generic (PLEG): container finished" podID="37f162c9-ecf4-489c-8f20-fe32ffdff906" containerID="16d77564c1d411cbbf77d3f58307a6a1efd44c74703500244c9a856265336a8d" exitCode=0 Dec 05 08:09:38 crc kubenswrapper[4997]: I1205 08:09:38.046693 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d5rwp" Dec 05 08:09:38 crc kubenswrapper[4997]: I1205 08:09:38.046689 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d5rwp" event={"ID":"37f162c9-ecf4-489c-8f20-fe32ffdff906","Type":"ContainerDied","Data":"16d77564c1d411cbbf77d3f58307a6a1efd44c74703500244c9a856265336a8d"} Dec 05 08:09:38 crc kubenswrapper[4997]: I1205 08:09:38.047328 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d5rwp" event={"ID":"37f162c9-ecf4-489c-8f20-fe32ffdff906","Type":"ContainerDied","Data":"b7b78e37e5315029a75c33ff894615cec4dd774689d63de8c30d38be36ac72c3"} Dec 05 08:09:38 crc kubenswrapper[4997]: I1205 08:09:38.047370 4997 scope.go:117] "RemoveContainer" containerID="16d77564c1d411cbbf77d3f58307a6a1efd44c74703500244c9a856265336a8d" Dec 05 08:09:38 crc kubenswrapper[4997]: I1205 08:09:38.077697 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d5rwp"] Dec 05 08:09:38 crc kubenswrapper[4997]: I1205 08:09:38.081850 4997 scope.go:117] "RemoveContainer" containerID="409593ce7b9f757911c9dcb11e1cd2378584f7231b062a74f6038bb293641b86" Dec 05 08:09:38 crc kubenswrapper[4997]: I1205 08:09:38.084012 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d5rwp"] Dec 05 08:09:38 crc kubenswrapper[4997]: I1205 08:09:38.617253 4997 scope.go:117] "RemoveContainer" containerID="24f313bc12a33950bc34659e4d80ae4d34a564a2fa1eba665b0cc68edc386bc1" Dec 05 08:09:38 crc kubenswrapper[4997]: I1205 08:09:38.650215 4997 scope.go:117] "RemoveContainer" containerID="16d77564c1d411cbbf77d3f58307a6a1efd44c74703500244c9a856265336a8d" Dec 05 08:09:38 crc kubenswrapper[4997]: E1205 08:09:38.650815 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16d77564c1d411cbbf77d3f58307a6a1efd44c74703500244c9a856265336a8d\": container with ID starting with 16d77564c1d411cbbf77d3f58307a6a1efd44c74703500244c9a856265336a8d not found: ID does not exist" containerID="16d77564c1d411cbbf77d3f58307a6a1efd44c74703500244c9a856265336a8d" Dec 05 08:09:38 crc kubenswrapper[4997]: I1205 08:09:38.650870 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16d77564c1d411cbbf77d3f58307a6a1efd44c74703500244c9a856265336a8d"} err="failed to get container status \"16d77564c1d411cbbf77d3f58307a6a1efd44c74703500244c9a856265336a8d\": rpc error: code = NotFound desc = could not find container \"16d77564c1d411cbbf77d3f58307a6a1efd44c74703500244c9a856265336a8d\": container with ID starting with 16d77564c1d411cbbf77d3f58307a6a1efd44c74703500244c9a856265336a8d not found: ID does not exist" Dec 05 08:09:38 crc kubenswrapper[4997]: I1205 08:09:38.650906 4997 scope.go:117] "RemoveContainer" containerID="409593ce7b9f757911c9dcb11e1cd2378584f7231b062a74f6038bb293641b86" Dec 05 08:09:38 crc kubenswrapper[4997]: E1205 08:09:38.651852 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"409593ce7b9f757911c9dcb11e1cd2378584f7231b062a74f6038bb293641b86\": container with ID starting with 409593ce7b9f757911c9dcb11e1cd2378584f7231b062a74f6038bb293641b86 not found: ID does not exist" containerID="409593ce7b9f757911c9dcb11e1cd2378584f7231b062a74f6038bb293641b86" Dec 05 08:09:38 crc kubenswrapper[4997]: I1205 08:09:38.651952 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"409593ce7b9f757911c9dcb11e1cd2378584f7231b062a74f6038bb293641b86"} err="failed to get container status \"409593ce7b9f757911c9dcb11e1cd2378584f7231b062a74f6038bb293641b86\": rpc error: code = NotFound desc = could not find container \"409593ce7b9f757911c9dcb11e1cd2378584f7231b062a74f6038bb293641b86\": container with ID starting with 409593ce7b9f757911c9dcb11e1cd2378584f7231b062a74f6038bb293641b86 not found: ID does not exist" Dec 05 08:09:38 crc kubenswrapper[4997]: I1205 08:09:38.652026 4997 scope.go:117] "RemoveContainer" containerID="24f313bc12a33950bc34659e4d80ae4d34a564a2fa1eba665b0cc68edc386bc1" Dec 05 08:09:38 crc kubenswrapper[4997]: E1205 08:09:38.652704 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24f313bc12a33950bc34659e4d80ae4d34a564a2fa1eba665b0cc68edc386bc1\": container with ID starting with 24f313bc12a33950bc34659e4d80ae4d34a564a2fa1eba665b0cc68edc386bc1 not found: ID does not exist" containerID="24f313bc12a33950bc34659e4d80ae4d34a564a2fa1eba665b0cc68edc386bc1" Dec 05 08:09:38 crc kubenswrapper[4997]: I1205 08:09:38.652756 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24f313bc12a33950bc34659e4d80ae4d34a564a2fa1eba665b0cc68edc386bc1"} err="failed to get container status \"24f313bc12a33950bc34659e4d80ae4d34a564a2fa1eba665b0cc68edc386bc1\": rpc error: code = NotFound desc = could not find container \"24f313bc12a33950bc34659e4d80ae4d34a564a2fa1eba665b0cc68edc386bc1\": container with ID starting with 24f313bc12a33950bc34659e4d80ae4d34a564a2fa1eba665b0cc68edc386bc1 not found: ID does not exist" Dec 05 08:09:39 crc kubenswrapper[4997]: I1205 08:09:39.764234 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37f162c9-ecf4-489c-8f20-fe32ffdff906" path="/var/lib/kubelet/pods/37f162c9-ecf4-489c-8f20-fe32ffdff906/volumes" Dec 05 08:09:49 crc kubenswrapper[4997]: I1205 08:09:49.770978 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:09:49 crc kubenswrapper[4997]: I1205 08:09:49.771797 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:10:19 crc kubenswrapper[4997]: I1205 08:10:19.770554 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:10:19 crc kubenswrapper[4997]: I1205 08:10:19.771365 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:10:49 crc kubenswrapper[4997]: I1205 08:10:49.770530 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:10:49 crc kubenswrapper[4997]: I1205 08:10:49.771146 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:10:49 crc kubenswrapper[4997]: I1205 08:10:49.771201 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 08:10:49 crc kubenswrapper[4997]: I1205 08:10:49.772035 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:10:49 crc kubenswrapper[4997]: I1205 08:10:49.772146 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" gracePeriod=600 Dec 05 08:10:49 crc kubenswrapper[4997]: E1205 08:10:49.895102 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:10:50 crc kubenswrapper[4997]: I1205 08:10:50.009511 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" exitCode=0 Dec 05 08:10:50 crc kubenswrapper[4997]: I1205 08:10:50.009553 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a"} Dec 05 08:10:50 crc kubenswrapper[4997]: I1205 08:10:50.009585 4997 scope.go:117] "RemoveContainer" containerID="e65f871a5d8f1b6b491a6cfe1489bad2714fbbf443695e3bcd99bcb1a0d11097" Dec 05 08:10:50 crc kubenswrapper[4997]: I1205 08:10:50.010065 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:10:50 crc kubenswrapper[4997]: E1205 08:10:50.010301 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:11:03 crc kubenswrapper[4997]: I1205 08:11:03.748836 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:11:03 crc kubenswrapper[4997]: E1205 08:11:03.749546 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:11:15 crc kubenswrapper[4997]: I1205 08:11:15.748556 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:11:15 crc kubenswrapper[4997]: E1205 08:11:15.749268 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:11:30 crc kubenswrapper[4997]: I1205 08:11:30.749581 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:11:30 crc kubenswrapper[4997]: E1205 08:11:30.750562 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:11:42 crc kubenswrapper[4997]: I1205 08:11:42.749656 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:11:42 crc kubenswrapper[4997]: E1205 08:11:42.750884 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:11:55 crc kubenswrapper[4997]: I1205 08:11:55.748819 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:11:55 crc kubenswrapper[4997]: E1205 08:11:55.749959 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:12:10 crc kubenswrapper[4997]: I1205 08:12:10.749743 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:12:10 crc kubenswrapper[4997]: E1205 08:12:10.750423 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:12:24 crc kubenswrapper[4997]: I1205 08:12:24.749943 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:12:24 crc kubenswrapper[4997]: E1205 08:12:24.752709 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:12:38 crc kubenswrapper[4997]: I1205 08:12:38.749354 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:12:38 crc kubenswrapper[4997]: E1205 08:12:38.750369 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:12:49 crc kubenswrapper[4997]: I1205 08:12:49.755186 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:12:49 crc kubenswrapper[4997]: E1205 08:12:49.755985 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:13:01 crc kubenswrapper[4997]: I1205 08:13:01.749589 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:13:01 crc kubenswrapper[4997]: E1205 08:13:01.750882 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:13:14 crc kubenswrapper[4997]: I1205 08:13:14.749791 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:13:14 crc kubenswrapper[4997]: E1205 08:13:14.751009 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:13:29 crc kubenswrapper[4997]: I1205 08:13:29.754564 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:13:29 crc kubenswrapper[4997]: E1205 08:13:29.755407 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:13:40 crc kubenswrapper[4997]: I1205 08:13:40.749767 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:13:40 crc kubenswrapper[4997]: E1205 08:13:40.750937 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:13:52 crc kubenswrapper[4997]: I1205 08:13:52.750023 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:13:52 crc kubenswrapper[4997]: E1205 08:13:52.751248 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:14:06 crc kubenswrapper[4997]: I1205 08:14:06.749195 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:14:06 crc kubenswrapper[4997]: E1205 08:14:06.750057 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:14:19 crc kubenswrapper[4997]: I1205 08:14:19.762347 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:14:19 crc kubenswrapper[4997]: E1205 08:14:19.763524 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:14:30 crc kubenswrapper[4997]: I1205 08:14:30.748408 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:14:30 crc kubenswrapper[4997]: E1205 08:14:30.749072 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:14:43 crc kubenswrapper[4997]: I1205 08:14:43.749526 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:14:43 crc kubenswrapper[4997]: E1205 08:14:43.750276 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.083268 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-98tqv"] Dec 05 08:14:50 crc kubenswrapper[4997]: E1205 08:14:50.084157 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37f162c9-ecf4-489c-8f20-fe32ffdff906" containerName="registry-server" Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.084177 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="37f162c9-ecf4-489c-8f20-fe32ffdff906" containerName="registry-server" Dec 05 08:14:50 crc kubenswrapper[4997]: E1205 08:14:50.084203 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37f162c9-ecf4-489c-8f20-fe32ffdff906" containerName="extract-content" Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.084211 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="37f162c9-ecf4-489c-8f20-fe32ffdff906" containerName="extract-content" Dec 05 08:14:50 crc kubenswrapper[4997]: E1205 08:14:50.084234 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37f162c9-ecf4-489c-8f20-fe32ffdff906" containerName="extract-utilities" Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.084243 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="37f162c9-ecf4-489c-8f20-fe32ffdff906" containerName="extract-utilities" Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.084434 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="37f162c9-ecf4-489c-8f20-fe32ffdff906" containerName="registry-server" Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.085704 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-98tqv" Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.099878 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-98tqv"] Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.191917 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49bea99f-636c-4f61-9a0c-6cf1ef50f038-catalog-content\") pod \"community-operators-98tqv\" (UID: \"49bea99f-636c-4f61-9a0c-6cf1ef50f038\") " pod="openshift-marketplace/community-operators-98tqv" Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.192030 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2xsp\" (UniqueName: \"kubernetes.io/projected/49bea99f-636c-4f61-9a0c-6cf1ef50f038-kube-api-access-z2xsp\") pod \"community-operators-98tqv\" (UID: \"49bea99f-636c-4f61-9a0c-6cf1ef50f038\") " pod="openshift-marketplace/community-operators-98tqv" Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.192155 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49bea99f-636c-4f61-9a0c-6cf1ef50f038-utilities\") pod \"community-operators-98tqv\" (UID: \"49bea99f-636c-4f61-9a0c-6cf1ef50f038\") " pod="openshift-marketplace/community-operators-98tqv" Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.293000 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49bea99f-636c-4f61-9a0c-6cf1ef50f038-catalog-content\") pod \"community-operators-98tqv\" (UID: \"49bea99f-636c-4f61-9a0c-6cf1ef50f038\") " pod="openshift-marketplace/community-operators-98tqv" Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.293056 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2xsp\" (UniqueName: \"kubernetes.io/projected/49bea99f-636c-4f61-9a0c-6cf1ef50f038-kube-api-access-z2xsp\") pod \"community-operators-98tqv\" (UID: \"49bea99f-636c-4f61-9a0c-6cf1ef50f038\") " pod="openshift-marketplace/community-operators-98tqv" Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.293158 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49bea99f-636c-4f61-9a0c-6cf1ef50f038-utilities\") pod \"community-operators-98tqv\" (UID: \"49bea99f-636c-4f61-9a0c-6cf1ef50f038\") " pod="openshift-marketplace/community-operators-98tqv" Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.293584 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49bea99f-636c-4f61-9a0c-6cf1ef50f038-utilities\") pod \"community-operators-98tqv\" (UID: \"49bea99f-636c-4f61-9a0c-6cf1ef50f038\") " pod="openshift-marketplace/community-operators-98tqv" Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.293853 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49bea99f-636c-4f61-9a0c-6cf1ef50f038-catalog-content\") pod \"community-operators-98tqv\" (UID: \"49bea99f-636c-4f61-9a0c-6cf1ef50f038\") " pod="openshift-marketplace/community-operators-98tqv" Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.313444 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2xsp\" (UniqueName: \"kubernetes.io/projected/49bea99f-636c-4f61-9a0c-6cf1ef50f038-kube-api-access-z2xsp\") pod \"community-operators-98tqv\" (UID: \"49bea99f-636c-4f61-9a0c-6cf1ef50f038\") " pod="openshift-marketplace/community-operators-98tqv" Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.405913 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-98tqv" Dec 05 08:14:50 crc kubenswrapper[4997]: I1205 08:14:50.711853 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-98tqv"] Dec 05 08:14:51 crc kubenswrapper[4997]: I1205 08:14:51.101276 4997 generic.go:334] "Generic (PLEG): container finished" podID="49bea99f-636c-4f61-9a0c-6cf1ef50f038" containerID="8945e735162e4547bbf13fbddc95239d0355119dbc4d0bd9d0c7d3a128ef436b" exitCode=0 Dec 05 08:14:51 crc kubenswrapper[4997]: I1205 08:14:51.101350 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98tqv" event={"ID":"49bea99f-636c-4f61-9a0c-6cf1ef50f038","Type":"ContainerDied","Data":"8945e735162e4547bbf13fbddc95239d0355119dbc4d0bd9d0c7d3a128ef436b"} Dec 05 08:14:51 crc kubenswrapper[4997]: I1205 08:14:51.101421 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98tqv" event={"ID":"49bea99f-636c-4f61-9a0c-6cf1ef50f038","Type":"ContainerStarted","Data":"d745da24d2726f58023d0ff705cae5d1a1a90156b82aef8bd2cd1fea5900391f"} Dec 05 08:14:51 crc kubenswrapper[4997]: I1205 08:14:51.103453 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 08:14:52 crc kubenswrapper[4997]: I1205 08:14:52.109846 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98tqv" event={"ID":"49bea99f-636c-4f61-9a0c-6cf1ef50f038","Type":"ContainerStarted","Data":"ab6777bd728643dd522bb57f952f67fdc240839bc2219325a0792b2114457942"} Dec 05 08:14:53 crc kubenswrapper[4997]: I1205 08:14:53.119207 4997 generic.go:334] "Generic (PLEG): container finished" podID="49bea99f-636c-4f61-9a0c-6cf1ef50f038" containerID="ab6777bd728643dd522bb57f952f67fdc240839bc2219325a0792b2114457942" exitCode=0 Dec 05 08:14:53 crc kubenswrapper[4997]: I1205 08:14:53.119283 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98tqv" event={"ID":"49bea99f-636c-4f61-9a0c-6cf1ef50f038","Type":"ContainerDied","Data":"ab6777bd728643dd522bb57f952f67fdc240839bc2219325a0792b2114457942"} Dec 05 08:14:54 crc kubenswrapper[4997]: I1205 08:14:54.128621 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98tqv" event={"ID":"49bea99f-636c-4f61-9a0c-6cf1ef50f038","Type":"ContainerStarted","Data":"ba44bed977da25a5e197ace2ff8c00b0c102a903e32b73e674544ea79aca4b06"} Dec 05 08:14:54 crc kubenswrapper[4997]: I1205 08:14:54.147940 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-98tqv" podStartSLOduration=1.6238497170000001 podStartE2EDuration="4.147924009s" podCreationTimestamp="2025-12-05 08:14:50 +0000 UTC" firstStartedPulling="2025-12-05 08:14:51.103177161 +0000 UTC m=+4791.632084422" lastFinishedPulling="2025-12-05 08:14:53.627251433 +0000 UTC m=+4794.156158714" observedRunningTime="2025-12-05 08:14:54.14498978 +0000 UTC m=+4794.673897051" watchObservedRunningTime="2025-12-05 08:14:54.147924009 +0000 UTC m=+4794.676831270" Dec 05 08:14:54 crc kubenswrapper[4997]: I1205 08:14:54.749719 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:14:54 crc kubenswrapper[4997]: E1205 08:14:54.750094 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.150284 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864"] Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.151820 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864" Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.153919 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.155538 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.160862 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864"] Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.246971 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-config-volume\") pod \"collect-profiles-29415375-9k864\" (UID: \"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864" Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.247044 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-secret-volume\") pod \"collect-profiles-29415375-9k864\" (UID: \"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864" Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.247081 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpmxk\" (UniqueName: \"kubernetes.io/projected/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-kube-api-access-lpmxk\") pod \"collect-profiles-29415375-9k864\" (UID: \"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864" Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.349014 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-config-volume\") pod \"collect-profiles-29415375-9k864\" (UID: \"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864" Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.349101 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-secret-volume\") pod \"collect-profiles-29415375-9k864\" (UID: \"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864" Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.349196 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpmxk\" (UniqueName: \"kubernetes.io/projected/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-kube-api-access-lpmxk\") pod \"collect-profiles-29415375-9k864\" (UID: \"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864" Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.350083 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-config-volume\") pod \"collect-profiles-29415375-9k864\" (UID: \"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864" Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.355009 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-secret-volume\") pod \"collect-profiles-29415375-9k864\" (UID: \"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864" Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.371497 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpmxk\" (UniqueName: \"kubernetes.io/projected/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-kube-api-access-lpmxk\") pod \"collect-profiles-29415375-9k864\" (UID: \"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864" Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.406869 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-98tqv" Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.407180 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-98tqv" Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.450449 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-98tqv" Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.477895 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864" Dec 05 08:15:00 crc kubenswrapper[4997]: I1205 08:15:00.686474 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864"] Dec 05 08:15:01 crc kubenswrapper[4997]: I1205 08:15:01.187529 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864" event={"ID":"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f","Type":"ContainerStarted","Data":"b2fb99755c8ff05fa5ed8e6ded6263f26c9266d0364ca17d0ecc989f0de79f6b"} Dec 05 08:15:01 crc kubenswrapper[4997]: I1205 08:15:01.250078 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-98tqv" Dec 05 08:15:01 crc kubenswrapper[4997]: I1205 08:15:01.324761 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-98tqv"] Dec 05 08:15:02 crc kubenswrapper[4997]: I1205 08:15:02.198439 4997 generic.go:334] "Generic (PLEG): container finished" podID="0228be20-bfff-4d5b-97d9-8b5f8ace9d5f" containerID="0ccda522da9ed3273e605f81df796f6d14bb94a65119065bdd8959c257ab6867" exitCode=0 Dec 05 08:15:02 crc kubenswrapper[4997]: I1205 08:15:02.198581 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864" event={"ID":"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f","Type":"ContainerDied","Data":"0ccda522da9ed3273e605f81df796f6d14bb94a65119065bdd8959c257ab6867"} Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.205798 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-98tqv" podUID="49bea99f-636c-4f61-9a0c-6cf1ef50f038" containerName="registry-server" containerID="cri-o://ba44bed977da25a5e197ace2ff8c00b0c102a903e32b73e674544ea79aca4b06" gracePeriod=2 Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.484991 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864" Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.597081 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpmxk\" (UniqueName: \"kubernetes.io/projected/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-kube-api-access-lpmxk\") pod \"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f\" (UID: \"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f\") " Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.597357 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-secret-volume\") pod \"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f\" (UID: \"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f\") " Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.597652 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-config-volume\") pod \"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f\" (UID: \"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f\") " Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.598741 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-config-volume" (OuterVolumeSpecName: "config-volume") pod "0228be20-bfff-4d5b-97d9-8b5f8ace9d5f" (UID: "0228be20-bfff-4d5b-97d9-8b5f8ace9d5f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.603270 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-kube-api-access-lpmxk" (OuterVolumeSpecName: "kube-api-access-lpmxk") pod "0228be20-bfff-4d5b-97d9-8b5f8ace9d5f" (UID: "0228be20-bfff-4d5b-97d9-8b5f8ace9d5f"). InnerVolumeSpecName "kube-api-access-lpmxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.603555 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0228be20-bfff-4d5b-97d9-8b5f8ace9d5f" (UID: "0228be20-bfff-4d5b-97d9-8b5f8ace9d5f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.604163 4997 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.604197 4997 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.604210 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpmxk\" (UniqueName: \"kubernetes.io/projected/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f-kube-api-access-lpmxk\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.621543 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-98tqv" Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.705072 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49bea99f-636c-4f61-9a0c-6cf1ef50f038-utilities\") pod \"49bea99f-636c-4f61-9a0c-6cf1ef50f038\" (UID: \"49bea99f-636c-4f61-9a0c-6cf1ef50f038\") " Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.705128 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49bea99f-636c-4f61-9a0c-6cf1ef50f038-catalog-content\") pod \"49bea99f-636c-4f61-9a0c-6cf1ef50f038\" (UID: \"49bea99f-636c-4f61-9a0c-6cf1ef50f038\") " Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.705178 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2xsp\" (UniqueName: \"kubernetes.io/projected/49bea99f-636c-4f61-9a0c-6cf1ef50f038-kube-api-access-z2xsp\") pod \"49bea99f-636c-4f61-9a0c-6cf1ef50f038\" (UID: \"49bea99f-636c-4f61-9a0c-6cf1ef50f038\") " Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.706270 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49bea99f-636c-4f61-9a0c-6cf1ef50f038-utilities" (OuterVolumeSpecName: "utilities") pod "49bea99f-636c-4f61-9a0c-6cf1ef50f038" (UID: "49bea99f-636c-4f61-9a0c-6cf1ef50f038"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.783142 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49bea99f-636c-4f61-9a0c-6cf1ef50f038-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49bea99f-636c-4f61-9a0c-6cf1ef50f038" (UID: "49bea99f-636c-4f61-9a0c-6cf1ef50f038"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.807271 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49bea99f-636c-4f61-9a0c-6cf1ef50f038-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:03 crc kubenswrapper[4997]: I1205 08:15:03.807304 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49bea99f-636c-4f61-9a0c-6cf1ef50f038-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.001785 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49bea99f-636c-4f61-9a0c-6cf1ef50f038-kube-api-access-z2xsp" (OuterVolumeSpecName: "kube-api-access-z2xsp") pod "49bea99f-636c-4f61-9a0c-6cf1ef50f038" (UID: "49bea99f-636c-4f61-9a0c-6cf1ef50f038"). InnerVolumeSpecName "kube-api-access-z2xsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.010324 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2xsp\" (UniqueName: \"kubernetes.io/projected/49bea99f-636c-4f61-9a0c-6cf1ef50f038-kube-api-access-z2xsp\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.218035 4997 generic.go:334] "Generic (PLEG): container finished" podID="49bea99f-636c-4f61-9a0c-6cf1ef50f038" containerID="ba44bed977da25a5e197ace2ff8c00b0c102a903e32b73e674544ea79aca4b06" exitCode=0 Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.218340 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98tqv" event={"ID":"49bea99f-636c-4f61-9a0c-6cf1ef50f038","Type":"ContainerDied","Data":"ba44bed977da25a5e197ace2ff8c00b0c102a903e32b73e674544ea79aca4b06"} Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.218360 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-98tqv" Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.218426 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98tqv" event={"ID":"49bea99f-636c-4f61-9a0c-6cf1ef50f038","Type":"ContainerDied","Data":"d745da24d2726f58023d0ff705cae5d1a1a90156b82aef8bd2cd1fea5900391f"} Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.218459 4997 scope.go:117] "RemoveContainer" containerID="ba44bed977da25a5e197ace2ff8c00b0c102a903e32b73e674544ea79aca4b06" Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.220142 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864" event={"ID":"0228be20-bfff-4d5b-97d9-8b5f8ace9d5f","Type":"ContainerDied","Data":"b2fb99755c8ff05fa5ed8e6ded6263f26c9266d0364ca17d0ecc989f0de79f6b"} Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.220184 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2fb99755c8ff05fa5ed8e6ded6263f26c9266d0364ca17d0ecc989f0de79f6b" Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.220159 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864" Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.238457 4997 scope.go:117] "RemoveContainer" containerID="ab6777bd728643dd522bb57f952f67fdc240839bc2219325a0792b2114457942" Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.264991 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-98tqv"] Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.269261 4997 scope.go:117] "RemoveContainer" containerID="8945e735162e4547bbf13fbddc95239d0355119dbc4d0bd9d0c7d3a128ef436b" Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.276745 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-98tqv"] Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.290473 4997 scope.go:117] "RemoveContainer" containerID="ba44bed977da25a5e197ace2ff8c00b0c102a903e32b73e674544ea79aca4b06" Dec 05 08:15:04 crc kubenswrapper[4997]: E1205 08:15:04.290934 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba44bed977da25a5e197ace2ff8c00b0c102a903e32b73e674544ea79aca4b06\": container with ID starting with ba44bed977da25a5e197ace2ff8c00b0c102a903e32b73e674544ea79aca4b06 not found: ID does not exist" containerID="ba44bed977da25a5e197ace2ff8c00b0c102a903e32b73e674544ea79aca4b06" Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.290965 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba44bed977da25a5e197ace2ff8c00b0c102a903e32b73e674544ea79aca4b06"} err="failed to get container status \"ba44bed977da25a5e197ace2ff8c00b0c102a903e32b73e674544ea79aca4b06\": rpc error: code = NotFound desc = could not find container \"ba44bed977da25a5e197ace2ff8c00b0c102a903e32b73e674544ea79aca4b06\": container with ID starting with ba44bed977da25a5e197ace2ff8c00b0c102a903e32b73e674544ea79aca4b06 not found: ID does not exist" Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.290986 4997 scope.go:117] "RemoveContainer" containerID="ab6777bd728643dd522bb57f952f67fdc240839bc2219325a0792b2114457942" Dec 05 08:15:04 crc kubenswrapper[4997]: E1205 08:15:04.291191 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab6777bd728643dd522bb57f952f67fdc240839bc2219325a0792b2114457942\": container with ID starting with ab6777bd728643dd522bb57f952f67fdc240839bc2219325a0792b2114457942 not found: ID does not exist" containerID="ab6777bd728643dd522bb57f952f67fdc240839bc2219325a0792b2114457942" Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.291213 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab6777bd728643dd522bb57f952f67fdc240839bc2219325a0792b2114457942"} err="failed to get container status \"ab6777bd728643dd522bb57f952f67fdc240839bc2219325a0792b2114457942\": rpc error: code = NotFound desc = could not find container \"ab6777bd728643dd522bb57f952f67fdc240839bc2219325a0792b2114457942\": container with ID starting with ab6777bd728643dd522bb57f952f67fdc240839bc2219325a0792b2114457942 not found: ID does not exist" Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.291226 4997 scope.go:117] "RemoveContainer" containerID="8945e735162e4547bbf13fbddc95239d0355119dbc4d0bd9d0c7d3a128ef436b" Dec 05 08:15:04 crc kubenswrapper[4997]: E1205 08:15:04.291498 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8945e735162e4547bbf13fbddc95239d0355119dbc4d0bd9d0c7d3a128ef436b\": container with ID starting with 8945e735162e4547bbf13fbddc95239d0355119dbc4d0bd9d0c7d3a128ef436b not found: ID does not exist" containerID="8945e735162e4547bbf13fbddc95239d0355119dbc4d0bd9d0c7d3a128ef436b" Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.291519 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8945e735162e4547bbf13fbddc95239d0355119dbc4d0bd9d0c7d3a128ef436b"} err="failed to get container status \"8945e735162e4547bbf13fbddc95239d0355119dbc4d0bd9d0c7d3a128ef436b\": rpc error: code = NotFound desc = could not find container \"8945e735162e4547bbf13fbddc95239d0355119dbc4d0bd9d0c7d3a128ef436b\": container with ID starting with 8945e735162e4547bbf13fbddc95239d0355119dbc4d0bd9d0c7d3a128ef436b not found: ID does not exist" Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.583502 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl"] Dec 05 08:15:04 crc kubenswrapper[4997]: I1205 08:15:04.594444 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415330-h9pcl"] Dec 05 08:15:05 crc kubenswrapper[4997]: I1205 08:15:05.764940 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49bea99f-636c-4f61-9a0c-6cf1ef50f038" path="/var/lib/kubelet/pods/49bea99f-636c-4f61-9a0c-6cf1ef50f038/volumes" Dec 05 08:15:05 crc kubenswrapper[4997]: I1205 08:15:05.767605 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f3272ac-8f40-4c96-95b0-31b7e78e6318" path="/var/lib/kubelet/pods/8f3272ac-8f40-4c96-95b0-31b7e78e6318/volumes" Dec 05 08:15:09 crc kubenswrapper[4997]: I1205 08:15:09.754741 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:15:09 crc kubenswrapper[4997]: E1205 08:15:09.755328 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:15:11 crc kubenswrapper[4997]: I1205 08:15:11.925685 4997 scope.go:117] "RemoveContainer" containerID="224eab4b424ce3861377ade439c8d1bea513ab185533a7173886823a2fa6936e" Dec 05 08:15:22 crc kubenswrapper[4997]: I1205 08:15:22.749344 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:15:22 crc kubenswrapper[4997]: E1205 08:15:22.750317 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:15:37 crc kubenswrapper[4997]: I1205 08:15:37.749679 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:15:37 crc kubenswrapper[4997]: E1205 08:15:37.750775 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:15:52 crc kubenswrapper[4997]: I1205 08:15:52.749046 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:15:53 crc kubenswrapper[4997]: I1205 08:15:53.644418 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"f3907db969deab75a3846872db0276a30995a72fb7e0b6fc2b81e6f61bfe6a31"} Dec 05 08:16:19 crc kubenswrapper[4997]: I1205 08:16:19.988981 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-twsjv"] Dec 05 08:16:19 crc kubenswrapper[4997]: E1205 08:16:19.989809 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49bea99f-636c-4f61-9a0c-6cf1ef50f038" containerName="registry-server" Dec 05 08:16:19 crc kubenswrapper[4997]: I1205 08:16:19.989827 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="49bea99f-636c-4f61-9a0c-6cf1ef50f038" containerName="registry-server" Dec 05 08:16:19 crc kubenswrapper[4997]: E1205 08:16:19.989840 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0228be20-bfff-4d5b-97d9-8b5f8ace9d5f" containerName="collect-profiles" Dec 05 08:16:19 crc kubenswrapper[4997]: I1205 08:16:19.989847 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0228be20-bfff-4d5b-97d9-8b5f8ace9d5f" containerName="collect-profiles" Dec 05 08:16:19 crc kubenswrapper[4997]: E1205 08:16:19.989859 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49bea99f-636c-4f61-9a0c-6cf1ef50f038" containerName="extract-content" Dec 05 08:16:19 crc kubenswrapper[4997]: I1205 08:16:19.989866 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="49bea99f-636c-4f61-9a0c-6cf1ef50f038" containerName="extract-content" Dec 05 08:16:19 crc kubenswrapper[4997]: E1205 08:16:19.989881 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49bea99f-636c-4f61-9a0c-6cf1ef50f038" containerName="extract-utilities" Dec 05 08:16:19 crc kubenswrapper[4997]: I1205 08:16:19.989886 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="49bea99f-636c-4f61-9a0c-6cf1ef50f038" containerName="extract-utilities" Dec 05 08:16:19 crc kubenswrapper[4997]: I1205 08:16:19.990024 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0228be20-bfff-4d5b-97d9-8b5f8ace9d5f" containerName="collect-profiles" Dec 05 08:16:19 crc kubenswrapper[4997]: I1205 08:16:19.990042 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="49bea99f-636c-4f61-9a0c-6cf1ef50f038" containerName="registry-server" Dec 05 08:16:19 crc kubenswrapper[4997]: I1205 08:16:19.991239 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-twsjv" Dec 05 08:16:20 crc kubenswrapper[4997]: I1205 08:16:20.032542 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-twsjv"] Dec 05 08:16:20 crc kubenswrapper[4997]: I1205 08:16:20.173965 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cngx6\" (UniqueName: \"kubernetes.io/projected/a80a1a81-c809-40d7-9c58-6943b7f22858-kube-api-access-cngx6\") pod \"certified-operators-twsjv\" (UID: \"a80a1a81-c809-40d7-9c58-6943b7f22858\") " pod="openshift-marketplace/certified-operators-twsjv" Dec 05 08:16:20 crc kubenswrapper[4997]: I1205 08:16:20.174062 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80a1a81-c809-40d7-9c58-6943b7f22858-utilities\") pod \"certified-operators-twsjv\" (UID: \"a80a1a81-c809-40d7-9c58-6943b7f22858\") " pod="openshift-marketplace/certified-operators-twsjv" Dec 05 08:16:20 crc kubenswrapper[4997]: I1205 08:16:20.174093 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80a1a81-c809-40d7-9c58-6943b7f22858-catalog-content\") pod \"certified-operators-twsjv\" (UID: \"a80a1a81-c809-40d7-9c58-6943b7f22858\") " pod="openshift-marketplace/certified-operators-twsjv" Dec 05 08:16:20 crc kubenswrapper[4997]: I1205 08:16:20.275524 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cngx6\" (UniqueName: \"kubernetes.io/projected/a80a1a81-c809-40d7-9c58-6943b7f22858-kube-api-access-cngx6\") pod \"certified-operators-twsjv\" (UID: \"a80a1a81-c809-40d7-9c58-6943b7f22858\") " pod="openshift-marketplace/certified-operators-twsjv" Dec 05 08:16:20 crc kubenswrapper[4997]: I1205 08:16:20.275635 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80a1a81-c809-40d7-9c58-6943b7f22858-utilities\") pod \"certified-operators-twsjv\" (UID: \"a80a1a81-c809-40d7-9c58-6943b7f22858\") " pod="openshift-marketplace/certified-operators-twsjv" Dec 05 08:16:20 crc kubenswrapper[4997]: I1205 08:16:20.275678 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80a1a81-c809-40d7-9c58-6943b7f22858-catalog-content\") pod \"certified-operators-twsjv\" (UID: \"a80a1a81-c809-40d7-9c58-6943b7f22858\") " pod="openshift-marketplace/certified-operators-twsjv" Dec 05 08:16:20 crc kubenswrapper[4997]: I1205 08:16:20.276191 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80a1a81-c809-40d7-9c58-6943b7f22858-utilities\") pod \"certified-operators-twsjv\" (UID: \"a80a1a81-c809-40d7-9c58-6943b7f22858\") " pod="openshift-marketplace/certified-operators-twsjv" Dec 05 08:16:20 crc kubenswrapper[4997]: I1205 08:16:20.276370 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80a1a81-c809-40d7-9c58-6943b7f22858-catalog-content\") pod \"certified-operators-twsjv\" (UID: \"a80a1a81-c809-40d7-9c58-6943b7f22858\") " pod="openshift-marketplace/certified-operators-twsjv" Dec 05 08:16:20 crc kubenswrapper[4997]: I1205 08:16:20.294257 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cngx6\" (UniqueName: \"kubernetes.io/projected/a80a1a81-c809-40d7-9c58-6943b7f22858-kube-api-access-cngx6\") pod \"certified-operators-twsjv\" (UID: \"a80a1a81-c809-40d7-9c58-6943b7f22858\") " pod="openshift-marketplace/certified-operators-twsjv" Dec 05 08:16:20 crc kubenswrapper[4997]: I1205 08:16:20.333479 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-twsjv" Dec 05 08:16:20 crc kubenswrapper[4997]: I1205 08:16:20.620823 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-twsjv"] Dec 05 08:16:20 crc kubenswrapper[4997]: W1205 08:16:20.638708 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda80a1a81_c809_40d7_9c58_6943b7f22858.slice/crio-dc8323fc6bfd4ac72b9f2195000d025e01583aeb9208f474d1f87c3010e8c0d3 WatchSource:0}: Error finding container dc8323fc6bfd4ac72b9f2195000d025e01583aeb9208f474d1f87c3010e8c0d3: Status 404 returned error can't find the container with id dc8323fc6bfd4ac72b9f2195000d025e01583aeb9208f474d1f87c3010e8c0d3 Dec 05 08:16:20 crc kubenswrapper[4997]: I1205 08:16:20.896181 4997 generic.go:334] "Generic (PLEG): container finished" podID="a80a1a81-c809-40d7-9c58-6943b7f22858" containerID="05b440cf37a3c7d281b7571f0ed8ebd4095cafccfc218b354576c41e8ebe5920" exitCode=0 Dec 05 08:16:20 crc kubenswrapper[4997]: I1205 08:16:20.896217 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-twsjv" event={"ID":"a80a1a81-c809-40d7-9c58-6943b7f22858","Type":"ContainerDied","Data":"05b440cf37a3c7d281b7571f0ed8ebd4095cafccfc218b354576c41e8ebe5920"} Dec 05 08:16:20 crc kubenswrapper[4997]: I1205 08:16:20.896466 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-twsjv" event={"ID":"a80a1a81-c809-40d7-9c58-6943b7f22858","Type":"ContainerStarted","Data":"dc8323fc6bfd4ac72b9f2195000d025e01583aeb9208f474d1f87c3010e8c0d3"} Dec 05 08:16:21 crc kubenswrapper[4997]: I1205 08:16:21.906176 4997 generic.go:334] "Generic (PLEG): container finished" podID="a80a1a81-c809-40d7-9c58-6943b7f22858" containerID="55090ec00f8c1dc1694479b4d69ac97dede3c2de72c884963483364fdbe5668d" exitCode=0 Dec 05 08:16:21 crc kubenswrapper[4997]: I1205 08:16:21.906265 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-twsjv" event={"ID":"a80a1a81-c809-40d7-9c58-6943b7f22858","Type":"ContainerDied","Data":"55090ec00f8c1dc1694479b4d69ac97dede3c2de72c884963483364fdbe5668d"} Dec 05 08:16:22 crc kubenswrapper[4997]: I1205 08:16:22.916560 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-twsjv" event={"ID":"a80a1a81-c809-40d7-9c58-6943b7f22858","Type":"ContainerStarted","Data":"c43e81763ef8c8f7acbe720f2018fc8a1e59e5da702b60feb0ee80dafe3a6ac4"} Dec 05 08:16:22 crc kubenswrapper[4997]: I1205 08:16:22.943276 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-twsjv" podStartSLOduration=2.54329107 podStartE2EDuration="3.94325289s" podCreationTimestamp="2025-12-05 08:16:19 +0000 UTC" firstStartedPulling="2025-12-05 08:16:20.898448937 +0000 UTC m=+4881.427356198" lastFinishedPulling="2025-12-05 08:16:22.298410737 +0000 UTC m=+4882.827318018" observedRunningTime="2025-12-05 08:16:22.938547304 +0000 UTC m=+4883.467454655" watchObservedRunningTime="2025-12-05 08:16:22.94325289 +0000 UTC m=+4883.472160161" Dec 05 08:16:30 crc kubenswrapper[4997]: I1205 08:16:30.334322 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-twsjv" Dec 05 08:16:30 crc kubenswrapper[4997]: I1205 08:16:30.336463 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-twsjv" Dec 05 08:16:30 crc kubenswrapper[4997]: I1205 08:16:30.402191 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-twsjv" Dec 05 08:16:31 crc kubenswrapper[4997]: I1205 08:16:31.060009 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-twsjv" Dec 05 08:16:31 crc kubenswrapper[4997]: I1205 08:16:31.123922 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-twsjv"] Dec 05 08:16:33 crc kubenswrapper[4997]: I1205 08:16:33.004891 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-twsjv" podUID="a80a1a81-c809-40d7-9c58-6943b7f22858" containerName="registry-server" containerID="cri-o://c43e81763ef8c8f7acbe720f2018fc8a1e59e5da702b60feb0ee80dafe3a6ac4" gracePeriod=2 Dec 05 08:16:34 crc kubenswrapper[4997]: I1205 08:16:34.012826 4997 generic.go:334] "Generic (PLEG): container finished" podID="a80a1a81-c809-40d7-9c58-6943b7f22858" containerID="c43e81763ef8c8f7acbe720f2018fc8a1e59e5da702b60feb0ee80dafe3a6ac4" exitCode=0 Dec 05 08:16:34 crc kubenswrapper[4997]: I1205 08:16:34.012871 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-twsjv" event={"ID":"a80a1a81-c809-40d7-9c58-6943b7f22858","Type":"ContainerDied","Data":"c43e81763ef8c8f7acbe720f2018fc8a1e59e5da702b60feb0ee80dafe3a6ac4"} Dec 05 08:16:34 crc kubenswrapper[4997]: I1205 08:16:34.012901 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-twsjv" event={"ID":"a80a1a81-c809-40d7-9c58-6943b7f22858","Type":"ContainerDied","Data":"dc8323fc6bfd4ac72b9f2195000d025e01583aeb9208f474d1f87c3010e8c0d3"} Dec 05 08:16:34 crc kubenswrapper[4997]: I1205 08:16:34.012916 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc8323fc6bfd4ac72b9f2195000d025e01583aeb9208f474d1f87c3010e8c0d3" Dec 05 08:16:34 crc kubenswrapper[4997]: I1205 08:16:34.037317 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-twsjv" Dec 05 08:16:34 crc kubenswrapper[4997]: I1205 08:16:34.182938 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80a1a81-c809-40d7-9c58-6943b7f22858-catalog-content\") pod \"a80a1a81-c809-40d7-9c58-6943b7f22858\" (UID: \"a80a1a81-c809-40d7-9c58-6943b7f22858\") " Dec 05 08:16:34 crc kubenswrapper[4997]: I1205 08:16:34.183006 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cngx6\" (UniqueName: \"kubernetes.io/projected/a80a1a81-c809-40d7-9c58-6943b7f22858-kube-api-access-cngx6\") pod \"a80a1a81-c809-40d7-9c58-6943b7f22858\" (UID: \"a80a1a81-c809-40d7-9c58-6943b7f22858\") " Dec 05 08:16:34 crc kubenswrapper[4997]: I1205 08:16:34.183055 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80a1a81-c809-40d7-9c58-6943b7f22858-utilities\") pod \"a80a1a81-c809-40d7-9c58-6943b7f22858\" (UID: \"a80a1a81-c809-40d7-9c58-6943b7f22858\") " Dec 05 08:16:34 crc kubenswrapper[4997]: I1205 08:16:34.184150 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a80a1a81-c809-40d7-9c58-6943b7f22858-utilities" (OuterVolumeSpecName: "utilities") pod "a80a1a81-c809-40d7-9c58-6943b7f22858" (UID: "a80a1a81-c809-40d7-9c58-6943b7f22858"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:16:34 crc kubenswrapper[4997]: I1205 08:16:34.188170 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a80a1a81-c809-40d7-9c58-6943b7f22858-kube-api-access-cngx6" (OuterVolumeSpecName: "kube-api-access-cngx6") pod "a80a1a81-c809-40d7-9c58-6943b7f22858" (UID: "a80a1a81-c809-40d7-9c58-6943b7f22858"). InnerVolumeSpecName "kube-api-access-cngx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:16:34 crc kubenswrapper[4997]: I1205 08:16:34.234214 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a80a1a81-c809-40d7-9c58-6943b7f22858-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a80a1a81-c809-40d7-9c58-6943b7f22858" (UID: "a80a1a81-c809-40d7-9c58-6943b7f22858"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:16:34 crc kubenswrapper[4997]: I1205 08:16:34.284845 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80a1a81-c809-40d7-9c58-6943b7f22858-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:16:34 crc kubenswrapper[4997]: I1205 08:16:34.284878 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cngx6\" (UniqueName: \"kubernetes.io/projected/a80a1a81-c809-40d7-9c58-6943b7f22858-kube-api-access-cngx6\") on node \"crc\" DevicePath \"\"" Dec 05 08:16:34 crc kubenswrapper[4997]: I1205 08:16:34.284892 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80a1a81-c809-40d7-9c58-6943b7f22858-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:16:35 crc kubenswrapper[4997]: I1205 08:16:35.023393 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-twsjv" Dec 05 08:16:35 crc kubenswrapper[4997]: I1205 08:16:35.087373 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-twsjv"] Dec 05 08:16:35 crc kubenswrapper[4997]: I1205 08:16:35.097041 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-twsjv"] Dec 05 08:16:35 crc kubenswrapper[4997]: I1205 08:16:35.765545 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a80a1a81-c809-40d7-9c58-6943b7f22858" path="/var/lib/kubelet/pods/a80a1a81-c809-40d7-9c58-6943b7f22858/volumes" Dec 05 08:18:19 crc kubenswrapper[4997]: I1205 08:18:19.770084 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:18:19 crc kubenswrapper[4997]: I1205 08:18:19.770782 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:18:49 crc kubenswrapper[4997]: I1205 08:18:49.770149 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:18:49 crc kubenswrapper[4997]: I1205 08:18:49.771007 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:19:19 crc kubenswrapper[4997]: I1205 08:19:19.770554 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:19:19 crc kubenswrapper[4997]: I1205 08:19:19.771072 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:19:19 crc kubenswrapper[4997]: I1205 08:19:19.771105 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 08:19:19 crc kubenswrapper[4997]: I1205 08:19:19.771571 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f3907db969deab75a3846872db0276a30995a72fb7e0b6fc2b81e6f61bfe6a31"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:19:19 crc kubenswrapper[4997]: I1205 08:19:19.771641 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://f3907db969deab75a3846872db0276a30995a72fb7e0b6fc2b81e6f61bfe6a31" gracePeriod=600 Dec 05 08:19:20 crc kubenswrapper[4997]: I1205 08:19:20.636639 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="f3907db969deab75a3846872db0276a30995a72fb7e0b6fc2b81e6f61bfe6a31" exitCode=0 Dec 05 08:19:20 crc kubenswrapper[4997]: I1205 08:19:20.636675 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"f3907db969deab75a3846872db0276a30995a72fb7e0b6fc2b81e6f61bfe6a31"} Dec 05 08:19:20 crc kubenswrapper[4997]: I1205 08:19:20.637268 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e"} Dec 05 08:19:20 crc kubenswrapper[4997]: I1205 08:19:20.637315 4997 scope.go:117] "RemoveContainer" containerID="fa3e0450d0e5475b27c3e586a791092c01c266bd0fce53aea77f158048e42a0a" Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.135797 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cg92w"] Dec 05 08:19:29 crc kubenswrapper[4997]: E1205 08:19:29.138267 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a80a1a81-c809-40d7-9c58-6943b7f22858" containerName="registry-server" Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.138402 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a80a1a81-c809-40d7-9c58-6943b7f22858" containerName="registry-server" Dec 05 08:19:29 crc kubenswrapper[4997]: E1205 08:19:29.138564 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a80a1a81-c809-40d7-9c58-6943b7f22858" containerName="extract-content" Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.138714 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a80a1a81-c809-40d7-9c58-6943b7f22858" containerName="extract-content" Dec 05 08:19:29 crc kubenswrapper[4997]: E1205 08:19:29.138853 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a80a1a81-c809-40d7-9c58-6943b7f22858" containerName="extract-utilities" Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.138964 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a80a1a81-c809-40d7-9c58-6943b7f22858" containerName="extract-utilities" Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.139269 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a80a1a81-c809-40d7-9c58-6943b7f22858" containerName="registry-server" Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.141031 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cg92w" Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.153078 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cg92w"] Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.316282 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfcmd\" (UniqueName: \"kubernetes.io/projected/895a6cef-c397-43cf-9aa3-6190b710b8c3-kube-api-access-pfcmd\") pod \"redhat-operators-cg92w\" (UID: \"895a6cef-c397-43cf-9aa3-6190b710b8c3\") " pod="openshift-marketplace/redhat-operators-cg92w" Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.316350 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/895a6cef-c397-43cf-9aa3-6190b710b8c3-catalog-content\") pod \"redhat-operators-cg92w\" (UID: \"895a6cef-c397-43cf-9aa3-6190b710b8c3\") " pod="openshift-marketplace/redhat-operators-cg92w" Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.316487 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/895a6cef-c397-43cf-9aa3-6190b710b8c3-utilities\") pod \"redhat-operators-cg92w\" (UID: \"895a6cef-c397-43cf-9aa3-6190b710b8c3\") " pod="openshift-marketplace/redhat-operators-cg92w" Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.418029 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/895a6cef-c397-43cf-9aa3-6190b710b8c3-utilities\") pod \"redhat-operators-cg92w\" (UID: \"895a6cef-c397-43cf-9aa3-6190b710b8c3\") " pod="openshift-marketplace/redhat-operators-cg92w" Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.418477 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfcmd\" (UniqueName: \"kubernetes.io/projected/895a6cef-c397-43cf-9aa3-6190b710b8c3-kube-api-access-pfcmd\") pod \"redhat-operators-cg92w\" (UID: \"895a6cef-c397-43cf-9aa3-6190b710b8c3\") " pod="openshift-marketplace/redhat-operators-cg92w" Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.419011 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/895a6cef-c397-43cf-9aa3-6190b710b8c3-catalog-content\") pod \"redhat-operators-cg92w\" (UID: \"895a6cef-c397-43cf-9aa3-6190b710b8c3\") " pod="openshift-marketplace/redhat-operators-cg92w" Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.418863 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/895a6cef-c397-43cf-9aa3-6190b710b8c3-utilities\") pod \"redhat-operators-cg92w\" (UID: \"895a6cef-c397-43cf-9aa3-6190b710b8c3\") " pod="openshift-marketplace/redhat-operators-cg92w" Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.419442 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/895a6cef-c397-43cf-9aa3-6190b710b8c3-catalog-content\") pod \"redhat-operators-cg92w\" (UID: \"895a6cef-c397-43cf-9aa3-6190b710b8c3\") " pod="openshift-marketplace/redhat-operators-cg92w" Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.437484 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfcmd\" (UniqueName: \"kubernetes.io/projected/895a6cef-c397-43cf-9aa3-6190b710b8c3-kube-api-access-pfcmd\") pod \"redhat-operators-cg92w\" (UID: \"895a6cef-c397-43cf-9aa3-6190b710b8c3\") " pod="openshift-marketplace/redhat-operators-cg92w" Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.475177 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cg92w" Dec 05 08:19:29 crc kubenswrapper[4997]: I1205 08:19:29.895163 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cg92w"] Dec 05 08:19:30 crc kubenswrapper[4997]: I1205 08:19:30.738672 4997 generic.go:334] "Generic (PLEG): container finished" podID="895a6cef-c397-43cf-9aa3-6190b710b8c3" containerID="352715e1ed6f48cd16e2433f52b6e0340d29edf8ef735d9220a9213cad76e961" exitCode=0 Dec 05 08:19:30 crc kubenswrapper[4997]: I1205 08:19:30.739119 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cg92w" event={"ID":"895a6cef-c397-43cf-9aa3-6190b710b8c3","Type":"ContainerDied","Data":"352715e1ed6f48cd16e2433f52b6e0340d29edf8ef735d9220a9213cad76e961"} Dec 05 08:19:30 crc kubenswrapper[4997]: I1205 08:19:30.739195 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cg92w" event={"ID":"895a6cef-c397-43cf-9aa3-6190b710b8c3","Type":"ContainerStarted","Data":"a938b5f4da626aec15849d36e680dd9cc3b5190e99d2220ddc743d6529b20f2d"} Dec 05 08:19:31 crc kubenswrapper[4997]: I1205 08:19:31.758009 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cg92w" event={"ID":"895a6cef-c397-43cf-9aa3-6190b710b8c3","Type":"ContainerStarted","Data":"24a021caab09d8dc620ac7bf99416b6ab7cef6acfb150fd6510a508b979bbd03"} Dec 05 08:19:32 crc kubenswrapper[4997]: I1205 08:19:32.782281 4997 generic.go:334] "Generic (PLEG): container finished" podID="895a6cef-c397-43cf-9aa3-6190b710b8c3" containerID="24a021caab09d8dc620ac7bf99416b6ab7cef6acfb150fd6510a508b979bbd03" exitCode=0 Dec 05 08:19:32 crc kubenswrapper[4997]: I1205 08:19:32.782644 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cg92w" event={"ID":"895a6cef-c397-43cf-9aa3-6190b710b8c3","Type":"ContainerDied","Data":"24a021caab09d8dc620ac7bf99416b6ab7cef6acfb150fd6510a508b979bbd03"} Dec 05 08:19:33 crc kubenswrapper[4997]: I1205 08:19:33.795814 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cg92w" event={"ID":"895a6cef-c397-43cf-9aa3-6190b710b8c3","Type":"ContainerStarted","Data":"ec64d23db402e17f3b0d5947b51805029a3d5056ac4e03b8446786c4661961c2"} Dec 05 08:19:38 crc kubenswrapper[4997]: I1205 08:19:38.496442 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cg92w" podStartSLOduration=6.907823072 podStartE2EDuration="9.496416834s" podCreationTimestamp="2025-12-05 08:19:29 +0000 UTC" firstStartedPulling="2025-12-05 08:19:30.741530193 +0000 UTC m=+5071.270437474" lastFinishedPulling="2025-12-05 08:19:33.330123975 +0000 UTC m=+5073.859031236" observedRunningTime="2025-12-05 08:19:33.816171248 +0000 UTC m=+5074.345078509" watchObservedRunningTime="2025-12-05 08:19:38.496416834 +0000 UTC m=+5079.025324095" Dec 05 08:19:38 crc kubenswrapper[4997]: I1205 08:19:38.500943 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vh2cc"] Dec 05 08:19:38 crc kubenswrapper[4997]: I1205 08:19:38.502390 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vh2cc" Dec 05 08:19:38 crc kubenswrapper[4997]: I1205 08:19:38.514908 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vh2cc"] Dec 05 08:19:38 crc kubenswrapper[4997]: I1205 08:19:38.662585 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b484f2b9-cbe0-46de-b352-3102de92627c-utilities\") pod \"redhat-marketplace-vh2cc\" (UID: \"b484f2b9-cbe0-46de-b352-3102de92627c\") " pod="openshift-marketplace/redhat-marketplace-vh2cc" Dec 05 08:19:38 crc kubenswrapper[4997]: I1205 08:19:38.662855 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b484f2b9-cbe0-46de-b352-3102de92627c-catalog-content\") pod \"redhat-marketplace-vh2cc\" (UID: \"b484f2b9-cbe0-46de-b352-3102de92627c\") " pod="openshift-marketplace/redhat-marketplace-vh2cc" Dec 05 08:19:38 crc kubenswrapper[4997]: I1205 08:19:38.662926 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45bc2\" (UniqueName: \"kubernetes.io/projected/b484f2b9-cbe0-46de-b352-3102de92627c-kube-api-access-45bc2\") pod \"redhat-marketplace-vh2cc\" (UID: \"b484f2b9-cbe0-46de-b352-3102de92627c\") " pod="openshift-marketplace/redhat-marketplace-vh2cc" Dec 05 08:19:38 crc kubenswrapper[4997]: I1205 08:19:38.765474 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b484f2b9-cbe0-46de-b352-3102de92627c-catalog-content\") pod \"redhat-marketplace-vh2cc\" (UID: \"b484f2b9-cbe0-46de-b352-3102de92627c\") " pod="openshift-marketplace/redhat-marketplace-vh2cc" Dec 05 08:19:38 crc kubenswrapper[4997]: I1205 08:19:38.765545 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45bc2\" (UniqueName: \"kubernetes.io/projected/b484f2b9-cbe0-46de-b352-3102de92627c-kube-api-access-45bc2\") pod \"redhat-marketplace-vh2cc\" (UID: \"b484f2b9-cbe0-46de-b352-3102de92627c\") " pod="openshift-marketplace/redhat-marketplace-vh2cc" Dec 05 08:19:38 crc kubenswrapper[4997]: I1205 08:19:38.765662 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b484f2b9-cbe0-46de-b352-3102de92627c-utilities\") pod \"redhat-marketplace-vh2cc\" (UID: \"b484f2b9-cbe0-46de-b352-3102de92627c\") " pod="openshift-marketplace/redhat-marketplace-vh2cc" Dec 05 08:19:38 crc kubenswrapper[4997]: I1205 08:19:38.766076 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b484f2b9-cbe0-46de-b352-3102de92627c-catalog-content\") pod \"redhat-marketplace-vh2cc\" (UID: \"b484f2b9-cbe0-46de-b352-3102de92627c\") " pod="openshift-marketplace/redhat-marketplace-vh2cc" Dec 05 08:19:38 crc kubenswrapper[4997]: I1205 08:19:38.766219 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b484f2b9-cbe0-46de-b352-3102de92627c-utilities\") pod \"redhat-marketplace-vh2cc\" (UID: \"b484f2b9-cbe0-46de-b352-3102de92627c\") " pod="openshift-marketplace/redhat-marketplace-vh2cc" Dec 05 08:19:38 crc kubenswrapper[4997]: I1205 08:19:38.785461 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45bc2\" (UniqueName: \"kubernetes.io/projected/b484f2b9-cbe0-46de-b352-3102de92627c-kube-api-access-45bc2\") pod \"redhat-marketplace-vh2cc\" (UID: \"b484f2b9-cbe0-46de-b352-3102de92627c\") " pod="openshift-marketplace/redhat-marketplace-vh2cc" Dec 05 08:19:38 crc kubenswrapper[4997]: I1205 08:19:38.821959 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vh2cc" Dec 05 08:19:39 crc kubenswrapper[4997]: I1205 08:19:39.267203 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vh2cc"] Dec 05 08:19:39 crc kubenswrapper[4997]: I1205 08:19:39.475997 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cg92w" Dec 05 08:19:39 crc kubenswrapper[4997]: I1205 08:19:39.476755 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cg92w" Dec 05 08:19:39 crc kubenswrapper[4997]: I1205 08:19:39.514482 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cg92w" Dec 05 08:19:39 crc kubenswrapper[4997]: I1205 08:19:39.844741 4997 generic.go:334] "Generic (PLEG): container finished" podID="b484f2b9-cbe0-46de-b352-3102de92627c" containerID="6f6b5a73f5c5b27e789d576a6acd2d4ee28bde1123a4a9a765b09982e69d0d09" exitCode=0 Dec 05 08:19:39 crc kubenswrapper[4997]: I1205 08:19:39.844864 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vh2cc" event={"ID":"b484f2b9-cbe0-46de-b352-3102de92627c","Type":"ContainerDied","Data":"6f6b5a73f5c5b27e789d576a6acd2d4ee28bde1123a4a9a765b09982e69d0d09"} Dec 05 08:19:39 crc kubenswrapper[4997]: I1205 08:19:39.844984 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vh2cc" event={"ID":"b484f2b9-cbe0-46de-b352-3102de92627c","Type":"ContainerStarted","Data":"d61aac214790776a60a0d19cd1178dd5a8c2d2127ef695a433b6a9ea32a255f5"} Dec 05 08:19:39 crc kubenswrapper[4997]: I1205 08:19:39.931923 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cg92w" Dec 05 08:19:40 crc kubenswrapper[4997]: I1205 08:19:40.858019 4997 generic.go:334] "Generic (PLEG): container finished" podID="b484f2b9-cbe0-46de-b352-3102de92627c" containerID="c4f62013b92183c34b7b2f8ade3bf554f5b060d818683909c4b7c86347f333f9" exitCode=0 Dec 05 08:19:40 crc kubenswrapper[4997]: I1205 08:19:40.858074 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vh2cc" event={"ID":"b484f2b9-cbe0-46de-b352-3102de92627c","Type":"ContainerDied","Data":"c4f62013b92183c34b7b2f8ade3bf554f5b060d818683909c4b7c86347f333f9"} Dec 05 08:19:42 crc kubenswrapper[4997]: I1205 08:19:42.878244 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vh2cc" event={"ID":"b484f2b9-cbe0-46de-b352-3102de92627c","Type":"ContainerStarted","Data":"25f9e3875f04967702452e606da93a9a359181ac4aeed3cacb22160907945c14"} Dec 05 08:19:42 crc kubenswrapper[4997]: I1205 08:19:42.903940 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vh2cc" podStartSLOduration=2.937762356 podStartE2EDuration="4.903921618s" podCreationTimestamp="2025-12-05 08:19:38 +0000 UTC" firstStartedPulling="2025-12-05 08:19:39.847845045 +0000 UTC m=+5080.376752336" lastFinishedPulling="2025-12-05 08:19:41.814004297 +0000 UTC m=+5082.342911598" observedRunningTime="2025-12-05 08:19:42.899404646 +0000 UTC m=+5083.428311927" watchObservedRunningTime="2025-12-05 08:19:42.903921618 +0000 UTC m=+5083.432828879" Dec 05 08:19:44 crc kubenswrapper[4997]: I1205 08:19:44.489986 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cg92w"] Dec 05 08:19:44 crc kubenswrapper[4997]: I1205 08:19:44.490262 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cg92w" podUID="895a6cef-c397-43cf-9aa3-6190b710b8c3" containerName="registry-server" containerID="cri-o://ec64d23db402e17f3b0d5947b51805029a3d5056ac4e03b8446786c4661961c2" gracePeriod=2 Dec 05 08:19:45 crc kubenswrapper[4997]: I1205 08:19:45.904725 4997 generic.go:334] "Generic (PLEG): container finished" podID="895a6cef-c397-43cf-9aa3-6190b710b8c3" containerID="ec64d23db402e17f3b0d5947b51805029a3d5056ac4e03b8446786c4661961c2" exitCode=0 Dec 05 08:19:45 crc kubenswrapper[4997]: I1205 08:19:45.904764 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cg92w" event={"ID":"895a6cef-c397-43cf-9aa3-6190b710b8c3","Type":"ContainerDied","Data":"ec64d23db402e17f3b0d5947b51805029a3d5056ac4e03b8446786c4661961c2"} Dec 05 08:19:45 crc kubenswrapper[4997]: I1205 08:19:45.982128 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cg92w" Dec 05 08:19:46 crc kubenswrapper[4997]: I1205 08:19:46.182964 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/895a6cef-c397-43cf-9aa3-6190b710b8c3-utilities\") pod \"895a6cef-c397-43cf-9aa3-6190b710b8c3\" (UID: \"895a6cef-c397-43cf-9aa3-6190b710b8c3\") " Dec 05 08:19:46 crc kubenswrapper[4997]: I1205 08:19:46.183140 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/895a6cef-c397-43cf-9aa3-6190b710b8c3-catalog-content\") pod \"895a6cef-c397-43cf-9aa3-6190b710b8c3\" (UID: \"895a6cef-c397-43cf-9aa3-6190b710b8c3\") " Dec 05 08:19:46 crc kubenswrapper[4997]: I1205 08:19:46.183274 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfcmd\" (UniqueName: \"kubernetes.io/projected/895a6cef-c397-43cf-9aa3-6190b710b8c3-kube-api-access-pfcmd\") pod \"895a6cef-c397-43cf-9aa3-6190b710b8c3\" (UID: \"895a6cef-c397-43cf-9aa3-6190b710b8c3\") " Dec 05 08:19:46 crc kubenswrapper[4997]: I1205 08:19:46.186536 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/895a6cef-c397-43cf-9aa3-6190b710b8c3-utilities" (OuterVolumeSpecName: "utilities") pod "895a6cef-c397-43cf-9aa3-6190b710b8c3" (UID: "895a6cef-c397-43cf-9aa3-6190b710b8c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:19:46 crc kubenswrapper[4997]: I1205 08:19:46.191792 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/895a6cef-c397-43cf-9aa3-6190b710b8c3-kube-api-access-pfcmd" (OuterVolumeSpecName: "kube-api-access-pfcmd") pod "895a6cef-c397-43cf-9aa3-6190b710b8c3" (UID: "895a6cef-c397-43cf-9aa3-6190b710b8c3"). InnerVolumeSpecName "kube-api-access-pfcmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:19:46 crc kubenswrapper[4997]: I1205 08:19:46.285298 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfcmd\" (UniqueName: \"kubernetes.io/projected/895a6cef-c397-43cf-9aa3-6190b710b8c3-kube-api-access-pfcmd\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:46 crc kubenswrapper[4997]: I1205 08:19:46.285387 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/895a6cef-c397-43cf-9aa3-6190b710b8c3-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:46 crc kubenswrapper[4997]: I1205 08:19:46.317120 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/895a6cef-c397-43cf-9aa3-6190b710b8c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "895a6cef-c397-43cf-9aa3-6190b710b8c3" (UID: "895a6cef-c397-43cf-9aa3-6190b710b8c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:19:46 crc kubenswrapper[4997]: I1205 08:19:46.387588 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/895a6cef-c397-43cf-9aa3-6190b710b8c3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:46 crc kubenswrapper[4997]: I1205 08:19:46.920275 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cg92w" event={"ID":"895a6cef-c397-43cf-9aa3-6190b710b8c3","Type":"ContainerDied","Data":"a938b5f4da626aec15849d36e680dd9cc3b5190e99d2220ddc743d6529b20f2d"} Dec 05 08:19:46 crc kubenswrapper[4997]: I1205 08:19:46.920329 4997 scope.go:117] "RemoveContainer" containerID="ec64d23db402e17f3b0d5947b51805029a3d5056ac4e03b8446786c4661961c2" Dec 05 08:19:46 crc kubenswrapper[4997]: I1205 08:19:46.920479 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cg92w" Dec 05 08:19:46 crc kubenswrapper[4997]: I1205 08:19:46.958728 4997 scope.go:117] "RemoveContainer" containerID="24a021caab09d8dc620ac7bf99416b6ab7cef6acfb150fd6510a508b979bbd03" Dec 05 08:19:46 crc kubenswrapper[4997]: I1205 08:19:46.973235 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cg92w"] Dec 05 08:19:46 crc kubenswrapper[4997]: I1205 08:19:46.979832 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cg92w"] Dec 05 08:19:46 crc kubenswrapper[4997]: I1205 08:19:46.990141 4997 scope.go:117] "RemoveContainer" containerID="352715e1ed6f48cd16e2433f52b6e0340d29edf8ef735d9220a9213cad76e961" Dec 05 08:19:47 crc kubenswrapper[4997]: I1205 08:19:47.759110 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="895a6cef-c397-43cf-9aa3-6190b710b8c3" path="/var/lib/kubelet/pods/895a6cef-c397-43cf-9aa3-6190b710b8c3/volumes" Dec 05 08:19:48 crc kubenswrapper[4997]: I1205 08:19:48.822937 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vh2cc" Dec 05 08:19:48 crc kubenswrapper[4997]: I1205 08:19:48.822988 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vh2cc" Dec 05 08:19:48 crc kubenswrapper[4997]: I1205 08:19:48.886400 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vh2cc" Dec 05 08:19:48 crc kubenswrapper[4997]: I1205 08:19:48.985595 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vh2cc" Dec 05 08:19:49 crc kubenswrapper[4997]: I1205 08:19:49.289940 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vh2cc"] Dec 05 08:19:50 crc kubenswrapper[4997]: I1205 08:19:50.949301 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vh2cc" podUID="b484f2b9-cbe0-46de-b352-3102de92627c" containerName="registry-server" containerID="cri-o://25f9e3875f04967702452e606da93a9a359181ac4aeed3cacb22160907945c14" gracePeriod=2 Dec 05 08:19:51 crc kubenswrapper[4997]: I1205 08:19:51.812534 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vh2cc" Dec 05 08:19:51 crc kubenswrapper[4997]: I1205 08:19:51.959130 4997 generic.go:334] "Generic (PLEG): container finished" podID="b484f2b9-cbe0-46de-b352-3102de92627c" containerID="25f9e3875f04967702452e606da93a9a359181ac4aeed3cacb22160907945c14" exitCode=0 Dec 05 08:19:51 crc kubenswrapper[4997]: I1205 08:19:51.959174 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vh2cc" event={"ID":"b484f2b9-cbe0-46de-b352-3102de92627c","Type":"ContainerDied","Data":"25f9e3875f04967702452e606da93a9a359181ac4aeed3cacb22160907945c14"} Dec 05 08:19:51 crc kubenswrapper[4997]: I1205 08:19:51.959191 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vh2cc" Dec 05 08:19:51 crc kubenswrapper[4997]: I1205 08:19:51.959203 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vh2cc" event={"ID":"b484f2b9-cbe0-46de-b352-3102de92627c","Type":"ContainerDied","Data":"d61aac214790776a60a0d19cd1178dd5a8c2d2127ef695a433b6a9ea32a255f5"} Dec 05 08:19:51 crc kubenswrapper[4997]: I1205 08:19:51.959221 4997 scope.go:117] "RemoveContainer" containerID="25f9e3875f04967702452e606da93a9a359181ac4aeed3cacb22160907945c14" Dec 05 08:19:51 crc kubenswrapper[4997]: I1205 08:19:51.963307 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b484f2b9-cbe0-46de-b352-3102de92627c-catalog-content\") pod \"b484f2b9-cbe0-46de-b352-3102de92627c\" (UID: \"b484f2b9-cbe0-46de-b352-3102de92627c\") " Dec 05 08:19:51 crc kubenswrapper[4997]: I1205 08:19:51.963355 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b484f2b9-cbe0-46de-b352-3102de92627c-utilities\") pod \"b484f2b9-cbe0-46de-b352-3102de92627c\" (UID: \"b484f2b9-cbe0-46de-b352-3102de92627c\") " Dec 05 08:19:51 crc kubenswrapper[4997]: I1205 08:19:51.963496 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45bc2\" (UniqueName: \"kubernetes.io/projected/b484f2b9-cbe0-46de-b352-3102de92627c-kube-api-access-45bc2\") pod \"b484f2b9-cbe0-46de-b352-3102de92627c\" (UID: \"b484f2b9-cbe0-46de-b352-3102de92627c\") " Dec 05 08:19:51 crc kubenswrapper[4997]: I1205 08:19:51.965294 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b484f2b9-cbe0-46de-b352-3102de92627c-utilities" (OuterVolumeSpecName: "utilities") pod "b484f2b9-cbe0-46de-b352-3102de92627c" (UID: "b484f2b9-cbe0-46de-b352-3102de92627c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:19:51 crc kubenswrapper[4997]: I1205 08:19:51.974509 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b484f2b9-cbe0-46de-b352-3102de92627c-kube-api-access-45bc2" (OuterVolumeSpecName: "kube-api-access-45bc2") pod "b484f2b9-cbe0-46de-b352-3102de92627c" (UID: "b484f2b9-cbe0-46de-b352-3102de92627c"). InnerVolumeSpecName "kube-api-access-45bc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:19:51 crc kubenswrapper[4997]: I1205 08:19:51.996470 4997 scope.go:117] "RemoveContainer" containerID="c4f62013b92183c34b7b2f8ade3bf554f5b060d818683909c4b7c86347f333f9" Dec 05 08:19:52 crc kubenswrapper[4997]: I1205 08:19:52.015162 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b484f2b9-cbe0-46de-b352-3102de92627c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b484f2b9-cbe0-46de-b352-3102de92627c" (UID: "b484f2b9-cbe0-46de-b352-3102de92627c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:19:52 crc kubenswrapper[4997]: I1205 08:19:52.044058 4997 scope.go:117] "RemoveContainer" containerID="6f6b5a73f5c5b27e789d576a6acd2d4ee28bde1123a4a9a765b09982e69d0d09" Dec 05 08:19:52 crc kubenswrapper[4997]: I1205 08:19:52.065465 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b484f2b9-cbe0-46de-b352-3102de92627c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:52 crc kubenswrapper[4997]: I1205 08:19:52.065509 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b484f2b9-cbe0-46de-b352-3102de92627c-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:52 crc kubenswrapper[4997]: I1205 08:19:52.065524 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45bc2\" (UniqueName: \"kubernetes.io/projected/b484f2b9-cbe0-46de-b352-3102de92627c-kube-api-access-45bc2\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:52 crc kubenswrapper[4997]: I1205 08:19:52.077298 4997 scope.go:117] "RemoveContainer" containerID="25f9e3875f04967702452e606da93a9a359181ac4aeed3cacb22160907945c14" Dec 05 08:19:52 crc kubenswrapper[4997]: E1205 08:19:52.077840 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25f9e3875f04967702452e606da93a9a359181ac4aeed3cacb22160907945c14\": container with ID starting with 25f9e3875f04967702452e606da93a9a359181ac4aeed3cacb22160907945c14 not found: ID does not exist" containerID="25f9e3875f04967702452e606da93a9a359181ac4aeed3cacb22160907945c14" Dec 05 08:19:52 crc kubenswrapper[4997]: I1205 08:19:52.077867 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25f9e3875f04967702452e606da93a9a359181ac4aeed3cacb22160907945c14"} err="failed to get container status \"25f9e3875f04967702452e606da93a9a359181ac4aeed3cacb22160907945c14\": rpc error: code = NotFound desc = could not find container \"25f9e3875f04967702452e606da93a9a359181ac4aeed3cacb22160907945c14\": container with ID starting with 25f9e3875f04967702452e606da93a9a359181ac4aeed3cacb22160907945c14 not found: ID does not exist" Dec 05 08:19:52 crc kubenswrapper[4997]: I1205 08:19:52.077887 4997 scope.go:117] "RemoveContainer" containerID="c4f62013b92183c34b7b2f8ade3bf554f5b060d818683909c4b7c86347f333f9" Dec 05 08:19:52 crc kubenswrapper[4997]: E1205 08:19:52.078923 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4f62013b92183c34b7b2f8ade3bf554f5b060d818683909c4b7c86347f333f9\": container with ID starting with c4f62013b92183c34b7b2f8ade3bf554f5b060d818683909c4b7c86347f333f9 not found: ID does not exist" containerID="c4f62013b92183c34b7b2f8ade3bf554f5b060d818683909c4b7c86347f333f9" Dec 05 08:19:52 crc kubenswrapper[4997]: I1205 08:19:52.078952 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4f62013b92183c34b7b2f8ade3bf554f5b060d818683909c4b7c86347f333f9"} err="failed to get container status \"c4f62013b92183c34b7b2f8ade3bf554f5b060d818683909c4b7c86347f333f9\": rpc error: code = NotFound desc = could not find container \"c4f62013b92183c34b7b2f8ade3bf554f5b060d818683909c4b7c86347f333f9\": container with ID starting with c4f62013b92183c34b7b2f8ade3bf554f5b060d818683909c4b7c86347f333f9 not found: ID does not exist" Dec 05 08:19:52 crc kubenswrapper[4997]: I1205 08:19:52.078967 4997 scope.go:117] "RemoveContainer" containerID="6f6b5a73f5c5b27e789d576a6acd2d4ee28bde1123a4a9a765b09982e69d0d09" Dec 05 08:19:52 crc kubenswrapper[4997]: E1205 08:19:52.079244 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f6b5a73f5c5b27e789d576a6acd2d4ee28bde1123a4a9a765b09982e69d0d09\": container with ID starting with 6f6b5a73f5c5b27e789d576a6acd2d4ee28bde1123a4a9a765b09982e69d0d09 not found: ID does not exist" containerID="6f6b5a73f5c5b27e789d576a6acd2d4ee28bde1123a4a9a765b09982e69d0d09" Dec 05 08:19:52 crc kubenswrapper[4997]: I1205 08:19:52.079267 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f6b5a73f5c5b27e789d576a6acd2d4ee28bde1123a4a9a765b09982e69d0d09"} err="failed to get container status \"6f6b5a73f5c5b27e789d576a6acd2d4ee28bde1123a4a9a765b09982e69d0d09\": rpc error: code = NotFound desc = could not find container \"6f6b5a73f5c5b27e789d576a6acd2d4ee28bde1123a4a9a765b09982e69d0d09\": container with ID starting with 6f6b5a73f5c5b27e789d576a6acd2d4ee28bde1123a4a9a765b09982e69d0d09 not found: ID does not exist" Dec 05 08:19:52 crc kubenswrapper[4997]: I1205 08:19:52.294087 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vh2cc"] Dec 05 08:19:52 crc kubenswrapper[4997]: I1205 08:19:52.299672 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vh2cc"] Dec 05 08:19:53 crc kubenswrapper[4997]: I1205 08:19:53.783258 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b484f2b9-cbe0-46de-b352-3102de92627c" path="/var/lib/kubelet/pods/b484f2b9-cbe0-46de-b352-3102de92627c/volumes" Dec 05 08:21:49 crc kubenswrapper[4997]: I1205 08:21:49.770400 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:21:49 crc kubenswrapper[4997]: I1205 08:21:49.771115 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:22:19 crc kubenswrapper[4997]: I1205 08:22:19.770161 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:22:19 crc kubenswrapper[4997]: I1205 08:22:19.770899 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:22:49 crc kubenswrapper[4997]: I1205 08:22:49.769804 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:22:49 crc kubenswrapper[4997]: I1205 08:22:49.770439 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:22:49 crc kubenswrapper[4997]: I1205 08:22:49.770747 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 08:22:49 crc kubenswrapper[4997]: I1205 08:22:49.771627 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:22:49 crc kubenswrapper[4997]: I1205 08:22:49.771690 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" gracePeriod=600 Dec 05 08:22:50 crc kubenswrapper[4997]: E1205 08:22:50.406418 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:22:50 crc kubenswrapper[4997]: I1205 08:22:50.717683 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" exitCode=0 Dec 05 08:22:50 crc kubenswrapper[4997]: I1205 08:22:50.717761 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e"} Dec 05 08:22:50 crc kubenswrapper[4997]: I1205 08:22:50.717840 4997 scope.go:117] "RemoveContainer" containerID="f3907db969deab75a3846872db0276a30995a72fb7e0b6fc2b81e6f61bfe6a31" Dec 05 08:22:50 crc kubenswrapper[4997]: I1205 08:22:50.718369 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:22:50 crc kubenswrapper[4997]: E1205 08:22:50.718629 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:23:02 crc kubenswrapper[4997]: I1205 08:23:02.748611 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:23:02 crc kubenswrapper[4997]: E1205 08:23:02.749393 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:23:12 crc kubenswrapper[4997]: I1205 08:23:12.166809 4997 scope.go:117] "RemoveContainer" containerID="c43e81763ef8c8f7acbe720f2018fc8a1e59e5da702b60feb0ee80dafe3a6ac4" Dec 05 08:23:12 crc kubenswrapper[4997]: I1205 08:23:12.186108 4997 scope.go:117] "RemoveContainer" containerID="55090ec00f8c1dc1694479b4d69ac97dede3c2de72c884963483364fdbe5668d" Dec 05 08:23:12 crc kubenswrapper[4997]: I1205 08:23:12.213035 4997 scope.go:117] "RemoveContainer" containerID="05b440cf37a3c7d281b7571f0ed8ebd4095cafccfc218b354576c41e8ebe5920" Dec 05 08:23:15 crc kubenswrapper[4997]: I1205 08:23:15.749035 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:23:15 crc kubenswrapper[4997]: E1205 08:23:15.749603 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:23:26 crc kubenswrapper[4997]: I1205 08:23:26.749901 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:23:26 crc kubenswrapper[4997]: E1205 08:23:26.751125 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:23:40 crc kubenswrapper[4997]: I1205 08:23:40.749204 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:23:40 crc kubenswrapper[4997]: E1205 08:23:40.750011 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:23:52 crc kubenswrapper[4997]: I1205 08:23:52.749638 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:23:52 crc kubenswrapper[4997]: E1205 08:23:52.750330 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:24:03 crc kubenswrapper[4997]: I1205 08:24:03.749732 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:24:03 crc kubenswrapper[4997]: E1205 08:24:03.750662 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:24:14 crc kubenswrapper[4997]: I1205 08:24:14.750581 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:24:14 crc kubenswrapper[4997]: E1205 08:24:14.751564 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:24:25 crc kubenswrapper[4997]: I1205 08:24:25.749197 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:24:25 crc kubenswrapper[4997]: E1205 08:24:25.750024 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:24:36 crc kubenswrapper[4997]: I1205 08:24:36.749280 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:24:36 crc kubenswrapper[4997]: E1205 08:24:36.750124 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:24:50 crc kubenswrapper[4997]: I1205 08:24:50.749100 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:24:50 crc kubenswrapper[4997]: E1205 08:24:50.750374 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:25:01 crc kubenswrapper[4997]: I1205 08:25:01.749880 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:25:01 crc kubenswrapper[4997]: E1205 08:25:01.751143 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:25:15 crc kubenswrapper[4997]: I1205 08:25:15.749282 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:25:15 crc kubenswrapper[4997]: E1205 08:25:15.750433 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:25:30 crc kubenswrapper[4997]: I1205 08:25:30.748775 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:25:30 crc kubenswrapper[4997]: E1205 08:25:30.749503 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.182380 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rqks4"] Dec 05 08:25:42 crc kubenswrapper[4997]: E1205 08:25:42.183435 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="895a6cef-c397-43cf-9aa3-6190b710b8c3" containerName="extract-utilities" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.183461 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="895a6cef-c397-43cf-9aa3-6190b710b8c3" containerName="extract-utilities" Dec 05 08:25:42 crc kubenswrapper[4997]: E1205 08:25:42.183483 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="895a6cef-c397-43cf-9aa3-6190b710b8c3" containerName="registry-server" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.183494 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="895a6cef-c397-43cf-9aa3-6190b710b8c3" containerName="registry-server" Dec 05 08:25:42 crc kubenswrapper[4997]: E1205 08:25:42.183512 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b484f2b9-cbe0-46de-b352-3102de92627c" containerName="extract-utilities" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.183522 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b484f2b9-cbe0-46de-b352-3102de92627c" containerName="extract-utilities" Dec 05 08:25:42 crc kubenswrapper[4997]: E1205 08:25:42.183539 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="895a6cef-c397-43cf-9aa3-6190b710b8c3" containerName="extract-content" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.183547 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="895a6cef-c397-43cf-9aa3-6190b710b8c3" containerName="extract-content" Dec 05 08:25:42 crc kubenswrapper[4997]: E1205 08:25:42.183567 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b484f2b9-cbe0-46de-b352-3102de92627c" containerName="registry-server" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.183575 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b484f2b9-cbe0-46de-b352-3102de92627c" containerName="registry-server" Dec 05 08:25:42 crc kubenswrapper[4997]: E1205 08:25:42.183594 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b484f2b9-cbe0-46de-b352-3102de92627c" containerName="extract-content" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.183603 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b484f2b9-cbe0-46de-b352-3102de92627c" containerName="extract-content" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.183877 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="b484f2b9-cbe0-46de-b352-3102de92627c" containerName="registry-server" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.183903 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="895a6cef-c397-43cf-9aa3-6190b710b8c3" containerName="registry-server" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.185670 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqks4" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.195610 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rqks4"] Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.259805 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b2bed97-1759-4cac-8df8-b994a162e81d-utilities\") pod \"community-operators-rqks4\" (UID: \"9b2bed97-1759-4cac-8df8-b994a162e81d\") " pod="openshift-marketplace/community-operators-rqks4" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.259890 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpjhs\" (UniqueName: \"kubernetes.io/projected/9b2bed97-1759-4cac-8df8-b994a162e81d-kube-api-access-tpjhs\") pod \"community-operators-rqks4\" (UID: \"9b2bed97-1759-4cac-8df8-b994a162e81d\") " pod="openshift-marketplace/community-operators-rqks4" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.259935 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b2bed97-1759-4cac-8df8-b994a162e81d-catalog-content\") pod \"community-operators-rqks4\" (UID: \"9b2bed97-1759-4cac-8df8-b994a162e81d\") " pod="openshift-marketplace/community-operators-rqks4" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.360531 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b2bed97-1759-4cac-8df8-b994a162e81d-catalog-content\") pod \"community-operators-rqks4\" (UID: \"9b2bed97-1759-4cac-8df8-b994a162e81d\") " pod="openshift-marketplace/community-operators-rqks4" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.360639 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b2bed97-1759-4cac-8df8-b994a162e81d-utilities\") pod \"community-operators-rqks4\" (UID: \"9b2bed97-1759-4cac-8df8-b994a162e81d\") " pod="openshift-marketplace/community-operators-rqks4" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.360689 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpjhs\" (UniqueName: \"kubernetes.io/projected/9b2bed97-1759-4cac-8df8-b994a162e81d-kube-api-access-tpjhs\") pod \"community-operators-rqks4\" (UID: \"9b2bed97-1759-4cac-8df8-b994a162e81d\") " pod="openshift-marketplace/community-operators-rqks4" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.361698 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b2bed97-1759-4cac-8df8-b994a162e81d-catalog-content\") pod \"community-operators-rqks4\" (UID: \"9b2bed97-1759-4cac-8df8-b994a162e81d\") " pod="openshift-marketplace/community-operators-rqks4" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.361924 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b2bed97-1759-4cac-8df8-b994a162e81d-utilities\") pod \"community-operators-rqks4\" (UID: \"9b2bed97-1759-4cac-8df8-b994a162e81d\") " pod="openshift-marketplace/community-operators-rqks4" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.388565 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpjhs\" (UniqueName: \"kubernetes.io/projected/9b2bed97-1759-4cac-8df8-b994a162e81d-kube-api-access-tpjhs\") pod \"community-operators-rqks4\" (UID: \"9b2bed97-1759-4cac-8df8-b994a162e81d\") " pod="openshift-marketplace/community-operators-rqks4" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.507972 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqks4" Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.831140 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rqks4"] Dec 05 08:25:42 crc kubenswrapper[4997]: I1205 08:25:42.987720 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqks4" event={"ID":"9b2bed97-1759-4cac-8df8-b994a162e81d","Type":"ContainerStarted","Data":"4b60713a0e3930baf3f6cd3f7fd16e9aea883475472092bedae50e7e2e11c165"} Dec 05 08:25:43 crc kubenswrapper[4997]: I1205 08:25:43.749468 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:25:43 crc kubenswrapper[4997]: E1205 08:25:43.749978 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:25:43 crc kubenswrapper[4997]: I1205 08:25:43.996981 4997 generic.go:334] "Generic (PLEG): container finished" podID="9b2bed97-1759-4cac-8df8-b994a162e81d" containerID="b0654f7782090ef561ea74fe9cf1d64066cd9055d6a898fe87d192b42ece9140" exitCode=0 Dec 05 08:25:43 crc kubenswrapper[4997]: I1205 08:25:43.997020 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqks4" event={"ID":"9b2bed97-1759-4cac-8df8-b994a162e81d","Type":"ContainerDied","Data":"b0654f7782090ef561ea74fe9cf1d64066cd9055d6a898fe87d192b42ece9140"} Dec 05 08:25:43 crc kubenswrapper[4997]: I1205 08:25:43.999089 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 08:25:45 crc kubenswrapper[4997]: E1205 08:25:45.519539 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b2bed97_1759_4cac_8df8_b994a162e81d.slice/crio-conmon-845b1e4b20a0cfeabef9a59a4f17dae25406486bb133c848bf0b8ad77414b35c.scope\": RecentStats: unable to find data in memory cache]" Dec 05 08:25:46 crc kubenswrapper[4997]: I1205 08:25:46.014840 4997 generic.go:334] "Generic (PLEG): container finished" podID="9b2bed97-1759-4cac-8df8-b994a162e81d" containerID="845b1e4b20a0cfeabef9a59a4f17dae25406486bb133c848bf0b8ad77414b35c" exitCode=0 Dec 05 08:25:46 crc kubenswrapper[4997]: I1205 08:25:46.014880 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqks4" event={"ID":"9b2bed97-1759-4cac-8df8-b994a162e81d","Type":"ContainerDied","Data":"845b1e4b20a0cfeabef9a59a4f17dae25406486bb133c848bf0b8ad77414b35c"} Dec 05 08:25:47 crc kubenswrapper[4997]: I1205 08:25:47.024402 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqks4" event={"ID":"9b2bed97-1759-4cac-8df8-b994a162e81d","Type":"ContainerStarted","Data":"d54d8716b91dd7333cdefb1a23c0665f65cfb8de41b710e0c1c8b747caa96188"} Dec 05 08:25:47 crc kubenswrapper[4997]: I1205 08:25:47.043059 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rqks4" podStartSLOduration=2.6434795380000002 podStartE2EDuration="5.043036823s" podCreationTimestamp="2025-12-05 08:25:42 +0000 UTC" firstStartedPulling="2025-12-05 08:25:43.998734995 +0000 UTC m=+5444.527642266" lastFinishedPulling="2025-12-05 08:25:46.39829229 +0000 UTC m=+5446.927199551" observedRunningTime="2025-12-05 08:25:47.042880989 +0000 UTC m=+5447.571788250" watchObservedRunningTime="2025-12-05 08:25:47.043036823 +0000 UTC m=+5447.571944084" Dec 05 08:25:52 crc kubenswrapper[4997]: I1205 08:25:52.509524 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rqks4" Dec 05 08:25:52 crc kubenswrapper[4997]: I1205 08:25:52.510163 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rqks4" Dec 05 08:25:52 crc kubenswrapper[4997]: I1205 08:25:52.553591 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rqks4" Dec 05 08:25:53 crc kubenswrapper[4997]: I1205 08:25:53.115601 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rqks4" Dec 05 08:25:53 crc kubenswrapper[4997]: I1205 08:25:53.175399 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rqks4"] Dec 05 08:25:55 crc kubenswrapper[4997]: I1205 08:25:55.084832 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rqks4" podUID="9b2bed97-1759-4cac-8df8-b994a162e81d" containerName="registry-server" containerID="cri-o://d54d8716b91dd7333cdefb1a23c0665f65cfb8de41b710e0c1c8b747caa96188" gracePeriod=2 Dec 05 08:25:55 crc kubenswrapper[4997]: I1205 08:25:55.468933 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqks4" Dec 05 08:25:55 crc kubenswrapper[4997]: I1205 08:25:55.567478 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b2bed97-1759-4cac-8df8-b994a162e81d-utilities\") pod \"9b2bed97-1759-4cac-8df8-b994a162e81d\" (UID: \"9b2bed97-1759-4cac-8df8-b994a162e81d\") " Dec 05 08:25:55 crc kubenswrapper[4997]: I1205 08:25:55.567569 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpjhs\" (UniqueName: \"kubernetes.io/projected/9b2bed97-1759-4cac-8df8-b994a162e81d-kube-api-access-tpjhs\") pod \"9b2bed97-1759-4cac-8df8-b994a162e81d\" (UID: \"9b2bed97-1759-4cac-8df8-b994a162e81d\") " Dec 05 08:25:55 crc kubenswrapper[4997]: I1205 08:25:55.567725 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b2bed97-1759-4cac-8df8-b994a162e81d-catalog-content\") pod \"9b2bed97-1759-4cac-8df8-b994a162e81d\" (UID: \"9b2bed97-1759-4cac-8df8-b994a162e81d\") " Dec 05 08:25:55 crc kubenswrapper[4997]: I1205 08:25:55.568774 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b2bed97-1759-4cac-8df8-b994a162e81d-utilities" (OuterVolumeSpecName: "utilities") pod "9b2bed97-1759-4cac-8df8-b994a162e81d" (UID: "9b2bed97-1759-4cac-8df8-b994a162e81d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:25:55 crc kubenswrapper[4997]: I1205 08:25:55.581216 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b2bed97-1759-4cac-8df8-b994a162e81d-kube-api-access-tpjhs" (OuterVolumeSpecName: "kube-api-access-tpjhs") pod "9b2bed97-1759-4cac-8df8-b994a162e81d" (UID: "9b2bed97-1759-4cac-8df8-b994a162e81d"). InnerVolumeSpecName "kube-api-access-tpjhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:25:55 crc kubenswrapper[4997]: I1205 08:25:55.635192 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b2bed97-1759-4cac-8df8-b994a162e81d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9b2bed97-1759-4cac-8df8-b994a162e81d" (UID: "9b2bed97-1759-4cac-8df8-b994a162e81d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:25:55 crc kubenswrapper[4997]: I1205 08:25:55.668821 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b2bed97-1759-4cac-8df8-b994a162e81d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:25:55 crc kubenswrapper[4997]: I1205 08:25:55.668869 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpjhs\" (UniqueName: \"kubernetes.io/projected/9b2bed97-1759-4cac-8df8-b994a162e81d-kube-api-access-tpjhs\") on node \"crc\" DevicePath \"\"" Dec 05 08:25:55 crc kubenswrapper[4997]: I1205 08:25:55.668884 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b2bed97-1759-4cac-8df8-b994a162e81d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:25:56 crc kubenswrapper[4997]: I1205 08:25:56.095168 4997 generic.go:334] "Generic (PLEG): container finished" podID="9b2bed97-1759-4cac-8df8-b994a162e81d" containerID="d54d8716b91dd7333cdefb1a23c0665f65cfb8de41b710e0c1c8b747caa96188" exitCode=0 Dec 05 08:25:56 crc kubenswrapper[4997]: I1205 08:25:56.095301 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqks4" Dec 05 08:25:56 crc kubenswrapper[4997]: I1205 08:25:56.095320 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqks4" event={"ID":"9b2bed97-1759-4cac-8df8-b994a162e81d","Type":"ContainerDied","Data":"d54d8716b91dd7333cdefb1a23c0665f65cfb8de41b710e0c1c8b747caa96188"} Dec 05 08:25:56 crc kubenswrapper[4997]: I1205 08:25:56.096391 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqks4" event={"ID":"9b2bed97-1759-4cac-8df8-b994a162e81d","Type":"ContainerDied","Data":"4b60713a0e3930baf3f6cd3f7fd16e9aea883475472092bedae50e7e2e11c165"} Dec 05 08:25:56 crc kubenswrapper[4997]: I1205 08:25:56.096419 4997 scope.go:117] "RemoveContainer" containerID="d54d8716b91dd7333cdefb1a23c0665f65cfb8de41b710e0c1c8b747caa96188" Dec 05 08:25:56 crc kubenswrapper[4997]: I1205 08:25:56.120551 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rqks4"] Dec 05 08:25:56 crc kubenswrapper[4997]: I1205 08:25:56.127457 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rqks4"] Dec 05 08:25:56 crc kubenswrapper[4997]: I1205 08:25:56.132207 4997 scope.go:117] "RemoveContainer" containerID="845b1e4b20a0cfeabef9a59a4f17dae25406486bb133c848bf0b8ad77414b35c" Dec 05 08:25:56 crc kubenswrapper[4997]: I1205 08:25:56.150188 4997 scope.go:117] "RemoveContainer" containerID="b0654f7782090ef561ea74fe9cf1d64066cd9055d6a898fe87d192b42ece9140" Dec 05 08:25:56 crc kubenswrapper[4997]: I1205 08:25:56.174799 4997 scope.go:117] "RemoveContainer" containerID="d54d8716b91dd7333cdefb1a23c0665f65cfb8de41b710e0c1c8b747caa96188" Dec 05 08:25:56 crc kubenswrapper[4997]: E1205 08:25:56.175312 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d54d8716b91dd7333cdefb1a23c0665f65cfb8de41b710e0c1c8b747caa96188\": container with ID starting with d54d8716b91dd7333cdefb1a23c0665f65cfb8de41b710e0c1c8b747caa96188 not found: ID does not exist" containerID="d54d8716b91dd7333cdefb1a23c0665f65cfb8de41b710e0c1c8b747caa96188" Dec 05 08:25:56 crc kubenswrapper[4997]: I1205 08:25:56.175352 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d54d8716b91dd7333cdefb1a23c0665f65cfb8de41b710e0c1c8b747caa96188"} err="failed to get container status \"d54d8716b91dd7333cdefb1a23c0665f65cfb8de41b710e0c1c8b747caa96188\": rpc error: code = NotFound desc = could not find container \"d54d8716b91dd7333cdefb1a23c0665f65cfb8de41b710e0c1c8b747caa96188\": container with ID starting with d54d8716b91dd7333cdefb1a23c0665f65cfb8de41b710e0c1c8b747caa96188 not found: ID does not exist" Dec 05 08:25:56 crc kubenswrapper[4997]: I1205 08:25:56.175374 4997 scope.go:117] "RemoveContainer" containerID="845b1e4b20a0cfeabef9a59a4f17dae25406486bb133c848bf0b8ad77414b35c" Dec 05 08:25:56 crc kubenswrapper[4997]: E1205 08:25:56.175681 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"845b1e4b20a0cfeabef9a59a4f17dae25406486bb133c848bf0b8ad77414b35c\": container with ID starting with 845b1e4b20a0cfeabef9a59a4f17dae25406486bb133c848bf0b8ad77414b35c not found: ID does not exist" containerID="845b1e4b20a0cfeabef9a59a4f17dae25406486bb133c848bf0b8ad77414b35c" Dec 05 08:25:56 crc kubenswrapper[4997]: I1205 08:25:56.175830 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"845b1e4b20a0cfeabef9a59a4f17dae25406486bb133c848bf0b8ad77414b35c"} err="failed to get container status \"845b1e4b20a0cfeabef9a59a4f17dae25406486bb133c848bf0b8ad77414b35c\": rpc error: code = NotFound desc = could not find container \"845b1e4b20a0cfeabef9a59a4f17dae25406486bb133c848bf0b8ad77414b35c\": container with ID starting with 845b1e4b20a0cfeabef9a59a4f17dae25406486bb133c848bf0b8ad77414b35c not found: ID does not exist" Dec 05 08:25:56 crc kubenswrapper[4997]: I1205 08:25:56.175852 4997 scope.go:117] "RemoveContainer" containerID="b0654f7782090ef561ea74fe9cf1d64066cd9055d6a898fe87d192b42ece9140" Dec 05 08:25:56 crc kubenswrapper[4997]: E1205 08:25:56.176332 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0654f7782090ef561ea74fe9cf1d64066cd9055d6a898fe87d192b42ece9140\": container with ID starting with b0654f7782090ef561ea74fe9cf1d64066cd9055d6a898fe87d192b42ece9140 not found: ID does not exist" containerID="b0654f7782090ef561ea74fe9cf1d64066cd9055d6a898fe87d192b42ece9140" Dec 05 08:25:56 crc kubenswrapper[4997]: I1205 08:25:56.176485 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0654f7782090ef561ea74fe9cf1d64066cd9055d6a898fe87d192b42ece9140"} err="failed to get container status \"b0654f7782090ef561ea74fe9cf1d64066cd9055d6a898fe87d192b42ece9140\": rpc error: code = NotFound desc = could not find container \"b0654f7782090ef561ea74fe9cf1d64066cd9055d6a898fe87d192b42ece9140\": container with ID starting with b0654f7782090ef561ea74fe9cf1d64066cd9055d6a898fe87d192b42ece9140 not found: ID does not exist" Dec 05 08:25:57 crc kubenswrapper[4997]: I1205 08:25:57.757464 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b2bed97-1759-4cac-8df8-b994a162e81d" path="/var/lib/kubelet/pods/9b2bed97-1759-4cac-8df8-b994a162e81d/volumes" Dec 05 08:25:58 crc kubenswrapper[4997]: I1205 08:25:58.749997 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:25:58 crc kubenswrapper[4997]: E1205 08:25:58.750547 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:26:13 crc kubenswrapper[4997]: I1205 08:26:13.750269 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:26:13 crc kubenswrapper[4997]: E1205 08:26:13.751156 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:26:25 crc kubenswrapper[4997]: I1205 08:26:25.748972 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:26:25 crc kubenswrapper[4997]: E1205 08:26:25.749712 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:26:36 crc kubenswrapper[4997]: I1205 08:26:36.749864 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:26:36 crc kubenswrapper[4997]: E1205 08:26:36.750766 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:26:48 crc kubenswrapper[4997]: I1205 08:26:48.749454 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:26:48 crc kubenswrapper[4997]: E1205 08:26:48.750223 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:26:59 crc kubenswrapper[4997]: I1205 08:26:59.758114 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:26:59 crc kubenswrapper[4997]: E1205 08:26:59.759572 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:27:11 crc kubenswrapper[4997]: I1205 08:27:11.748785 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:27:11 crc kubenswrapper[4997]: E1205 08:27:11.749476 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.191039 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j2nf6"] Dec 05 08:27:22 crc kubenswrapper[4997]: E1205 08:27:22.191696 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b2bed97-1759-4cac-8df8-b994a162e81d" containerName="extract-utilities" Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.191717 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b2bed97-1759-4cac-8df8-b994a162e81d" containerName="extract-utilities" Dec 05 08:27:22 crc kubenswrapper[4997]: E1205 08:27:22.191743 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b2bed97-1759-4cac-8df8-b994a162e81d" containerName="registry-server" Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.191749 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b2bed97-1759-4cac-8df8-b994a162e81d" containerName="registry-server" Dec 05 08:27:22 crc kubenswrapper[4997]: E1205 08:27:22.191759 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b2bed97-1759-4cac-8df8-b994a162e81d" containerName="extract-content" Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.191765 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b2bed97-1759-4cac-8df8-b994a162e81d" containerName="extract-content" Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.191902 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b2bed97-1759-4cac-8df8-b994a162e81d" containerName="registry-server" Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.193007 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j2nf6" Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.204813 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j2nf6"] Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.294697 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gp6s6\" (UniqueName: \"kubernetes.io/projected/55279eea-54cd-488c-acd8-41b1a86aefec-kube-api-access-gp6s6\") pod \"certified-operators-j2nf6\" (UID: \"55279eea-54cd-488c-acd8-41b1a86aefec\") " pod="openshift-marketplace/certified-operators-j2nf6" Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.294913 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55279eea-54cd-488c-acd8-41b1a86aefec-utilities\") pod \"certified-operators-j2nf6\" (UID: \"55279eea-54cd-488c-acd8-41b1a86aefec\") " pod="openshift-marketplace/certified-operators-j2nf6" Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.294984 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55279eea-54cd-488c-acd8-41b1a86aefec-catalog-content\") pod \"certified-operators-j2nf6\" (UID: \"55279eea-54cd-488c-acd8-41b1a86aefec\") " pod="openshift-marketplace/certified-operators-j2nf6" Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.396692 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gp6s6\" (UniqueName: \"kubernetes.io/projected/55279eea-54cd-488c-acd8-41b1a86aefec-kube-api-access-gp6s6\") pod \"certified-operators-j2nf6\" (UID: \"55279eea-54cd-488c-acd8-41b1a86aefec\") " pod="openshift-marketplace/certified-operators-j2nf6" Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.396769 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55279eea-54cd-488c-acd8-41b1a86aefec-utilities\") pod \"certified-operators-j2nf6\" (UID: \"55279eea-54cd-488c-acd8-41b1a86aefec\") " pod="openshift-marketplace/certified-operators-j2nf6" Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.396795 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55279eea-54cd-488c-acd8-41b1a86aefec-catalog-content\") pod \"certified-operators-j2nf6\" (UID: \"55279eea-54cd-488c-acd8-41b1a86aefec\") " pod="openshift-marketplace/certified-operators-j2nf6" Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.397314 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55279eea-54cd-488c-acd8-41b1a86aefec-catalog-content\") pod \"certified-operators-j2nf6\" (UID: \"55279eea-54cd-488c-acd8-41b1a86aefec\") " pod="openshift-marketplace/certified-operators-j2nf6" Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.397465 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55279eea-54cd-488c-acd8-41b1a86aefec-utilities\") pod \"certified-operators-j2nf6\" (UID: \"55279eea-54cd-488c-acd8-41b1a86aefec\") " pod="openshift-marketplace/certified-operators-j2nf6" Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.421378 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gp6s6\" (UniqueName: \"kubernetes.io/projected/55279eea-54cd-488c-acd8-41b1a86aefec-kube-api-access-gp6s6\") pod \"certified-operators-j2nf6\" (UID: \"55279eea-54cd-488c-acd8-41b1a86aefec\") " pod="openshift-marketplace/certified-operators-j2nf6" Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.522161 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j2nf6" Dec 05 08:27:22 crc kubenswrapper[4997]: I1205 08:27:22.814539 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j2nf6"] Dec 05 08:27:23 crc kubenswrapper[4997]: I1205 08:27:23.749040 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:27:23 crc kubenswrapper[4997]: E1205 08:27:23.749353 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:27:23 crc kubenswrapper[4997]: I1205 08:27:23.750261 4997 generic.go:334] "Generic (PLEG): container finished" podID="55279eea-54cd-488c-acd8-41b1a86aefec" containerID="57fdafe9557083088fcfb2589900f9858b66483e0793945d5b744f3fed4a2d14" exitCode=0 Dec 05 08:27:23 crc kubenswrapper[4997]: I1205 08:27:23.757948 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2nf6" event={"ID":"55279eea-54cd-488c-acd8-41b1a86aefec","Type":"ContainerDied","Data":"57fdafe9557083088fcfb2589900f9858b66483e0793945d5b744f3fed4a2d14"} Dec 05 08:27:23 crc kubenswrapper[4997]: I1205 08:27:23.757988 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2nf6" event={"ID":"55279eea-54cd-488c-acd8-41b1a86aefec","Type":"ContainerStarted","Data":"244fc608f6555d8e21f89b8b4682d23d43efc66c2c2749b6df53a0bc911de5cf"} Dec 05 08:27:24 crc kubenswrapper[4997]: I1205 08:27:24.760076 4997 generic.go:334] "Generic (PLEG): container finished" podID="55279eea-54cd-488c-acd8-41b1a86aefec" containerID="5e208cb25fe9bb5f2c84d87bf798f1ee9dd8804c54491f7e14944b67562e14b9" exitCode=0 Dec 05 08:27:24 crc kubenswrapper[4997]: I1205 08:27:24.760140 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2nf6" event={"ID":"55279eea-54cd-488c-acd8-41b1a86aefec","Type":"ContainerDied","Data":"5e208cb25fe9bb5f2c84d87bf798f1ee9dd8804c54491f7e14944b67562e14b9"} Dec 05 08:27:25 crc kubenswrapper[4997]: I1205 08:27:25.771332 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2nf6" event={"ID":"55279eea-54cd-488c-acd8-41b1a86aefec","Type":"ContainerStarted","Data":"a75da0bd55222e7945a42161095519c54b1d1b431e1894ffaeb53ea1b4f712b1"} Dec 05 08:27:25 crc kubenswrapper[4997]: I1205 08:27:25.795060 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j2nf6" podStartSLOduration=2.377722212 podStartE2EDuration="3.795038209s" podCreationTimestamp="2025-12-05 08:27:22 +0000 UTC" firstStartedPulling="2025-12-05 08:27:23.751344674 +0000 UTC m=+5544.280251935" lastFinishedPulling="2025-12-05 08:27:25.168660671 +0000 UTC m=+5545.697567932" observedRunningTime="2025-12-05 08:27:25.790485276 +0000 UTC m=+5546.319392557" watchObservedRunningTime="2025-12-05 08:27:25.795038209 +0000 UTC m=+5546.323945480" Dec 05 08:27:32 crc kubenswrapper[4997]: I1205 08:27:32.522766 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j2nf6" Dec 05 08:27:32 crc kubenswrapper[4997]: I1205 08:27:32.523356 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j2nf6" Dec 05 08:27:32 crc kubenswrapper[4997]: I1205 08:27:32.569599 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j2nf6" Dec 05 08:27:32 crc kubenswrapper[4997]: I1205 08:27:32.865219 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j2nf6" Dec 05 08:27:32 crc kubenswrapper[4997]: I1205 08:27:32.920140 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j2nf6"] Dec 05 08:27:35 crc kubenswrapper[4997]: I1205 08:27:35.143276 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j2nf6" podUID="55279eea-54cd-488c-acd8-41b1a86aefec" containerName="registry-server" containerID="cri-o://a75da0bd55222e7945a42161095519c54b1d1b431e1894ffaeb53ea1b4f712b1" gracePeriod=2 Dec 05 08:27:37 crc kubenswrapper[4997]: I1205 08:27:37.160077 4997 generic.go:334] "Generic (PLEG): container finished" podID="55279eea-54cd-488c-acd8-41b1a86aefec" containerID="a75da0bd55222e7945a42161095519c54b1d1b431e1894ffaeb53ea1b4f712b1" exitCode=0 Dec 05 08:27:37 crc kubenswrapper[4997]: I1205 08:27:37.160124 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2nf6" event={"ID":"55279eea-54cd-488c-acd8-41b1a86aefec","Type":"ContainerDied","Data":"a75da0bd55222e7945a42161095519c54b1d1b431e1894ffaeb53ea1b4f712b1"} Dec 05 08:27:37 crc kubenswrapper[4997]: I1205 08:27:37.358230 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j2nf6" Dec 05 08:27:37 crc kubenswrapper[4997]: I1205 08:27:37.409974 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55279eea-54cd-488c-acd8-41b1a86aefec-catalog-content\") pod \"55279eea-54cd-488c-acd8-41b1a86aefec\" (UID: \"55279eea-54cd-488c-acd8-41b1a86aefec\") " Dec 05 08:27:37 crc kubenswrapper[4997]: I1205 08:27:37.410109 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55279eea-54cd-488c-acd8-41b1a86aefec-utilities\") pod \"55279eea-54cd-488c-acd8-41b1a86aefec\" (UID: \"55279eea-54cd-488c-acd8-41b1a86aefec\") " Dec 05 08:27:37 crc kubenswrapper[4997]: I1205 08:27:37.410142 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gp6s6\" (UniqueName: \"kubernetes.io/projected/55279eea-54cd-488c-acd8-41b1a86aefec-kube-api-access-gp6s6\") pod \"55279eea-54cd-488c-acd8-41b1a86aefec\" (UID: \"55279eea-54cd-488c-acd8-41b1a86aefec\") " Dec 05 08:27:37 crc kubenswrapper[4997]: I1205 08:27:37.411513 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55279eea-54cd-488c-acd8-41b1a86aefec-utilities" (OuterVolumeSpecName: "utilities") pod "55279eea-54cd-488c-acd8-41b1a86aefec" (UID: "55279eea-54cd-488c-acd8-41b1a86aefec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:27:37 crc kubenswrapper[4997]: I1205 08:27:37.417472 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55279eea-54cd-488c-acd8-41b1a86aefec-kube-api-access-gp6s6" (OuterVolumeSpecName: "kube-api-access-gp6s6") pod "55279eea-54cd-488c-acd8-41b1a86aefec" (UID: "55279eea-54cd-488c-acd8-41b1a86aefec"). InnerVolumeSpecName "kube-api-access-gp6s6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:27:37 crc kubenswrapper[4997]: I1205 08:27:37.464896 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55279eea-54cd-488c-acd8-41b1a86aefec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55279eea-54cd-488c-acd8-41b1a86aefec" (UID: "55279eea-54cd-488c-acd8-41b1a86aefec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:27:37 crc kubenswrapper[4997]: I1205 08:27:37.511332 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55279eea-54cd-488c-acd8-41b1a86aefec-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:27:37 crc kubenswrapper[4997]: I1205 08:27:37.511375 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55279eea-54cd-488c-acd8-41b1a86aefec-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:27:37 crc kubenswrapper[4997]: I1205 08:27:37.511392 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gp6s6\" (UniqueName: \"kubernetes.io/projected/55279eea-54cd-488c-acd8-41b1a86aefec-kube-api-access-gp6s6\") on node \"crc\" DevicePath \"\"" Dec 05 08:27:38 crc kubenswrapper[4997]: I1205 08:27:38.169795 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2nf6" event={"ID":"55279eea-54cd-488c-acd8-41b1a86aefec","Type":"ContainerDied","Data":"244fc608f6555d8e21f89b8b4682d23d43efc66c2c2749b6df53a0bc911de5cf"} Dec 05 08:27:38 crc kubenswrapper[4997]: I1205 08:27:38.169863 4997 scope.go:117] "RemoveContainer" containerID="a75da0bd55222e7945a42161095519c54b1d1b431e1894ffaeb53ea1b4f712b1" Dec 05 08:27:38 crc kubenswrapper[4997]: I1205 08:27:38.169870 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j2nf6" Dec 05 08:27:38 crc kubenswrapper[4997]: I1205 08:27:38.194458 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j2nf6"] Dec 05 08:27:38 crc kubenswrapper[4997]: I1205 08:27:38.200211 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j2nf6"] Dec 05 08:27:38 crc kubenswrapper[4997]: I1205 08:27:38.201867 4997 scope.go:117] "RemoveContainer" containerID="5e208cb25fe9bb5f2c84d87bf798f1ee9dd8804c54491f7e14944b67562e14b9" Dec 05 08:27:38 crc kubenswrapper[4997]: I1205 08:27:38.222144 4997 scope.go:117] "RemoveContainer" containerID="57fdafe9557083088fcfb2589900f9858b66483e0793945d5b744f3fed4a2d14" Dec 05 08:27:38 crc kubenswrapper[4997]: I1205 08:27:38.748798 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:27:38 crc kubenswrapper[4997]: E1205 08:27:38.749125 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:27:39 crc kubenswrapper[4997]: I1205 08:27:39.759822 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55279eea-54cd-488c-acd8-41b1a86aefec" path="/var/lib/kubelet/pods/55279eea-54cd-488c-acd8-41b1a86aefec/volumes" Dec 05 08:27:52 crc kubenswrapper[4997]: I1205 08:27:52.749213 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:27:53 crc kubenswrapper[4997]: I1205 08:27:53.278903 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"896ded924d3256236025e635c3d3f7a52372f71081a01d64d6aec438295ddd47"} Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.345091 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-28x98"] Dec 05 08:29:39 crc kubenswrapper[4997]: E1205 08:29:39.346250 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55279eea-54cd-488c-acd8-41b1a86aefec" containerName="extract-utilities" Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.346273 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="55279eea-54cd-488c-acd8-41b1a86aefec" containerName="extract-utilities" Dec 05 08:29:39 crc kubenswrapper[4997]: E1205 08:29:39.346284 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55279eea-54cd-488c-acd8-41b1a86aefec" containerName="extract-content" Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.346294 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="55279eea-54cd-488c-acd8-41b1a86aefec" containerName="extract-content" Dec 05 08:29:39 crc kubenswrapper[4997]: E1205 08:29:39.346310 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55279eea-54cd-488c-acd8-41b1a86aefec" containerName="registry-server" Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.346317 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="55279eea-54cd-488c-acd8-41b1a86aefec" containerName="registry-server" Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.346531 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="55279eea-54cd-488c-acd8-41b1a86aefec" containerName="registry-server" Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.347782 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-28x98" Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.356885 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-28x98"] Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.453151 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54081bb2-921c-4edd-a300-e19114e2ce0a-catalog-content\") pod \"redhat-operators-28x98\" (UID: \"54081bb2-921c-4edd-a300-e19114e2ce0a\") " pod="openshift-marketplace/redhat-operators-28x98" Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.453242 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42s42\" (UniqueName: \"kubernetes.io/projected/54081bb2-921c-4edd-a300-e19114e2ce0a-kube-api-access-42s42\") pod \"redhat-operators-28x98\" (UID: \"54081bb2-921c-4edd-a300-e19114e2ce0a\") " pod="openshift-marketplace/redhat-operators-28x98" Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.453293 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54081bb2-921c-4edd-a300-e19114e2ce0a-utilities\") pod \"redhat-operators-28x98\" (UID: \"54081bb2-921c-4edd-a300-e19114e2ce0a\") " pod="openshift-marketplace/redhat-operators-28x98" Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.555097 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54081bb2-921c-4edd-a300-e19114e2ce0a-catalog-content\") pod \"redhat-operators-28x98\" (UID: \"54081bb2-921c-4edd-a300-e19114e2ce0a\") " pod="openshift-marketplace/redhat-operators-28x98" Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.555203 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42s42\" (UniqueName: \"kubernetes.io/projected/54081bb2-921c-4edd-a300-e19114e2ce0a-kube-api-access-42s42\") pod \"redhat-operators-28x98\" (UID: \"54081bb2-921c-4edd-a300-e19114e2ce0a\") " pod="openshift-marketplace/redhat-operators-28x98" Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.555248 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54081bb2-921c-4edd-a300-e19114e2ce0a-utilities\") pod \"redhat-operators-28x98\" (UID: \"54081bb2-921c-4edd-a300-e19114e2ce0a\") " pod="openshift-marketplace/redhat-operators-28x98" Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.555787 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54081bb2-921c-4edd-a300-e19114e2ce0a-utilities\") pod \"redhat-operators-28x98\" (UID: \"54081bb2-921c-4edd-a300-e19114e2ce0a\") " pod="openshift-marketplace/redhat-operators-28x98" Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.556117 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54081bb2-921c-4edd-a300-e19114e2ce0a-catalog-content\") pod \"redhat-operators-28x98\" (UID: \"54081bb2-921c-4edd-a300-e19114e2ce0a\") " pod="openshift-marketplace/redhat-operators-28x98" Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.577159 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42s42\" (UniqueName: \"kubernetes.io/projected/54081bb2-921c-4edd-a300-e19114e2ce0a-kube-api-access-42s42\") pod \"redhat-operators-28x98\" (UID: \"54081bb2-921c-4edd-a300-e19114e2ce0a\") " pod="openshift-marketplace/redhat-operators-28x98" Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.679425 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-28x98" Dec 05 08:29:39 crc kubenswrapper[4997]: I1205 08:29:39.899225 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-28x98"] Dec 05 08:29:40 crc kubenswrapper[4997]: I1205 08:29:40.079370 4997 generic.go:334] "Generic (PLEG): container finished" podID="54081bb2-921c-4edd-a300-e19114e2ce0a" containerID="79232dbfa2ddc72976402cf5d88f0ea557568be635788cc08e3ff06bd1fd232c" exitCode=0 Dec 05 08:29:40 crc kubenswrapper[4997]: I1205 08:29:40.079410 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-28x98" event={"ID":"54081bb2-921c-4edd-a300-e19114e2ce0a","Type":"ContainerDied","Data":"79232dbfa2ddc72976402cf5d88f0ea557568be635788cc08e3ff06bd1fd232c"} Dec 05 08:29:40 crc kubenswrapper[4997]: I1205 08:29:40.079435 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-28x98" event={"ID":"54081bb2-921c-4edd-a300-e19114e2ce0a","Type":"ContainerStarted","Data":"7dfc4b2f028b969dfdb29cf7f9bcfb7adc4248e1aceca54f97300e6ba93641a4"} Dec 05 08:29:41 crc kubenswrapper[4997]: I1205 08:29:41.086720 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-28x98" event={"ID":"54081bb2-921c-4edd-a300-e19114e2ce0a","Type":"ContainerStarted","Data":"37ee6f1ffd526878a59859c9b6fee2542a83edf9aa4b20ad6849ed888a90a3c7"} Dec 05 08:29:42 crc kubenswrapper[4997]: I1205 08:29:42.096699 4997 generic.go:334] "Generic (PLEG): container finished" podID="54081bb2-921c-4edd-a300-e19114e2ce0a" containerID="37ee6f1ffd526878a59859c9b6fee2542a83edf9aa4b20ad6849ed888a90a3c7" exitCode=0 Dec 05 08:29:42 crc kubenswrapper[4997]: I1205 08:29:42.096776 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-28x98" event={"ID":"54081bb2-921c-4edd-a300-e19114e2ce0a","Type":"ContainerDied","Data":"37ee6f1ffd526878a59859c9b6fee2542a83edf9aa4b20ad6849ed888a90a3c7"} Dec 05 08:29:43 crc kubenswrapper[4997]: I1205 08:29:43.107411 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-28x98" event={"ID":"54081bb2-921c-4edd-a300-e19114e2ce0a","Type":"ContainerStarted","Data":"eef350e763e326fcddccfb2eaaf3364a95afe7750add34ec7c55850d32437bac"} Dec 05 08:29:43 crc kubenswrapper[4997]: I1205 08:29:43.124809 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-28x98" podStartSLOduration=1.682715508 podStartE2EDuration="4.124787881s" podCreationTimestamp="2025-12-05 08:29:39 +0000 UTC" firstStartedPulling="2025-12-05 08:29:40.080996575 +0000 UTC m=+5680.609903846" lastFinishedPulling="2025-12-05 08:29:42.523068958 +0000 UTC m=+5683.051976219" observedRunningTime="2025-12-05 08:29:43.123819115 +0000 UTC m=+5683.652726386" watchObservedRunningTime="2025-12-05 08:29:43.124787881 +0000 UTC m=+5683.653695152" Dec 05 08:29:49 crc kubenswrapper[4997]: I1205 08:29:49.680355 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-28x98" Dec 05 08:29:49 crc kubenswrapper[4997]: I1205 08:29:49.680995 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-28x98" Dec 05 08:29:49 crc kubenswrapper[4997]: I1205 08:29:49.730331 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-28x98" Dec 05 08:29:50 crc kubenswrapper[4997]: I1205 08:29:50.199179 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-28x98" Dec 05 08:29:50 crc kubenswrapper[4997]: I1205 08:29:50.243130 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-28x98"] Dec 05 08:29:52 crc kubenswrapper[4997]: I1205 08:29:52.174391 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-28x98" podUID="54081bb2-921c-4edd-a300-e19114e2ce0a" containerName="registry-server" containerID="cri-o://eef350e763e326fcddccfb2eaaf3364a95afe7750add34ec7c55850d32437bac" gracePeriod=2 Dec 05 08:29:54 crc kubenswrapper[4997]: I1205 08:29:54.596169 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-28x98" Dec 05 08:29:54 crc kubenswrapper[4997]: I1205 08:29:54.708323 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42s42\" (UniqueName: \"kubernetes.io/projected/54081bb2-921c-4edd-a300-e19114e2ce0a-kube-api-access-42s42\") pod \"54081bb2-921c-4edd-a300-e19114e2ce0a\" (UID: \"54081bb2-921c-4edd-a300-e19114e2ce0a\") " Dec 05 08:29:54 crc kubenswrapper[4997]: I1205 08:29:54.708457 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54081bb2-921c-4edd-a300-e19114e2ce0a-utilities\") pod \"54081bb2-921c-4edd-a300-e19114e2ce0a\" (UID: \"54081bb2-921c-4edd-a300-e19114e2ce0a\") " Dec 05 08:29:54 crc kubenswrapper[4997]: I1205 08:29:54.708576 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54081bb2-921c-4edd-a300-e19114e2ce0a-catalog-content\") pod \"54081bb2-921c-4edd-a300-e19114e2ce0a\" (UID: \"54081bb2-921c-4edd-a300-e19114e2ce0a\") " Dec 05 08:29:54 crc kubenswrapper[4997]: I1205 08:29:54.710530 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54081bb2-921c-4edd-a300-e19114e2ce0a-utilities" (OuterVolumeSpecName: "utilities") pod "54081bb2-921c-4edd-a300-e19114e2ce0a" (UID: "54081bb2-921c-4edd-a300-e19114e2ce0a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:29:54 crc kubenswrapper[4997]: I1205 08:29:54.714465 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54081bb2-921c-4edd-a300-e19114e2ce0a-kube-api-access-42s42" (OuterVolumeSpecName: "kube-api-access-42s42") pod "54081bb2-921c-4edd-a300-e19114e2ce0a" (UID: "54081bb2-921c-4edd-a300-e19114e2ce0a"). InnerVolumeSpecName "kube-api-access-42s42". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:29:54 crc kubenswrapper[4997]: I1205 08:29:54.810912 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42s42\" (UniqueName: \"kubernetes.io/projected/54081bb2-921c-4edd-a300-e19114e2ce0a-kube-api-access-42s42\") on node \"crc\" DevicePath \"\"" Dec 05 08:29:54 crc kubenswrapper[4997]: I1205 08:29:54.810954 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54081bb2-921c-4edd-a300-e19114e2ce0a-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:29:54 crc kubenswrapper[4997]: I1205 08:29:54.822530 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54081bb2-921c-4edd-a300-e19114e2ce0a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54081bb2-921c-4edd-a300-e19114e2ce0a" (UID: "54081bb2-921c-4edd-a300-e19114e2ce0a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:29:54 crc kubenswrapper[4997]: I1205 08:29:54.912715 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54081bb2-921c-4edd-a300-e19114e2ce0a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:29:55 crc kubenswrapper[4997]: I1205 08:29:55.200976 4997 generic.go:334] "Generic (PLEG): container finished" podID="54081bb2-921c-4edd-a300-e19114e2ce0a" containerID="eef350e763e326fcddccfb2eaaf3364a95afe7750add34ec7c55850d32437bac" exitCode=0 Dec 05 08:29:55 crc kubenswrapper[4997]: I1205 08:29:55.201026 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-28x98" event={"ID":"54081bb2-921c-4edd-a300-e19114e2ce0a","Type":"ContainerDied","Data":"eef350e763e326fcddccfb2eaaf3364a95afe7750add34ec7c55850d32437bac"} Dec 05 08:29:55 crc kubenswrapper[4997]: I1205 08:29:55.201056 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-28x98" event={"ID":"54081bb2-921c-4edd-a300-e19114e2ce0a","Type":"ContainerDied","Data":"7dfc4b2f028b969dfdb29cf7f9bcfb7adc4248e1aceca54f97300e6ba93641a4"} Dec 05 08:29:55 crc kubenswrapper[4997]: I1205 08:29:55.201076 4997 scope.go:117] "RemoveContainer" containerID="eef350e763e326fcddccfb2eaaf3364a95afe7750add34ec7c55850d32437bac" Dec 05 08:29:55 crc kubenswrapper[4997]: I1205 08:29:55.201216 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-28x98" Dec 05 08:29:55 crc kubenswrapper[4997]: I1205 08:29:55.220306 4997 scope.go:117] "RemoveContainer" containerID="37ee6f1ffd526878a59859c9b6fee2542a83edf9aa4b20ad6849ed888a90a3c7" Dec 05 08:29:55 crc kubenswrapper[4997]: I1205 08:29:55.236966 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-28x98"] Dec 05 08:29:55 crc kubenswrapper[4997]: I1205 08:29:55.243250 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-28x98"] Dec 05 08:29:55 crc kubenswrapper[4997]: I1205 08:29:55.255417 4997 scope.go:117] "RemoveContainer" containerID="79232dbfa2ddc72976402cf5d88f0ea557568be635788cc08e3ff06bd1fd232c" Dec 05 08:29:55 crc kubenswrapper[4997]: I1205 08:29:55.270709 4997 scope.go:117] "RemoveContainer" containerID="eef350e763e326fcddccfb2eaaf3364a95afe7750add34ec7c55850d32437bac" Dec 05 08:29:55 crc kubenswrapper[4997]: E1205 08:29:55.271250 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eef350e763e326fcddccfb2eaaf3364a95afe7750add34ec7c55850d32437bac\": container with ID starting with eef350e763e326fcddccfb2eaaf3364a95afe7750add34ec7c55850d32437bac not found: ID does not exist" containerID="eef350e763e326fcddccfb2eaaf3364a95afe7750add34ec7c55850d32437bac" Dec 05 08:29:55 crc kubenswrapper[4997]: I1205 08:29:55.271311 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eef350e763e326fcddccfb2eaaf3364a95afe7750add34ec7c55850d32437bac"} err="failed to get container status \"eef350e763e326fcddccfb2eaaf3364a95afe7750add34ec7c55850d32437bac\": rpc error: code = NotFound desc = could not find container \"eef350e763e326fcddccfb2eaaf3364a95afe7750add34ec7c55850d32437bac\": container with ID starting with eef350e763e326fcddccfb2eaaf3364a95afe7750add34ec7c55850d32437bac not found: ID does not exist" Dec 05 08:29:55 crc kubenswrapper[4997]: I1205 08:29:55.271361 4997 scope.go:117] "RemoveContainer" containerID="37ee6f1ffd526878a59859c9b6fee2542a83edf9aa4b20ad6849ed888a90a3c7" Dec 05 08:29:55 crc kubenswrapper[4997]: E1205 08:29:55.271789 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37ee6f1ffd526878a59859c9b6fee2542a83edf9aa4b20ad6849ed888a90a3c7\": container with ID starting with 37ee6f1ffd526878a59859c9b6fee2542a83edf9aa4b20ad6849ed888a90a3c7 not found: ID does not exist" containerID="37ee6f1ffd526878a59859c9b6fee2542a83edf9aa4b20ad6849ed888a90a3c7" Dec 05 08:29:55 crc kubenswrapper[4997]: I1205 08:29:55.271830 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37ee6f1ffd526878a59859c9b6fee2542a83edf9aa4b20ad6849ed888a90a3c7"} err="failed to get container status \"37ee6f1ffd526878a59859c9b6fee2542a83edf9aa4b20ad6849ed888a90a3c7\": rpc error: code = NotFound desc = could not find container \"37ee6f1ffd526878a59859c9b6fee2542a83edf9aa4b20ad6849ed888a90a3c7\": container with ID starting with 37ee6f1ffd526878a59859c9b6fee2542a83edf9aa4b20ad6849ed888a90a3c7 not found: ID does not exist" Dec 05 08:29:55 crc kubenswrapper[4997]: I1205 08:29:55.271861 4997 scope.go:117] "RemoveContainer" containerID="79232dbfa2ddc72976402cf5d88f0ea557568be635788cc08e3ff06bd1fd232c" Dec 05 08:29:55 crc kubenswrapper[4997]: E1205 08:29:55.272163 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79232dbfa2ddc72976402cf5d88f0ea557568be635788cc08e3ff06bd1fd232c\": container with ID starting with 79232dbfa2ddc72976402cf5d88f0ea557568be635788cc08e3ff06bd1fd232c not found: ID does not exist" containerID="79232dbfa2ddc72976402cf5d88f0ea557568be635788cc08e3ff06bd1fd232c" Dec 05 08:29:55 crc kubenswrapper[4997]: I1205 08:29:55.272205 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79232dbfa2ddc72976402cf5d88f0ea557568be635788cc08e3ff06bd1fd232c"} err="failed to get container status \"79232dbfa2ddc72976402cf5d88f0ea557568be635788cc08e3ff06bd1fd232c\": rpc error: code = NotFound desc = could not find container \"79232dbfa2ddc72976402cf5d88f0ea557568be635788cc08e3ff06bd1fd232c\": container with ID starting with 79232dbfa2ddc72976402cf5d88f0ea557568be635788cc08e3ff06bd1fd232c not found: ID does not exist" Dec 05 08:29:55 crc kubenswrapper[4997]: I1205 08:29:55.764883 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54081bb2-921c-4edd-a300-e19114e2ce0a" path="/var/lib/kubelet/pods/54081bb2-921c-4edd-a300-e19114e2ce0a/volumes" Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.124031 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mghch"] Dec 05 08:29:57 crc kubenswrapper[4997]: E1205 08:29:57.125096 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54081bb2-921c-4edd-a300-e19114e2ce0a" containerName="registry-server" Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.125143 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="54081bb2-921c-4edd-a300-e19114e2ce0a" containerName="registry-server" Dec 05 08:29:57 crc kubenswrapper[4997]: E1205 08:29:57.125213 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54081bb2-921c-4edd-a300-e19114e2ce0a" containerName="extract-utilities" Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.125232 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="54081bb2-921c-4edd-a300-e19114e2ce0a" containerName="extract-utilities" Dec 05 08:29:57 crc kubenswrapper[4997]: E1205 08:29:57.125255 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54081bb2-921c-4edd-a300-e19114e2ce0a" containerName="extract-content" Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.125271 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="54081bb2-921c-4edd-a300-e19114e2ce0a" containerName="extract-content" Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.125579 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="54081bb2-921c-4edd-a300-e19114e2ce0a" containerName="registry-server" Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.128177 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mghch" Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.132756 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mghch"] Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.142660 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7d5c257-54e0-4739-921b-805de66493c1-utilities\") pod \"redhat-marketplace-mghch\" (UID: \"a7d5c257-54e0-4739-921b-805de66493c1\") " pod="openshift-marketplace/redhat-marketplace-mghch" Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.142799 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7d5c257-54e0-4739-921b-805de66493c1-catalog-content\") pod \"redhat-marketplace-mghch\" (UID: \"a7d5c257-54e0-4739-921b-805de66493c1\") " pod="openshift-marketplace/redhat-marketplace-mghch" Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.244547 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7d5c257-54e0-4739-921b-805de66493c1-utilities\") pod \"redhat-marketplace-mghch\" (UID: \"a7d5c257-54e0-4739-921b-805de66493c1\") " pod="openshift-marketplace/redhat-marketplace-mghch" Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.244584 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7d5c257-54e0-4739-921b-805de66493c1-catalog-content\") pod \"redhat-marketplace-mghch\" (UID: \"a7d5c257-54e0-4739-921b-805de66493c1\") " pod="openshift-marketplace/redhat-marketplace-mghch" Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.244742 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tw52\" (UniqueName: \"kubernetes.io/projected/a7d5c257-54e0-4739-921b-805de66493c1-kube-api-access-5tw52\") pod \"redhat-marketplace-mghch\" (UID: \"a7d5c257-54e0-4739-921b-805de66493c1\") " pod="openshift-marketplace/redhat-marketplace-mghch" Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.245128 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7d5c257-54e0-4739-921b-805de66493c1-catalog-content\") pod \"redhat-marketplace-mghch\" (UID: \"a7d5c257-54e0-4739-921b-805de66493c1\") " pod="openshift-marketplace/redhat-marketplace-mghch" Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.245139 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7d5c257-54e0-4739-921b-805de66493c1-utilities\") pod \"redhat-marketplace-mghch\" (UID: \"a7d5c257-54e0-4739-921b-805de66493c1\") " pod="openshift-marketplace/redhat-marketplace-mghch" Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.346568 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tw52\" (UniqueName: \"kubernetes.io/projected/a7d5c257-54e0-4739-921b-805de66493c1-kube-api-access-5tw52\") pod \"redhat-marketplace-mghch\" (UID: \"a7d5c257-54e0-4739-921b-805de66493c1\") " pod="openshift-marketplace/redhat-marketplace-mghch" Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.365425 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tw52\" (UniqueName: \"kubernetes.io/projected/a7d5c257-54e0-4739-921b-805de66493c1-kube-api-access-5tw52\") pod \"redhat-marketplace-mghch\" (UID: \"a7d5c257-54e0-4739-921b-805de66493c1\") " pod="openshift-marketplace/redhat-marketplace-mghch" Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.458285 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mghch" Dec 05 08:29:57 crc kubenswrapper[4997]: I1205 08:29:57.705798 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mghch"] Dec 05 08:29:58 crc kubenswrapper[4997]: I1205 08:29:58.222100 4997 generic.go:334] "Generic (PLEG): container finished" podID="a7d5c257-54e0-4739-921b-805de66493c1" containerID="f4f500cb927f0d9436dfe7171b6ff620455b6ea43caf0a9b2fde45bb63b247ed" exitCode=0 Dec 05 08:29:58 crc kubenswrapper[4997]: I1205 08:29:58.222148 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mghch" event={"ID":"a7d5c257-54e0-4739-921b-805de66493c1","Type":"ContainerDied","Data":"f4f500cb927f0d9436dfe7171b6ff620455b6ea43caf0a9b2fde45bb63b247ed"} Dec 05 08:29:58 crc kubenswrapper[4997]: I1205 08:29:58.222178 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mghch" event={"ID":"a7d5c257-54e0-4739-921b-805de66493c1","Type":"ContainerStarted","Data":"bf7914600164570af0d71bba42a10740043021a98d10f4a9fba68a6c50577c72"} Dec 05 08:29:59 crc kubenswrapper[4997]: I1205 08:29:59.230583 4997 generic.go:334] "Generic (PLEG): container finished" podID="a7d5c257-54e0-4739-921b-805de66493c1" containerID="ae7ebaa6d2e2c16897e0143e39788ceed2d30635cf3de4dce406658eb32a320e" exitCode=0 Dec 05 08:29:59 crc kubenswrapper[4997]: I1205 08:29:59.230652 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mghch" event={"ID":"a7d5c257-54e0-4739-921b-805de66493c1","Type":"ContainerDied","Data":"ae7ebaa6d2e2c16897e0143e39788ceed2d30635cf3de4dce406658eb32a320e"} Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.148654 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g"] Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.150039 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g" Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.152606 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.152860 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.160866 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g"] Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.239052 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mghch" event={"ID":"a7d5c257-54e0-4739-921b-805de66493c1","Type":"ContainerStarted","Data":"bbaeb04744b6900fc7aedb337a612440104dc5a048c57c8ca82fc2986e2e18fd"} Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.261499 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mghch" podStartSLOduration=1.835519962 podStartE2EDuration="3.261475771s" podCreationTimestamp="2025-12-05 08:29:57 +0000 UTC" firstStartedPulling="2025-12-05 08:29:58.223802949 +0000 UTC m=+5698.752710210" lastFinishedPulling="2025-12-05 08:29:59.649758738 +0000 UTC m=+5700.178666019" observedRunningTime="2025-12-05 08:30:00.258806539 +0000 UTC m=+5700.787713820" watchObservedRunningTime="2025-12-05 08:30:00.261475771 +0000 UTC m=+5700.790383052" Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.286167 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-secret-volume\") pod \"collect-profiles-29415390-2kb7g\" (UID: \"708bfbf7-22c5-4cbf-9dca-c7220fdd1615\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g" Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.286247 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-config-volume\") pod \"collect-profiles-29415390-2kb7g\" (UID: \"708bfbf7-22c5-4cbf-9dca-c7220fdd1615\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g" Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.286322 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbtkq\" (UniqueName: \"kubernetes.io/projected/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-kube-api-access-qbtkq\") pod \"collect-profiles-29415390-2kb7g\" (UID: \"708bfbf7-22c5-4cbf-9dca-c7220fdd1615\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g" Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.388223 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-secret-volume\") pod \"collect-profiles-29415390-2kb7g\" (UID: \"708bfbf7-22c5-4cbf-9dca-c7220fdd1615\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g" Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.388286 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-config-volume\") pod \"collect-profiles-29415390-2kb7g\" (UID: \"708bfbf7-22c5-4cbf-9dca-c7220fdd1615\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g" Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.388344 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbtkq\" (UniqueName: \"kubernetes.io/projected/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-kube-api-access-qbtkq\") pod \"collect-profiles-29415390-2kb7g\" (UID: \"708bfbf7-22c5-4cbf-9dca-c7220fdd1615\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g" Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.389561 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-config-volume\") pod \"collect-profiles-29415390-2kb7g\" (UID: \"708bfbf7-22c5-4cbf-9dca-c7220fdd1615\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g" Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.395479 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-secret-volume\") pod \"collect-profiles-29415390-2kb7g\" (UID: \"708bfbf7-22c5-4cbf-9dca-c7220fdd1615\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g" Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.406499 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbtkq\" (UniqueName: \"kubernetes.io/projected/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-kube-api-access-qbtkq\") pod \"collect-profiles-29415390-2kb7g\" (UID: \"708bfbf7-22c5-4cbf-9dca-c7220fdd1615\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g" Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.470878 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g" Dec 05 08:30:00 crc kubenswrapper[4997]: I1205 08:30:00.869726 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g"] Dec 05 08:30:00 crc kubenswrapper[4997]: W1205 08:30:00.874825 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod708bfbf7_22c5_4cbf_9dca_c7220fdd1615.slice/crio-740048f5105d0f67b04a833fe888c6200d2062ae50b61738dc15002aa9c62fca WatchSource:0}: Error finding container 740048f5105d0f67b04a833fe888c6200d2062ae50b61738dc15002aa9c62fca: Status 404 returned error can't find the container with id 740048f5105d0f67b04a833fe888c6200d2062ae50b61738dc15002aa9c62fca Dec 05 08:30:01 crc kubenswrapper[4997]: I1205 08:30:01.248022 4997 generic.go:334] "Generic (PLEG): container finished" podID="708bfbf7-22c5-4cbf-9dca-c7220fdd1615" containerID="100d4817715ec67fcf1a312c1b88613e5b24a5ce57475e06b138946098010fc0" exitCode=0 Dec 05 08:30:01 crc kubenswrapper[4997]: I1205 08:30:01.248219 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g" event={"ID":"708bfbf7-22c5-4cbf-9dca-c7220fdd1615","Type":"ContainerDied","Data":"100d4817715ec67fcf1a312c1b88613e5b24a5ce57475e06b138946098010fc0"} Dec 05 08:30:01 crc kubenswrapper[4997]: I1205 08:30:01.248304 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g" event={"ID":"708bfbf7-22c5-4cbf-9dca-c7220fdd1615","Type":"ContainerStarted","Data":"740048f5105d0f67b04a833fe888c6200d2062ae50b61738dc15002aa9c62fca"} Dec 05 08:30:02 crc kubenswrapper[4997]: I1205 08:30:02.506824 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g" Dec 05 08:30:02 crc kubenswrapper[4997]: I1205 08:30:02.619560 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbtkq\" (UniqueName: \"kubernetes.io/projected/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-kube-api-access-qbtkq\") pod \"708bfbf7-22c5-4cbf-9dca-c7220fdd1615\" (UID: \"708bfbf7-22c5-4cbf-9dca-c7220fdd1615\") " Dec 05 08:30:02 crc kubenswrapper[4997]: I1205 08:30:02.619716 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-config-volume\") pod \"708bfbf7-22c5-4cbf-9dca-c7220fdd1615\" (UID: \"708bfbf7-22c5-4cbf-9dca-c7220fdd1615\") " Dec 05 08:30:02 crc kubenswrapper[4997]: I1205 08:30:02.619764 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-secret-volume\") pod \"708bfbf7-22c5-4cbf-9dca-c7220fdd1615\" (UID: \"708bfbf7-22c5-4cbf-9dca-c7220fdd1615\") " Dec 05 08:30:02 crc kubenswrapper[4997]: I1205 08:30:02.620510 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-config-volume" (OuterVolumeSpecName: "config-volume") pod "708bfbf7-22c5-4cbf-9dca-c7220fdd1615" (UID: "708bfbf7-22c5-4cbf-9dca-c7220fdd1615"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:30:02 crc kubenswrapper[4997]: I1205 08:30:02.625065 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "708bfbf7-22c5-4cbf-9dca-c7220fdd1615" (UID: "708bfbf7-22c5-4cbf-9dca-c7220fdd1615"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:30:02 crc kubenswrapper[4997]: I1205 08:30:02.626693 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-kube-api-access-qbtkq" (OuterVolumeSpecName: "kube-api-access-qbtkq") pod "708bfbf7-22c5-4cbf-9dca-c7220fdd1615" (UID: "708bfbf7-22c5-4cbf-9dca-c7220fdd1615"). InnerVolumeSpecName "kube-api-access-qbtkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:30:02 crc kubenswrapper[4997]: I1205 08:30:02.721332 4997 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:02 crc kubenswrapper[4997]: I1205 08:30:02.721378 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbtkq\" (UniqueName: \"kubernetes.io/projected/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-kube-api-access-qbtkq\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:02 crc kubenswrapper[4997]: I1205 08:30:02.721391 4997 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/708bfbf7-22c5-4cbf-9dca-c7220fdd1615-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:03 crc kubenswrapper[4997]: I1205 08:30:03.263984 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g" event={"ID":"708bfbf7-22c5-4cbf-9dca-c7220fdd1615","Type":"ContainerDied","Data":"740048f5105d0f67b04a833fe888c6200d2062ae50b61738dc15002aa9c62fca"} Dec 05 08:30:03 crc kubenswrapper[4997]: I1205 08:30:03.264032 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="740048f5105d0f67b04a833fe888c6200d2062ae50b61738dc15002aa9c62fca" Dec 05 08:30:03 crc kubenswrapper[4997]: I1205 08:30:03.264054 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g" Dec 05 08:30:03 crc kubenswrapper[4997]: I1205 08:30:03.579291 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth"] Dec 05 08:30:03 crc kubenswrapper[4997]: I1205 08:30:03.590707 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415345-s5dth"] Dec 05 08:30:03 crc kubenswrapper[4997]: I1205 08:30:03.765766 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="898582ab-9959-4f46-8fb8-f363db61cb7f" path="/var/lib/kubelet/pods/898582ab-9959-4f46-8fb8-f363db61cb7f/volumes" Dec 05 08:30:07 crc kubenswrapper[4997]: I1205 08:30:07.462936 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mghch" Dec 05 08:30:07 crc kubenswrapper[4997]: I1205 08:30:07.463377 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mghch" Dec 05 08:30:07 crc kubenswrapper[4997]: I1205 08:30:07.506933 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mghch" Dec 05 08:30:08 crc kubenswrapper[4997]: I1205 08:30:08.341415 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mghch" Dec 05 08:30:08 crc kubenswrapper[4997]: I1205 08:30:08.386861 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mghch"] Dec 05 08:30:10 crc kubenswrapper[4997]: I1205 08:30:10.311921 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mghch" podUID="a7d5c257-54e0-4739-921b-805de66493c1" containerName="registry-server" containerID="cri-o://bbaeb04744b6900fc7aedb337a612440104dc5a048c57c8ca82fc2986e2e18fd" gracePeriod=2 Dec 05 08:30:10 crc kubenswrapper[4997]: E1205 08:30:10.482012 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7d5c257_54e0_4739_921b_805de66493c1.slice/crio-conmon-bbaeb04744b6900fc7aedb337a612440104dc5a048c57c8ca82fc2986e2e18fd.scope\": RecentStats: unable to find data in memory cache]" Dec 05 08:30:10 crc kubenswrapper[4997]: I1205 08:30:10.714241 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mghch" Dec 05 08:30:10 crc kubenswrapper[4997]: I1205 08:30:10.838283 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7d5c257-54e0-4739-921b-805de66493c1-catalog-content\") pod \"a7d5c257-54e0-4739-921b-805de66493c1\" (UID: \"a7d5c257-54e0-4739-921b-805de66493c1\") " Dec 05 08:30:10 crc kubenswrapper[4997]: I1205 08:30:10.838403 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tw52\" (UniqueName: \"kubernetes.io/projected/a7d5c257-54e0-4739-921b-805de66493c1-kube-api-access-5tw52\") pod \"a7d5c257-54e0-4739-921b-805de66493c1\" (UID: \"a7d5c257-54e0-4739-921b-805de66493c1\") " Dec 05 08:30:10 crc kubenswrapper[4997]: I1205 08:30:10.838447 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7d5c257-54e0-4739-921b-805de66493c1-utilities\") pod \"a7d5c257-54e0-4739-921b-805de66493c1\" (UID: \"a7d5c257-54e0-4739-921b-805de66493c1\") " Dec 05 08:30:10 crc kubenswrapper[4997]: I1205 08:30:10.839387 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7d5c257-54e0-4739-921b-805de66493c1-utilities" (OuterVolumeSpecName: "utilities") pod "a7d5c257-54e0-4739-921b-805de66493c1" (UID: "a7d5c257-54e0-4739-921b-805de66493c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:30:10 crc kubenswrapper[4997]: I1205 08:30:10.848120 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7d5c257-54e0-4739-921b-805de66493c1-kube-api-access-5tw52" (OuterVolumeSpecName: "kube-api-access-5tw52") pod "a7d5c257-54e0-4739-921b-805de66493c1" (UID: "a7d5c257-54e0-4739-921b-805de66493c1"). InnerVolumeSpecName "kube-api-access-5tw52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:30:10 crc kubenswrapper[4997]: I1205 08:30:10.871990 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7d5c257-54e0-4739-921b-805de66493c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7d5c257-54e0-4739-921b-805de66493c1" (UID: "a7d5c257-54e0-4739-921b-805de66493c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:30:10 crc kubenswrapper[4997]: I1205 08:30:10.940500 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7d5c257-54e0-4739-921b-805de66493c1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:10 crc kubenswrapper[4997]: I1205 08:30:10.940541 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tw52\" (UniqueName: \"kubernetes.io/projected/a7d5c257-54e0-4739-921b-805de66493c1-kube-api-access-5tw52\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:10 crc kubenswrapper[4997]: I1205 08:30:10.940552 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7d5c257-54e0-4739-921b-805de66493c1-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:11 crc kubenswrapper[4997]: I1205 08:30:11.322206 4997 generic.go:334] "Generic (PLEG): container finished" podID="a7d5c257-54e0-4739-921b-805de66493c1" containerID="bbaeb04744b6900fc7aedb337a612440104dc5a048c57c8ca82fc2986e2e18fd" exitCode=0 Dec 05 08:30:11 crc kubenswrapper[4997]: I1205 08:30:11.322311 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mghch" Dec 05 08:30:11 crc kubenswrapper[4997]: I1205 08:30:11.322309 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mghch" event={"ID":"a7d5c257-54e0-4739-921b-805de66493c1","Type":"ContainerDied","Data":"bbaeb04744b6900fc7aedb337a612440104dc5a048c57c8ca82fc2986e2e18fd"} Dec 05 08:30:11 crc kubenswrapper[4997]: I1205 08:30:11.322692 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mghch" event={"ID":"a7d5c257-54e0-4739-921b-805de66493c1","Type":"ContainerDied","Data":"bf7914600164570af0d71bba42a10740043021a98d10f4a9fba68a6c50577c72"} Dec 05 08:30:11 crc kubenswrapper[4997]: I1205 08:30:11.322726 4997 scope.go:117] "RemoveContainer" containerID="bbaeb04744b6900fc7aedb337a612440104dc5a048c57c8ca82fc2986e2e18fd" Dec 05 08:30:11 crc kubenswrapper[4997]: I1205 08:30:11.353808 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mghch"] Dec 05 08:30:11 crc kubenswrapper[4997]: I1205 08:30:11.353875 4997 scope.go:117] "RemoveContainer" containerID="ae7ebaa6d2e2c16897e0143e39788ceed2d30635cf3de4dce406658eb32a320e" Dec 05 08:30:11 crc kubenswrapper[4997]: I1205 08:30:11.359838 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mghch"] Dec 05 08:30:11 crc kubenswrapper[4997]: I1205 08:30:11.377717 4997 scope.go:117] "RemoveContainer" containerID="f4f500cb927f0d9436dfe7171b6ff620455b6ea43caf0a9b2fde45bb63b247ed" Dec 05 08:30:11 crc kubenswrapper[4997]: I1205 08:30:11.397424 4997 scope.go:117] "RemoveContainer" containerID="bbaeb04744b6900fc7aedb337a612440104dc5a048c57c8ca82fc2986e2e18fd" Dec 05 08:30:11 crc kubenswrapper[4997]: E1205 08:30:11.397950 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbaeb04744b6900fc7aedb337a612440104dc5a048c57c8ca82fc2986e2e18fd\": container with ID starting with bbaeb04744b6900fc7aedb337a612440104dc5a048c57c8ca82fc2986e2e18fd not found: ID does not exist" containerID="bbaeb04744b6900fc7aedb337a612440104dc5a048c57c8ca82fc2986e2e18fd" Dec 05 08:30:11 crc kubenswrapper[4997]: I1205 08:30:11.398028 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbaeb04744b6900fc7aedb337a612440104dc5a048c57c8ca82fc2986e2e18fd"} err="failed to get container status \"bbaeb04744b6900fc7aedb337a612440104dc5a048c57c8ca82fc2986e2e18fd\": rpc error: code = NotFound desc = could not find container \"bbaeb04744b6900fc7aedb337a612440104dc5a048c57c8ca82fc2986e2e18fd\": container with ID starting with bbaeb04744b6900fc7aedb337a612440104dc5a048c57c8ca82fc2986e2e18fd not found: ID does not exist" Dec 05 08:30:11 crc kubenswrapper[4997]: I1205 08:30:11.398112 4997 scope.go:117] "RemoveContainer" containerID="ae7ebaa6d2e2c16897e0143e39788ceed2d30635cf3de4dce406658eb32a320e" Dec 05 08:30:11 crc kubenswrapper[4997]: E1205 08:30:11.398482 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae7ebaa6d2e2c16897e0143e39788ceed2d30635cf3de4dce406658eb32a320e\": container with ID starting with ae7ebaa6d2e2c16897e0143e39788ceed2d30635cf3de4dce406658eb32a320e not found: ID does not exist" containerID="ae7ebaa6d2e2c16897e0143e39788ceed2d30635cf3de4dce406658eb32a320e" Dec 05 08:30:11 crc kubenswrapper[4997]: I1205 08:30:11.398522 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae7ebaa6d2e2c16897e0143e39788ceed2d30635cf3de4dce406658eb32a320e"} err="failed to get container status \"ae7ebaa6d2e2c16897e0143e39788ceed2d30635cf3de4dce406658eb32a320e\": rpc error: code = NotFound desc = could not find container \"ae7ebaa6d2e2c16897e0143e39788ceed2d30635cf3de4dce406658eb32a320e\": container with ID starting with ae7ebaa6d2e2c16897e0143e39788ceed2d30635cf3de4dce406658eb32a320e not found: ID does not exist" Dec 05 08:30:11 crc kubenswrapper[4997]: I1205 08:30:11.398548 4997 scope.go:117] "RemoveContainer" containerID="f4f500cb927f0d9436dfe7171b6ff620455b6ea43caf0a9b2fde45bb63b247ed" Dec 05 08:30:11 crc kubenswrapper[4997]: E1205 08:30:11.399011 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4f500cb927f0d9436dfe7171b6ff620455b6ea43caf0a9b2fde45bb63b247ed\": container with ID starting with f4f500cb927f0d9436dfe7171b6ff620455b6ea43caf0a9b2fde45bb63b247ed not found: ID does not exist" containerID="f4f500cb927f0d9436dfe7171b6ff620455b6ea43caf0a9b2fde45bb63b247ed" Dec 05 08:30:11 crc kubenswrapper[4997]: I1205 08:30:11.399044 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4f500cb927f0d9436dfe7171b6ff620455b6ea43caf0a9b2fde45bb63b247ed"} err="failed to get container status \"f4f500cb927f0d9436dfe7171b6ff620455b6ea43caf0a9b2fde45bb63b247ed\": rpc error: code = NotFound desc = could not find container \"f4f500cb927f0d9436dfe7171b6ff620455b6ea43caf0a9b2fde45bb63b247ed\": container with ID starting with f4f500cb927f0d9436dfe7171b6ff620455b6ea43caf0a9b2fde45bb63b247ed not found: ID does not exist" Dec 05 08:30:11 crc kubenswrapper[4997]: I1205 08:30:11.765404 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7d5c257-54e0-4739-921b-805de66493c1" path="/var/lib/kubelet/pods/a7d5c257-54e0-4739-921b-805de66493c1/volumes" Dec 05 08:30:12 crc kubenswrapper[4997]: I1205 08:30:12.411448 4997 scope.go:117] "RemoveContainer" containerID="c790c32fe98012838b02251f466cc90aa644532caee16c2661f4557eb3fee57c" Dec 05 08:30:19 crc kubenswrapper[4997]: I1205 08:30:19.770094 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:30:19 crc kubenswrapper[4997]: I1205 08:30:19.770938 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:30:49 crc kubenswrapper[4997]: I1205 08:30:49.770342 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:30:49 crc kubenswrapper[4997]: I1205 08:30:49.770892 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:31:19 crc kubenswrapper[4997]: I1205 08:31:19.770697 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:31:19 crc kubenswrapper[4997]: I1205 08:31:19.771196 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:31:19 crc kubenswrapper[4997]: I1205 08:31:19.771241 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 08:31:19 crc kubenswrapper[4997]: I1205 08:31:19.771870 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"896ded924d3256236025e635c3d3f7a52372f71081a01d64d6aec438295ddd47"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:31:19 crc kubenswrapper[4997]: I1205 08:31:19.771928 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://896ded924d3256236025e635c3d3f7a52372f71081a01d64d6aec438295ddd47" gracePeriod=600 Dec 05 08:31:20 crc kubenswrapper[4997]: I1205 08:31:20.833605 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="896ded924d3256236025e635c3d3f7a52372f71081a01d64d6aec438295ddd47" exitCode=0 Dec 05 08:31:20 crc kubenswrapper[4997]: I1205 08:31:20.833657 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"896ded924d3256236025e635c3d3f7a52372f71081a01d64d6aec438295ddd47"} Dec 05 08:31:20 crc kubenswrapper[4997]: I1205 08:31:20.833939 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b"} Dec 05 08:31:20 crc kubenswrapper[4997]: I1205 08:31:20.833960 4997 scope.go:117] "RemoveContainer" containerID="64ab89131aaf7252a095145a125b345f21896769ca302b090d9748f08149b92e" Dec 05 08:33:49 crc kubenswrapper[4997]: I1205 08:33:49.770480 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:33:49 crc kubenswrapper[4997]: I1205 08:33:49.771093 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:34:04 crc kubenswrapper[4997]: I1205 08:34:04.850725 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-5qdxt"] Dec 05 08:34:04 crc kubenswrapper[4997]: I1205 08:34:04.857814 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-5qdxt"] Dec 05 08:34:04 crc kubenswrapper[4997]: I1205 08:34:04.979762 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-wlhk6"] Dec 05 08:34:04 crc kubenswrapper[4997]: E1205 08:34:04.980050 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="708bfbf7-22c5-4cbf-9dca-c7220fdd1615" containerName="collect-profiles" Dec 05 08:34:04 crc kubenswrapper[4997]: I1205 08:34:04.980070 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="708bfbf7-22c5-4cbf-9dca-c7220fdd1615" containerName="collect-profiles" Dec 05 08:34:04 crc kubenswrapper[4997]: E1205 08:34:04.980081 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d5c257-54e0-4739-921b-805de66493c1" containerName="registry-server" Dec 05 08:34:04 crc kubenswrapper[4997]: I1205 08:34:04.980096 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d5c257-54e0-4739-921b-805de66493c1" containerName="registry-server" Dec 05 08:34:04 crc kubenswrapper[4997]: E1205 08:34:04.980106 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d5c257-54e0-4739-921b-805de66493c1" containerName="extract-content" Dec 05 08:34:04 crc kubenswrapper[4997]: I1205 08:34:04.980113 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d5c257-54e0-4739-921b-805de66493c1" containerName="extract-content" Dec 05 08:34:04 crc kubenswrapper[4997]: E1205 08:34:04.980130 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d5c257-54e0-4739-921b-805de66493c1" containerName="extract-utilities" Dec 05 08:34:04 crc kubenswrapper[4997]: I1205 08:34:04.980136 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d5c257-54e0-4739-921b-805de66493c1" containerName="extract-utilities" Dec 05 08:34:04 crc kubenswrapper[4997]: I1205 08:34:04.980300 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7d5c257-54e0-4739-921b-805de66493c1" containerName="registry-server" Dec 05 08:34:04 crc kubenswrapper[4997]: I1205 08:34:04.980313 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="708bfbf7-22c5-4cbf-9dca-c7220fdd1615" containerName="collect-profiles" Dec 05 08:34:04 crc kubenswrapper[4997]: I1205 08:34:04.981865 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wlhk6" Dec 05 08:34:04 crc kubenswrapper[4997]: I1205 08:34:04.985514 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 05 08:34:04 crc kubenswrapper[4997]: I1205 08:34:04.985947 4997 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-dtsv6" Dec 05 08:34:04 crc kubenswrapper[4997]: I1205 08:34:04.985969 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 05 08:34:04 crc kubenswrapper[4997]: I1205 08:34:04.986188 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 05 08:34:05 crc kubenswrapper[4997]: I1205 08:34:05.010506 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-wlhk6"] Dec 05 08:34:05 crc kubenswrapper[4997]: I1205 08:34:05.089745 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6fa2591f-a893-47da-948f-7cf5932db00b-node-mnt\") pod \"crc-storage-crc-wlhk6\" (UID: \"6fa2591f-a893-47da-948f-7cf5932db00b\") " pod="crc-storage/crc-storage-crc-wlhk6" Dec 05 08:34:05 crc kubenswrapper[4997]: I1205 08:34:05.090030 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6fa2591f-a893-47da-948f-7cf5932db00b-crc-storage\") pod \"crc-storage-crc-wlhk6\" (UID: \"6fa2591f-a893-47da-948f-7cf5932db00b\") " pod="crc-storage/crc-storage-crc-wlhk6" Dec 05 08:34:05 crc kubenswrapper[4997]: I1205 08:34:05.090168 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75kkz\" (UniqueName: \"kubernetes.io/projected/6fa2591f-a893-47da-948f-7cf5932db00b-kube-api-access-75kkz\") pod \"crc-storage-crc-wlhk6\" (UID: \"6fa2591f-a893-47da-948f-7cf5932db00b\") " pod="crc-storage/crc-storage-crc-wlhk6" Dec 05 08:34:05 crc kubenswrapper[4997]: I1205 08:34:05.191108 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6fa2591f-a893-47da-948f-7cf5932db00b-crc-storage\") pod \"crc-storage-crc-wlhk6\" (UID: \"6fa2591f-a893-47da-948f-7cf5932db00b\") " pod="crc-storage/crc-storage-crc-wlhk6" Dec 05 08:34:05 crc kubenswrapper[4997]: I1205 08:34:05.191675 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75kkz\" (UniqueName: \"kubernetes.io/projected/6fa2591f-a893-47da-948f-7cf5932db00b-kube-api-access-75kkz\") pod \"crc-storage-crc-wlhk6\" (UID: \"6fa2591f-a893-47da-948f-7cf5932db00b\") " pod="crc-storage/crc-storage-crc-wlhk6" Dec 05 08:34:05 crc kubenswrapper[4997]: I1205 08:34:05.191791 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6fa2591f-a893-47da-948f-7cf5932db00b-crc-storage\") pod \"crc-storage-crc-wlhk6\" (UID: \"6fa2591f-a893-47da-948f-7cf5932db00b\") " pod="crc-storage/crc-storage-crc-wlhk6" Dec 05 08:34:05 crc kubenswrapper[4997]: I1205 08:34:05.191960 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6fa2591f-a893-47da-948f-7cf5932db00b-node-mnt\") pod \"crc-storage-crc-wlhk6\" (UID: \"6fa2591f-a893-47da-948f-7cf5932db00b\") " pod="crc-storage/crc-storage-crc-wlhk6" Dec 05 08:34:05 crc kubenswrapper[4997]: I1205 08:34:05.192126 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6fa2591f-a893-47da-948f-7cf5932db00b-node-mnt\") pod \"crc-storage-crc-wlhk6\" (UID: \"6fa2591f-a893-47da-948f-7cf5932db00b\") " pod="crc-storage/crc-storage-crc-wlhk6" Dec 05 08:34:05 crc kubenswrapper[4997]: I1205 08:34:05.211766 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75kkz\" (UniqueName: \"kubernetes.io/projected/6fa2591f-a893-47da-948f-7cf5932db00b-kube-api-access-75kkz\") pod \"crc-storage-crc-wlhk6\" (UID: \"6fa2591f-a893-47da-948f-7cf5932db00b\") " pod="crc-storage/crc-storage-crc-wlhk6" Dec 05 08:34:05 crc kubenswrapper[4997]: I1205 08:34:05.313335 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wlhk6" Dec 05 08:34:05 crc kubenswrapper[4997]: I1205 08:34:05.764505 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf6e1555-1042-4e81-b239-1289b9ed77cd" path="/var/lib/kubelet/pods/bf6e1555-1042-4e81-b239-1289b9ed77cd/volumes" Dec 05 08:34:05 crc kubenswrapper[4997]: I1205 08:34:05.800398 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-wlhk6"] Dec 05 08:34:05 crc kubenswrapper[4997]: W1205 08:34:05.815557 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fa2591f_a893_47da_948f_7cf5932db00b.slice/crio-d899c56c300e82c7e5e41ab4cfabc3da70b575ed141ea492cde801ee3e7cf221 WatchSource:0}: Error finding container d899c56c300e82c7e5e41ab4cfabc3da70b575ed141ea492cde801ee3e7cf221: Status 404 returned error can't find the container with id d899c56c300e82c7e5e41ab4cfabc3da70b575ed141ea492cde801ee3e7cf221 Dec 05 08:34:05 crc kubenswrapper[4997]: I1205 08:34:05.821404 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 08:34:06 crc kubenswrapper[4997]: I1205 08:34:06.009986 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-wlhk6" event={"ID":"6fa2591f-a893-47da-948f-7cf5932db00b","Type":"ContainerStarted","Data":"d899c56c300e82c7e5e41ab4cfabc3da70b575ed141ea492cde801ee3e7cf221"} Dec 05 08:34:07 crc kubenswrapper[4997]: I1205 08:34:07.020167 4997 generic.go:334] "Generic (PLEG): container finished" podID="6fa2591f-a893-47da-948f-7cf5932db00b" containerID="cce72cf13bf12378384ca3a19d45d45b633e16a929d088ee083df1ab78f57c6c" exitCode=0 Dec 05 08:34:07 crc kubenswrapper[4997]: I1205 08:34:07.020236 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-wlhk6" event={"ID":"6fa2591f-a893-47da-948f-7cf5932db00b","Type":"ContainerDied","Data":"cce72cf13bf12378384ca3a19d45d45b633e16a929d088ee083df1ab78f57c6c"} Dec 05 08:34:08 crc kubenswrapper[4997]: I1205 08:34:08.303457 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wlhk6" Dec 05 08:34:08 crc kubenswrapper[4997]: I1205 08:34:08.440049 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6fa2591f-a893-47da-948f-7cf5932db00b-crc-storage\") pod \"6fa2591f-a893-47da-948f-7cf5932db00b\" (UID: \"6fa2591f-a893-47da-948f-7cf5932db00b\") " Dec 05 08:34:08 crc kubenswrapper[4997]: I1205 08:34:08.440170 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6fa2591f-a893-47da-948f-7cf5932db00b-node-mnt\") pod \"6fa2591f-a893-47da-948f-7cf5932db00b\" (UID: \"6fa2591f-a893-47da-948f-7cf5932db00b\") " Dec 05 08:34:08 crc kubenswrapper[4997]: I1205 08:34:08.440224 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75kkz\" (UniqueName: \"kubernetes.io/projected/6fa2591f-a893-47da-948f-7cf5932db00b-kube-api-access-75kkz\") pod \"6fa2591f-a893-47da-948f-7cf5932db00b\" (UID: \"6fa2591f-a893-47da-948f-7cf5932db00b\") " Dec 05 08:34:08 crc kubenswrapper[4997]: I1205 08:34:08.440703 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6fa2591f-a893-47da-948f-7cf5932db00b-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "6fa2591f-a893-47da-948f-7cf5932db00b" (UID: "6fa2591f-a893-47da-948f-7cf5932db00b"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:34:08 crc kubenswrapper[4997]: I1205 08:34:08.445698 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fa2591f-a893-47da-948f-7cf5932db00b-kube-api-access-75kkz" (OuterVolumeSpecName: "kube-api-access-75kkz") pod "6fa2591f-a893-47da-948f-7cf5932db00b" (UID: "6fa2591f-a893-47da-948f-7cf5932db00b"). InnerVolumeSpecName "kube-api-access-75kkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:34:08 crc kubenswrapper[4997]: I1205 08:34:08.460939 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fa2591f-a893-47da-948f-7cf5932db00b-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "6fa2591f-a893-47da-948f-7cf5932db00b" (UID: "6fa2591f-a893-47da-948f-7cf5932db00b"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:08 crc kubenswrapper[4997]: I1205 08:34:08.541578 4997 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6fa2591f-a893-47da-948f-7cf5932db00b-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:08 crc kubenswrapper[4997]: I1205 08:34:08.541839 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75kkz\" (UniqueName: \"kubernetes.io/projected/6fa2591f-a893-47da-948f-7cf5932db00b-kube-api-access-75kkz\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:08 crc kubenswrapper[4997]: I1205 08:34:08.541852 4997 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6fa2591f-a893-47da-948f-7cf5932db00b-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:09 crc kubenswrapper[4997]: I1205 08:34:09.042755 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-wlhk6" event={"ID":"6fa2591f-a893-47da-948f-7cf5932db00b","Type":"ContainerDied","Data":"d899c56c300e82c7e5e41ab4cfabc3da70b575ed141ea492cde801ee3e7cf221"} Dec 05 08:34:09 crc kubenswrapper[4997]: I1205 08:34:09.042801 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d899c56c300e82c7e5e41ab4cfabc3da70b575ed141ea492cde801ee3e7cf221" Dec 05 08:34:09 crc kubenswrapper[4997]: I1205 08:34:09.042923 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wlhk6" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.513671 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-wlhk6"] Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.518425 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-wlhk6"] Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.632562 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-gnvbb"] Dec 05 08:34:10 crc kubenswrapper[4997]: E1205 08:34:10.632932 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fa2591f-a893-47da-948f-7cf5932db00b" containerName="storage" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.632955 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fa2591f-a893-47da-948f-7cf5932db00b" containerName="storage" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.633114 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fa2591f-a893-47da-948f-7cf5932db00b" containerName="storage" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.633580 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-gnvbb" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.638458 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.638528 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.639192 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.644285 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-gnvbb"] Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.647455 4997 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-dtsv6" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.773111 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr8m9\" (UniqueName: \"kubernetes.io/projected/8d33518d-912e-4159-93fa-2c0281c4ae46-kube-api-access-rr8m9\") pod \"crc-storage-crc-gnvbb\" (UID: \"8d33518d-912e-4159-93fa-2c0281c4ae46\") " pod="crc-storage/crc-storage-crc-gnvbb" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.773174 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/8d33518d-912e-4159-93fa-2c0281c4ae46-node-mnt\") pod \"crc-storage-crc-gnvbb\" (UID: \"8d33518d-912e-4159-93fa-2c0281c4ae46\") " pod="crc-storage/crc-storage-crc-gnvbb" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.773307 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/8d33518d-912e-4159-93fa-2c0281c4ae46-crc-storage\") pod \"crc-storage-crc-gnvbb\" (UID: \"8d33518d-912e-4159-93fa-2c0281c4ae46\") " pod="crc-storage/crc-storage-crc-gnvbb" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.874597 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr8m9\" (UniqueName: \"kubernetes.io/projected/8d33518d-912e-4159-93fa-2c0281c4ae46-kube-api-access-rr8m9\") pod \"crc-storage-crc-gnvbb\" (UID: \"8d33518d-912e-4159-93fa-2c0281c4ae46\") " pod="crc-storage/crc-storage-crc-gnvbb" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.874718 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/8d33518d-912e-4159-93fa-2c0281c4ae46-node-mnt\") pod \"crc-storage-crc-gnvbb\" (UID: \"8d33518d-912e-4159-93fa-2c0281c4ae46\") " pod="crc-storage/crc-storage-crc-gnvbb" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.874827 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/8d33518d-912e-4159-93fa-2c0281c4ae46-crc-storage\") pod \"crc-storage-crc-gnvbb\" (UID: \"8d33518d-912e-4159-93fa-2c0281c4ae46\") " pod="crc-storage/crc-storage-crc-gnvbb" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.875202 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/8d33518d-912e-4159-93fa-2c0281c4ae46-node-mnt\") pod \"crc-storage-crc-gnvbb\" (UID: \"8d33518d-912e-4159-93fa-2c0281c4ae46\") " pod="crc-storage/crc-storage-crc-gnvbb" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.875554 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/8d33518d-912e-4159-93fa-2c0281c4ae46-crc-storage\") pod \"crc-storage-crc-gnvbb\" (UID: \"8d33518d-912e-4159-93fa-2c0281c4ae46\") " pod="crc-storage/crc-storage-crc-gnvbb" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.900339 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr8m9\" (UniqueName: \"kubernetes.io/projected/8d33518d-912e-4159-93fa-2c0281c4ae46-kube-api-access-rr8m9\") pod \"crc-storage-crc-gnvbb\" (UID: \"8d33518d-912e-4159-93fa-2c0281c4ae46\") " pod="crc-storage/crc-storage-crc-gnvbb" Dec 05 08:34:10 crc kubenswrapper[4997]: I1205 08:34:10.948838 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-gnvbb" Dec 05 08:34:11 crc kubenswrapper[4997]: I1205 08:34:11.430948 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-gnvbb"] Dec 05 08:34:11 crc kubenswrapper[4997]: I1205 08:34:11.760292 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fa2591f-a893-47da-948f-7cf5932db00b" path="/var/lib/kubelet/pods/6fa2591f-a893-47da-948f-7cf5932db00b/volumes" Dec 05 08:34:12 crc kubenswrapper[4997]: I1205 08:34:12.074343 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-gnvbb" event={"ID":"8d33518d-912e-4159-93fa-2c0281c4ae46","Type":"ContainerStarted","Data":"866b6472b65da7458533e49797c8b5d2440ffd5c6874ac6e9811c7d192101b3a"} Dec 05 08:34:12 crc kubenswrapper[4997]: I1205 08:34:12.074674 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-gnvbb" event={"ID":"8d33518d-912e-4159-93fa-2c0281c4ae46","Type":"ContainerStarted","Data":"23bca154f5ede854207e1ef46e7b2b5bed6458caf840e228a413bf9a18e1a0b0"} Dec 05 08:34:12 crc kubenswrapper[4997]: I1205 08:34:12.092233 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="crc-storage/crc-storage-crc-gnvbb" podStartSLOduration=1.691226172 podStartE2EDuration="2.092192839s" podCreationTimestamp="2025-12-05 08:34:10 +0000 UTC" firstStartedPulling="2025-12-05 08:34:11.445223076 +0000 UTC m=+5951.974130347" lastFinishedPulling="2025-12-05 08:34:11.846189753 +0000 UTC m=+5952.375097014" observedRunningTime="2025-12-05 08:34:12.089479036 +0000 UTC m=+5952.618386307" watchObservedRunningTime="2025-12-05 08:34:12.092192839 +0000 UTC m=+5952.621100100" Dec 05 08:34:12 crc kubenswrapper[4997]: I1205 08:34:12.550320 4997 scope.go:117] "RemoveContainer" containerID="b609fe619a1e20f6758f3c4f05f2e54a4c673bbf3098d4459662bf0c7cd96193" Dec 05 08:34:13 crc kubenswrapper[4997]: I1205 08:34:13.083179 4997 generic.go:334] "Generic (PLEG): container finished" podID="8d33518d-912e-4159-93fa-2c0281c4ae46" containerID="866b6472b65da7458533e49797c8b5d2440ffd5c6874ac6e9811c7d192101b3a" exitCode=0 Dec 05 08:34:13 crc kubenswrapper[4997]: I1205 08:34:13.083265 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-gnvbb" event={"ID":"8d33518d-912e-4159-93fa-2c0281c4ae46","Type":"ContainerDied","Data":"866b6472b65da7458533e49797c8b5d2440ffd5c6874ac6e9811c7d192101b3a"} Dec 05 08:34:14 crc kubenswrapper[4997]: I1205 08:34:14.363959 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-gnvbb" Dec 05 08:34:14 crc kubenswrapper[4997]: I1205 08:34:14.462177 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/8d33518d-912e-4159-93fa-2c0281c4ae46-crc-storage\") pod \"8d33518d-912e-4159-93fa-2c0281c4ae46\" (UID: \"8d33518d-912e-4159-93fa-2c0281c4ae46\") " Dec 05 08:34:14 crc kubenswrapper[4997]: I1205 08:34:14.462284 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/8d33518d-912e-4159-93fa-2c0281c4ae46-node-mnt\") pod \"8d33518d-912e-4159-93fa-2c0281c4ae46\" (UID: \"8d33518d-912e-4159-93fa-2c0281c4ae46\") " Dec 05 08:34:14 crc kubenswrapper[4997]: I1205 08:34:14.462367 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rr8m9\" (UniqueName: \"kubernetes.io/projected/8d33518d-912e-4159-93fa-2c0281c4ae46-kube-api-access-rr8m9\") pod \"8d33518d-912e-4159-93fa-2c0281c4ae46\" (UID: \"8d33518d-912e-4159-93fa-2c0281c4ae46\") " Dec 05 08:34:14 crc kubenswrapper[4997]: I1205 08:34:14.462603 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8d33518d-912e-4159-93fa-2c0281c4ae46-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "8d33518d-912e-4159-93fa-2c0281c4ae46" (UID: "8d33518d-912e-4159-93fa-2c0281c4ae46"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:34:14 crc kubenswrapper[4997]: I1205 08:34:14.463021 4997 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/8d33518d-912e-4159-93fa-2c0281c4ae46-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:14 crc kubenswrapper[4997]: I1205 08:34:14.467804 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d33518d-912e-4159-93fa-2c0281c4ae46-kube-api-access-rr8m9" (OuterVolumeSpecName: "kube-api-access-rr8m9") pod "8d33518d-912e-4159-93fa-2c0281c4ae46" (UID: "8d33518d-912e-4159-93fa-2c0281c4ae46"). InnerVolumeSpecName "kube-api-access-rr8m9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:34:14 crc kubenswrapper[4997]: I1205 08:34:14.482688 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d33518d-912e-4159-93fa-2c0281c4ae46-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "8d33518d-912e-4159-93fa-2c0281c4ae46" (UID: "8d33518d-912e-4159-93fa-2c0281c4ae46"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:14 crc kubenswrapper[4997]: I1205 08:34:14.563702 4997 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/8d33518d-912e-4159-93fa-2c0281c4ae46-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:14 crc kubenswrapper[4997]: I1205 08:34:14.563748 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rr8m9\" (UniqueName: \"kubernetes.io/projected/8d33518d-912e-4159-93fa-2c0281c4ae46-kube-api-access-rr8m9\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:15 crc kubenswrapper[4997]: I1205 08:34:15.104360 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-gnvbb" event={"ID":"8d33518d-912e-4159-93fa-2c0281c4ae46","Type":"ContainerDied","Data":"23bca154f5ede854207e1ef46e7b2b5bed6458caf840e228a413bf9a18e1a0b0"} Dec 05 08:34:15 crc kubenswrapper[4997]: I1205 08:34:15.104880 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23bca154f5ede854207e1ef46e7b2b5bed6458caf840e228a413bf9a18e1a0b0" Dec 05 08:34:15 crc kubenswrapper[4997]: I1205 08:34:15.104427 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-gnvbb" Dec 05 08:34:19 crc kubenswrapper[4997]: I1205 08:34:19.770351 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:34:19 crc kubenswrapper[4997]: I1205 08:34:19.770743 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:34:49 crc kubenswrapper[4997]: I1205 08:34:49.769844 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:34:49 crc kubenswrapper[4997]: I1205 08:34:49.770443 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:34:49 crc kubenswrapper[4997]: I1205 08:34:49.770492 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 08:34:49 crc kubenswrapper[4997]: I1205 08:34:49.771212 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:34:49 crc kubenswrapper[4997]: I1205 08:34:49.771276 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" gracePeriod=600 Dec 05 08:34:50 crc kubenswrapper[4997]: I1205 08:34:50.363100 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" exitCode=0 Dec 05 08:34:50 crc kubenswrapper[4997]: I1205 08:34:50.363156 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b"} Dec 05 08:34:50 crc kubenswrapper[4997]: I1205 08:34:50.363195 4997 scope.go:117] "RemoveContainer" containerID="896ded924d3256236025e635c3d3f7a52372f71081a01d64d6aec438295ddd47" Dec 05 08:34:50 crc kubenswrapper[4997]: E1205 08:34:50.632478 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:34:51 crc kubenswrapper[4997]: I1205 08:34:51.371677 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:34:51 crc kubenswrapper[4997]: E1205 08:34:51.372017 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:35:02 crc kubenswrapper[4997]: I1205 08:35:02.748918 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:35:02 crc kubenswrapper[4997]: E1205 08:35:02.749807 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:35:15 crc kubenswrapper[4997]: I1205 08:35:15.749407 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:35:15 crc kubenswrapper[4997]: E1205 08:35:15.750159 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:35:26 crc kubenswrapper[4997]: I1205 08:35:26.749398 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:35:26 crc kubenswrapper[4997]: E1205 08:35:26.751655 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:35:38 crc kubenswrapper[4997]: I1205 08:35:38.748848 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:35:38 crc kubenswrapper[4997]: E1205 08:35:38.749590 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:35:53 crc kubenswrapper[4997]: I1205 08:35:53.749559 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:35:53 crc kubenswrapper[4997]: E1205 08:35:53.750339 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:36:05 crc kubenswrapper[4997]: I1205 08:36:05.750287 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:36:05 crc kubenswrapper[4997]: E1205 08:36:05.750867 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:36:16 crc kubenswrapper[4997]: I1205 08:36:16.749115 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:36:16 crc kubenswrapper[4997]: E1205 08:36:16.749808 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.193366 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7688886755-zxr2l"] Dec 05 08:36:22 crc kubenswrapper[4997]: E1205 08:36:22.194456 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d33518d-912e-4159-93fa-2c0281c4ae46" containerName="storage" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.194474 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d33518d-912e-4159-93fa-2c0281c4ae46" containerName="storage" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.194676 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d33518d-912e-4159-93fa-2c0281c4ae46" containerName="storage" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.195607 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7688886755-zxr2l" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.201209 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7688886755-zxr2l"] Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.204964 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.205343 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.205555 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-62dnh" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.205742 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.205958 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.305590 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69967565-7a54-4904-9cca-41dba04dbef1-dns-svc\") pod \"dnsmasq-dns-7688886755-zxr2l\" (UID: \"69967565-7a54-4904-9cca-41dba04dbef1\") " pod="openstack/dnsmasq-dns-7688886755-zxr2l" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.305755 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69967565-7a54-4904-9cca-41dba04dbef1-config\") pod \"dnsmasq-dns-7688886755-zxr2l\" (UID: \"69967565-7a54-4904-9cca-41dba04dbef1\") " pod="openstack/dnsmasq-dns-7688886755-zxr2l" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.305895 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmghs\" (UniqueName: \"kubernetes.io/projected/69967565-7a54-4904-9cca-41dba04dbef1-kube-api-access-jmghs\") pod \"dnsmasq-dns-7688886755-zxr2l\" (UID: \"69967565-7a54-4904-9cca-41dba04dbef1\") " pod="openstack/dnsmasq-dns-7688886755-zxr2l" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.406780 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69967565-7a54-4904-9cca-41dba04dbef1-dns-svc\") pod \"dnsmasq-dns-7688886755-zxr2l\" (UID: \"69967565-7a54-4904-9cca-41dba04dbef1\") " pod="openstack/dnsmasq-dns-7688886755-zxr2l" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.406870 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69967565-7a54-4904-9cca-41dba04dbef1-config\") pod \"dnsmasq-dns-7688886755-zxr2l\" (UID: \"69967565-7a54-4904-9cca-41dba04dbef1\") " pod="openstack/dnsmasq-dns-7688886755-zxr2l" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.406933 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmghs\" (UniqueName: \"kubernetes.io/projected/69967565-7a54-4904-9cca-41dba04dbef1-kube-api-access-jmghs\") pod \"dnsmasq-dns-7688886755-zxr2l\" (UID: \"69967565-7a54-4904-9cca-41dba04dbef1\") " pod="openstack/dnsmasq-dns-7688886755-zxr2l" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.407728 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69967565-7a54-4904-9cca-41dba04dbef1-dns-svc\") pod \"dnsmasq-dns-7688886755-zxr2l\" (UID: \"69967565-7a54-4904-9cca-41dba04dbef1\") " pod="openstack/dnsmasq-dns-7688886755-zxr2l" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.407918 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69967565-7a54-4904-9cca-41dba04dbef1-config\") pod \"dnsmasq-dns-7688886755-zxr2l\" (UID: \"69967565-7a54-4904-9cca-41dba04dbef1\") " pod="openstack/dnsmasq-dns-7688886755-zxr2l" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.432027 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84496478f-jmzvf"] Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.433367 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84496478f-jmzvf" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.443678 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmghs\" (UniqueName: \"kubernetes.io/projected/69967565-7a54-4904-9cca-41dba04dbef1-kube-api-access-jmghs\") pod \"dnsmasq-dns-7688886755-zxr2l\" (UID: \"69967565-7a54-4904-9cca-41dba04dbef1\") " pod="openstack/dnsmasq-dns-7688886755-zxr2l" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.449466 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84496478f-jmzvf"] Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.524259 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7688886755-zxr2l" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.610364 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9355424-f0c2-4a5c-9fce-5a48bc52887a-config\") pod \"dnsmasq-dns-84496478f-jmzvf\" (UID: \"b9355424-f0c2-4a5c-9fce-5a48bc52887a\") " pod="openstack/dnsmasq-dns-84496478f-jmzvf" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.610425 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s88h\" (UniqueName: \"kubernetes.io/projected/b9355424-f0c2-4a5c-9fce-5a48bc52887a-kube-api-access-4s88h\") pod \"dnsmasq-dns-84496478f-jmzvf\" (UID: \"b9355424-f0c2-4a5c-9fce-5a48bc52887a\") " pod="openstack/dnsmasq-dns-84496478f-jmzvf" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.610463 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9355424-f0c2-4a5c-9fce-5a48bc52887a-dns-svc\") pod \"dnsmasq-dns-84496478f-jmzvf\" (UID: \"b9355424-f0c2-4a5c-9fce-5a48bc52887a\") " pod="openstack/dnsmasq-dns-84496478f-jmzvf" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.712314 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9355424-f0c2-4a5c-9fce-5a48bc52887a-config\") pod \"dnsmasq-dns-84496478f-jmzvf\" (UID: \"b9355424-f0c2-4a5c-9fce-5a48bc52887a\") " pod="openstack/dnsmasq-dns-84496478f-jmzvf" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.712401 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s88h\" (UniqueName: \"kubernetes.io/projected/b9355424-f0c2-4a5c-9fce-5a48bc52887a-kube-api-access-4s88h\") pod \"dnsmasq-dns-84496478f-jmzvf\" (UID: \"b9355424-f0c2-4a5c-9fce-5a48bc52887a\") " pod="openstack/dnsmasq-dns-84496478f-jmzvf" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.712442 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9355424-f0c2-4a5c-9fce-5a48bc52887a-dns-svc\") pod \"dnsmasq-dns-84496478f-jmzvf\" (UID: \"b9355424-f0c2-4a5c-9fce-5a48bc52887a\") " pod="openstack/dnsmasq-dns-84496478f-jmzvf" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.713506 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9355424-f0c2-4a5c-9fce-5a48bc52887a-config\") pod \"dnsmasq-dns-84496478f-jmzvf\" (UID: \"b9355424-f0c2-4a5c-9fce-5a48bc52887a\") " pod="openstack/dnsmasq-dns-84496478f-jmzvf" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.715876 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9355424-f0c2-4a5c-9fce-5a48bc52887a-dns-svc\") pod \"dnsmasq-dns-84496478f-jmzvf\" (UID: \"b9355424-f0c2-4a5c-9fce-5a48bc52887a\") " pod="openstack/dnsmasq-dns-84496478f-jmzvf" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.772112 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s88h\" (UniqueName: \"kubernetes.io/projected/b9355424-f0c2-4a5c-9fce-5a48bc52887a-kube-api-access-4s88h\") pod \"dnsmasq-dns-84496478f-jmzvf\" (UID: \"b9355424-f0c2-4a5c-9fce-5a48bc52887a\") " pod="openstack/dnsmasq-dns-84496478f-jmzvf" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.775426 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84496478f-jmzvf" Dec 05 08:36:22 crc kubenswrapper[4997]: I1205 08:36:22.832340 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7688886755-zxr2l"] Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.024679 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7688886755-zxr2l" event={"ID":"69967565-7a54-4904-9cca-41dba04dbef1","Type":"ContainerStarted","Data":"761cf8c0894921937646ded4b0cb69c3f454dd4917ffcddd3e376a30e1bf82b8"} Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.041794 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84496478f-jmzvf"] Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.313540 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.315723 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.319224 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.319419 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.319665 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.319768 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-56hjt" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.321461 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.328635 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.424951 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/235b6759-15f6-41b7-a3ed-151f56da12f6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.425100 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.425124 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.425149 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/235b6759-15f6-41b7-a3ed-151f56da12f6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.425210 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/235b6759-15f6-41b7-a3ed-151f56da12f6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.425233 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.425251 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.425269 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jrp2\" (UniqueName: \"kubernetes.io/projected/235b6759-15f6-41b7-a3ed-151f56da12f6-kube-api-access-2jrp2\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.425290 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/235b6759-15f6-41b7-a3ed-151f56da12f6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.526065 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/235b6759-15f6-41b7-a3ed-151f56da12f6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.526111 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.526134 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.526153 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jrp2\" (UniqueName: \"kubernetes.io/projected/235b6759-15f6-41b7-a3ed-151f56da12f6-kube-api-access-2jrp2\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.526173 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/235b6759-15f6-41b7-a3ed-151f56da12f6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.526193 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/235b6759-15f6-41b7-a3ed-151f56da12f6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.526657 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.527315 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.528563 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/235b6759-15f6-41b7-a3ed-151f56da12f6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.528605 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.528668 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/235b6759-15f6-41b7-a3ed-151f56da12f6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.529250 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.529860 4997 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.529889 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/cebfedb8bf3ab96dff23009c7dd389f6e95ad5aceed82f3636d5fe256275b59d/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.532360 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/235b6759-15f6-41b7-a3ed-151f56da12f6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.535843 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/235b6759-15f6-41b7-a3ed-151f56da12f6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.536242 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.537807 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/235b6759-15f6-41b7-a3ed-151f56da12f6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.545009 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jrp2\" (UniqueName: \"kubernetes.io/projected/235b6759-15f6-41b7-a3ed-151f56da12f6-kube-api-access-2jrp2\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.568017 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\") pod \"rabbitmq-server-0\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.598683 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.607924 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.610413 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.610561 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.610645 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-khzbp" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.611873 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.613521 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.614092 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.648410 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.732759 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c67f66b9-d729-463a-8a23-b24870c6ec50-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.732864 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.732912 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5a76facf-8690-4b19-80a2-011b963e7f01\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5a76facf-8690-4b19-80a2-011b963e7f01\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.732965 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.733085 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c67f66b9-d729-463a-8a23-b24870c6ec50-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.733117 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt9ps\" (UniqueName: \"kubernetes.io/projected/c67f66b9-d729-463a-8a23-b24870c6ec50-kube-api-access-wt9ps\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.733168 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c67f66b9-d729-463a-8a23-b24870c6ec50-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.733197 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c67f66b9-d729-463a-8a23-b24870c6ec50-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.733252 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.834976 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c67f66b9-d729-463a-8a23-b24870c6ec50-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.835067 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.835115 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5a76facf-8690-4b19-80a2-011b963e7f01\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5a76facf-8690-4b19-80a2-011b963e7f01\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.835142 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.835195 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c67f66b9-d729-463a-8a23-b24870c6ec50-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.835217 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt9ps\" (UniqueName: \"kubernetes.io/projected/c67f66b9-d729-463a-8a23-b24870c6ec50-kube-api-access-wt9ps\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.835236 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c67f66b9-d729-463a-8a23-b24870c6ec50-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.835280 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c67f66b9-d729-463a-8a23-b24870c6ec50-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.835303 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.839461 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.839852 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.840538 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c67f66b9-d729-463a-8a23-b24870c6ec50-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.851252 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c67f66b9-d729-463a-8a23-b24870c6ec50-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.852117 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.852255 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c67f66b9-d729-463a-8a23-b24870c6ec50-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.852601 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c67f66b9-d729-463a-8a23-b24870c6ec50-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.853589 4997 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.853628 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5a76facf-8690-4b19-80a2-011b963e7f01\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5a76facf-8690-4b19-80a2-011b963e7f01\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/13f336a81097285c82b8991b303326fa943937a5f7cb23c86fbd23e7b1c598a3/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.855985 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt9ps\" (UniqueName: \"kubernetes.io/projected/c67f66b9-d729-463a-8a23-b24870c6ec50-kube-api-access-wt9ps\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.919221 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5a76facf-8690-4b19-80a2-011b963e7f01\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5a76facf-8690-4b19-80a2-011b963e7f01\") pod \"rabbitmq-cell1-server-0\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.943348 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:36:23 crc kubenswrapper[4997]: I1205 08:36:23.947187 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.005353 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.006533 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.008742 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.011865 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-fc4jc" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.012945 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.061607 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"235b6759-15f6-41b7-a3ed-151f56da12f6","Type":"ContainerStarted","Data":"e8443b1838c57ae9f8129b83eb55ba1b6e6b9196a81de29d5740c2187e1726a0"} Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.072165 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84496478f-jmzvf" event={"ID":"b9355424-f0c2-4a5c-9fce-5a48bc52887a","Type":"ContainerStarted","Data":"ff9407c03bc127c1e0066f3b6dde0a9f2b7280fd785fa3e818320658da8673f7"} Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.106474 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.108660 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.114935 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.115107 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.115142 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.120898 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-6tvz7" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.125459 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.126285 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.138597 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2800a118-2688-4e45-8554-6d613c61d3ad-config-data\") pod \"memcached-0\" (UID: \"2800a118-2688-4e45-8554-6d613c61d3ad\") " pod="openstack/memcached-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.138801 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2800a118-2688-4e45-8554-6d613c61d3ad-kolla-config\") pod \"memcached-0\" (UID: \"2800a118-2688-4e45-8554-6d613c61d3ad\") " pod="openstack/memcached-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.138926 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj4r9\" (UniqueName: \"kubernetes.io/projected/2800a118-2688-4e45-8554-6d613c61d3ad-kube-api-access-vj4r9\") pod \"memcached-0\" (UID: \"2800a118-2688-4e45-8554-6d613c61d3ad\") " pod="openstack/memcached-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.240386 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/108e77e9-9f51-4686-bac0-1e0397f33c28-config-data-default\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.240440 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/108e77e9-9f51-4686-bac0-1e0397f33c28-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.240513 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj4r9\" (UniqueName: \"kubernetes.io/projected/2800a118-2688-4e45-8554-6d613c61d3ad-kube-api-access-vj4r9\") pod \"memcached-0\" (UID: \"2800a118-2688-4e45-8554-6d613c61d3ad\") " pod="openstack/memcached-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.240567 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0c5da443-23a2-454c-a933-17eafb3cbe63\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0c5da443-23a2-454c-a933-17eafb3cbe63\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.240593 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/108e77e9-9f51-4686-bac0-1e0397f33c28-operator-scripts\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.240631 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/108e77e9-9f51-4686-bac0-1e0397f33c28-kolla-config\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.240658 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/108e77e9-9f51-4686-bac0-1e0397f33c28-config-data-generated\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.240689 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpgtg\" (UniqueName: \"kubernetes.io/projected/108e77e9-9f51-4686-bac0-1e0397f33c28-kube-api-access-zpgtg\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.240839 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/108e77e9-9f51-4686-bac0-1e0397f33c28-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.240949 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2800a118-2688-4e45-8554-6d613c61d3ad-config-data\") pod \"memcached-0\" (UID: \"2800a118-2688-4e45-8554-6d613c61d3ad\") " pod="openstack/memcached-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.241113 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2800a118-2688-4e45-8554-6d613c61d3ad-kolla-config\") pod \"memcached-0\" (UID: \"2800a118-2688-4e45-8554-6d613c61d3ad\") " pod="openstack/memcached-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.241935 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2800a118-2688-4e45-8554-6d613c61d3ad-config-data\") pod \"memcached-0\" (UID: \"2800a118-2688-4e45-8554-6d613c61d3ad\") " pod="openstack/memcached-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.242277 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2800a118-2688-4e45-8554-6d613c61d3ad-kolla-config\") pod \"memcached-0\" (UID: \"2800a118-2688-4e45-8554-6d613c61d3ad\") " pod="openstack/memcached-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.286742 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj4r9\" (UniqueName: \"kubernetes.io/projected/2800a118-2688-4e45-8554-6d613c61d3ad-kube-api-access-vj4r9\") pod \"memcached-0\" (UID: \"2800a118-2688-4e45-8554-6d613c61d3ad\") " pod="openstack/memcached-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.331912 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.343288 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/108e77e9-9f51-4686-bac0-1e0397f33c28-config-data-default\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.343340 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/108e77e9-9f51-4686-bac0-1e0397f33c28-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.343394 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0c5da443-23a2-454c-a933-17eafb3cbe63\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0c5da443-23a2-454c-a933-17eafb3cbe63\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.343466 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/108e77e9-9f51-4686-bac0-1e0397f33c28-operator-scripts\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.343493 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/108e77e9-9f51-4686-bac0-1e0397f33c28-kolla-config\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.343520 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/108e77e9-9f51-4686-bac0-1e0397f33c28-config-data-generated\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.343547 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpgtg\" (UniqueName: \"kubernetes.io/projected/108e77e9-9f51-4686-bac0-1e0397f33c28-kube-api-access-zpgtg\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.343572 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/108e77e9-9f51-4686-bac0-1e0397f33c28-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.349503 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/108e77e9-9f51-4686-bac0-1e0397f33c28-config-data-default\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.350117 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/108e77e9-9f51-4686-bac0-1e0397f33c28-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.350665 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/108e77e9-9f51-4686-bac0-1e0397f33c28-operator-scripts\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.353301 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/108e77e9-9f51-4686-bac0-1e0397f33c28-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.354753 4997 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.354799 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0c5da443-23a2-454c-a933-17eafb3cbe63\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0c5da443-23a2-454c-a933-17eafb3cbe63\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1250bb41b583a834a32625eaa8f3f50eb3c32cb5d8bfabd005b753febe31c4ea/globalmount\"" pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.358738 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/108e77e9-9f51-4686-bac0-1e0397f33c28-config-data-generated\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.358796 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/108e77e9-9f51-4686-bac0-1e0397f33c28-kolla-config\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.361629 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpgtg\" (UniqueName: \"kubernetes.io/projected/108e77e9-9f51-4686-bac0-1e0397f33c28-kube-api-access-zpgtg\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.388080 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0c5da443-23a2-454c-a933-17eafb3cbe63\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0c5da443-23a2-454c-a933-17eafb3cbe63\") pod \"openstack-galera-0\" (UID: \"108e77e9-9f51-4686-bac0-1e0397f33c28\") " pod="openstack/openstack-galera-0" Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.427784 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:36:24 crc kubenswrapper[4997]: I1205 08:36:24.457118 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:24.595570 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:24.919265 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:24.921351 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:24.925630 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:24.925645 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-26bgz" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:24.926021 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:24.926360 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:24.938735 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.067596 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0d25601d-3268-4c04-86c7-fc1a60f6a379\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0d25601d-3268-4c04-86c7-fc1a60f6a379\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.067688 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/351c29ac-7ec2-4c65-95a1-040707a8d01f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.067743 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcnz9\" (UniqueName: \"kubernetes.io/projected/351c29ac-7ec2-4c65-95a1-040707a8d01f-kube-api-access-mcnz9\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.067778 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/351c29ac-7ec2-4c65-95a1-040707a8d01f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.068738 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/351c29ac-7ec2-4c65-95a1-040707a8d01f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.068782 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/351c29ac-7ec2-4c65-95a1-040707a8d01f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.068830 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/351c29ac-7ec2-4c65-95a1-040707a8d01f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.068877 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/351c29ac-7ec2-4c65-95a1-040707a8d01f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.081037 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"2800a118-2688-4e45-8554-6d613c61d3ad","Type":"ContainerStarted","Data":"63331b24aefe6a39908fbecf41b083dc59ea3f43f2aa5bfb7d6830f94014dfdd"} Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.082321 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c67f66b9-d729-463a-8a23-b24870c6ec50","Type":"ContainerStarted","Data":"0e98739ad027b512b84e7ff54a61b09797442c21641432f87499b73dbd0425db"} Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.170701 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcnz9\" (UniqueName: \"kubernetes.io/projected/351c29ac-7ec2-4c65-95a1-040707a8d01f-kube-api-access-mcnz9\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.170759 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/351c29ac-7ec2-4c65-95a1-040707a8d01f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.170801 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/351c29ac-7ec2-4c65-95a1-040707a8d01f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.170818 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/351c29ac-7ec2-4c65-95a1-040707a8d01f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.170839 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/351c29ac-7ec2-4c65-95a1-040707a8d01f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.170860 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/351c29ac-7ec2-4c65-95a1-040707a8d01f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.170931 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0d25601d-3268-4c04-86c7-fc1a60f6a379\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0d25601d-3268-4c04-86c7-fc1a60f6a379\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.170971 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/351c29ac-7ec2-4c65-95a1-040707a8d01f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.172210 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/351c29ac-7ec2-4c65-95a1-040707a8d01f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.172307 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/351c29ac-7ec2-4c65-95a1-040707a8d01f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.172820 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/351c29ac-7ec2-4c65-95a1-040707a8d01f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.173385 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/351c29ac-7ec2-4c65-95a1-040707a8d01f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.174698 4997 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.174722 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0d25601d-3268-4c04-86c7-fc1a60f6a379\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0d25601d-3268-4c04-86c7-fc1a60f6a379\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/73b114467f831efd73d8b7269daa643d93f3229623d1436fb31730b2357c55a6/globalmount\"" pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.189425 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/351c29ac-7ec2-4c65-95a1-040707a8d01f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.190454 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/351c29ac-7ec2-4c65-95a1-040707a8d01f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.194077 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcnz9\" (UniqueName: \"kubernetes.io/projected/351c29ac-7ec2-4c65-95a1-040707a8d01f-kube-api-access-mcnz9\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.227205 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0d25601d-3268-4c04-86c7-fc1a60f6a379\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0d25601d-3268-4c04-86c7-fc1a60f6a379\") pod \"openstack-cell1-galera-0\" (UID: \"351c29ac-7ec2-4c65-95a1-040707a8d01f\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:26 crc kubenswrapper[4997]: I1205 08:36:25.250648 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:27 crc kubenswrapper[4997]: I1205 08:36:27.111718 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 08:36:27 crc kubenswrapper[4997]: I1205 08:36:27.125594 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 08:36:27 crc kubenswrapper[4997]: W1205 08:36:27.136773 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod108e77e9_9f51_4686_bac0_1e0397f33c28.slice/crio-67a99a0f0a1acb35b1a051e309c28417290828d1d0853b166e2a4936c1e7a055 WatchSource:0}: Error finding container 67a99a0f0a1acb35b1a051e309c28417290828d1d0853b166e2a4936c1e7a055: Status 404 returned error can't find the container with id 67a99a0f0a1acb35b1a051e309c28417290828d1d0853b166e2a4936c1e7a055 Dec 05 08:36:28 crc kubenswrapper[4997]: I1205 08:36:28.117426 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"108e77e9-9f51-4686-bac0-1e0397f33c28","Type":"ContainerStarted","Data":"67a99a0f0a1acb35b1a051e309c28417290828d1d0853b166e2a4936c1e7a055"} Dec 05 08:36:28 crc kubenswrapper[4997]: I1205 08:36:28.119380 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"351c29ac-7ec2-4c65-95a1-040707a8d01f","Type":"ContainerStarted","Data":"c2665e0d217e2d3dc21f9c9b7e8e1a8a87b0a8e7e9f958f31141ef28b35cdca1"} Dec 05 08:36:30 crc kubenswrapper[4997]: I1205 08:36:30.749580 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:36:30 crc kubenswrapper[4997]: E1205 08:36:30.750110 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:36:42 crc kubenswrapper[4997]: I1205 08:36:42.748741 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:36:42 crc kubenswrapper[4997]: E1205 08:36:42.749421 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:36:45 crc kubenswrapper[4997]: E1205 08:36:45.305919 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-memcached:65066e8ca260a75886ae57f157049605" Dec 05 08:36:45 crc kubenswrapper[4997]: E1205 08:36:45.306028 4997 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-memcached:65066e8ca260a75886ae57f157049605" Dec 05 08:36:45 crc kubenswrapper[4997]: E1205 08:36:45.306294 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-memcached:65066e8ca260a75886ae57f157049605,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:nf5h5dchc8h568hb6h556h95h58ch5cfh5dh96h658h595hfbh587h58dh644h656h84h6h86h96h5d8h68h676h59bh66hf6h5fch656h79h6fq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vj4r9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(2800a118-2688-4e45-8554-6d613c61d3ad): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 08:36:45 crc kubenswrapper[4997]: E1205 08:36:45.307660 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="2800a118-2688-4e45-8554-6d613c61d3ad" Dec 05 08:36:46 crc kubenswrapper[4997]: E1205 08:36:46.271420 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-memcached:65066e8ca260a75886ae57f157049605\\\"\"" pod="openstack/memcached-0" podUID="2800a118-2688-4e45-8554-6d613c61d3ad" Dec 05 08:36:48 crc kubenswrapper[4997]: I1205 08:36:48.283049 4997 generic.go:334] "Generic (PLEG): container finished" podID="b9355424-f0c2-4a5c-9fce-5a48bc52887a" containerID="38d158200b6b86401b2f099d58e007a6b1b337495674dd1adc42bc344ea8d4c4" exitCode=0 Dec 05 08:36:48 crc kubenswrapper[4997]: I1205 08:36:48.283158 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84496478f-jmzvf" event={"ID":"b9355424-f0c2-4a5c-9fce-5a48bc52887a","Type":"ContainerDied","Data":"38d158200b6b86401b2f099d58e007a6b1b337495674dd1adc42bc344ea8d4c4"} Dec 05 08:36:48 crc kubenswrapper[4997]: I1205 08:36:48.289296 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"351c29ac-7ec2-4c65-95a1-040707a8d01f","Type":"ContainerStarted","Data":"a9c7d4490fe1e7dfea116b79fcd56493b5509de6103c3f5bd6101d2711434128"} Dec 05 08:36:48 crc kubenswrapper[4997]: I1205 08:36:48.291062 4997 generic.go:334] "Generic (PLEG): container finished" podID="69967565-7a54-4904-9cca-41dba04dbef1" containerID="de8e9b5efc1448aa1979e0ba3b5f1eac3d2d92464c143ede24fe0a28990d0b5f" exitCode=0 Dec 05 08:36:48 crc kubenswrapper[4997]: I1205 08:36:48.291119 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7688886755-zxr2l" event={"ID":"69967565-7a54-4904-9cca-41dba04dbef1","Type":"ContainerDied","Data":"de8e9b5efc1448aa1979e0ba3b5f1eac3d2d92464c143ede24fe0a28990d0b5f"} Dec 05 08:36:48 crc kubenswrapper[4997]: I1205 08:36:48.293730 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"108e77e9-9f51-4686-bac0-1e0397f33c28","Type":"ContainerStarted","Data":"a356d4c2ca4c957af74feedbfe5279244ae62dbd35600f079e3b610a039ba8c7"} Dec 05 08:36:49 crc kubenswrapper[4997]: I1205 08:36:49.302200 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7688886755-zxr2l" event={"ID":"69967565-7a54-4904-9cca-41dba04dbef1","Type":"ContainerStarted","Data":"b2d7b05ce9cb4f564021005614067d2f47e0994386cf49d0fc3ff90cafffa6fc"} Dec 05 08:36:49 crc kubenswrapper[4997]: I1205 08:36:49.302347 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7688886755-zxr2l" Dec 05 08:36:49 crc kubenswrapper[4997]: I1205 08:36:49.303812 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"235b6759-15f6-41b7-a3ed-151f56da12f6","Type":"ContainerStarted","Data":"03267fcc99286e2f62f3ae2c1d5fd66aef47dc6f7e6d0e6a11692dec3ebf3e72"} Dec 05 08:36:49 crc kubenswrapper[4997]: I1205 08:36:49.305485 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c67f66b9-d729-463a-8a23-b24870c6ec50","Type":"ContainerStarted","Data":"1989c9041938c4d4c76b60e44b4e3378aa6fded73aa42770e050df62c35aed10"} Dec 05 08:36:49 crc kubenswrapper[4997]: I1205 08:36:49.308102 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84496478f-jmzvf" event={"ID":"b9355424-f0c2-4a5c-9fce-5a48bc52887a","Type":"ContainerStarted","Data":"5091aa1c750377aa763a37282f33b5e3b01fdc56def3ca0ac73d6abf3e4004a7"} Dec 05 08:36:49 crc kubenswrapper[4997]: I1205 08:36:49.323648 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7688886755-zxr2l" podStartSLOduration=2.702975477 podStartE2EDuration="27.323600585s" podCreationTimestamp="2025-12-05 08:36:22 +0000 UTC" firstStartedPulling="2025-12-05 08:36:22.852126695 +0000 UTC m=+6083.381033956" lastFinishedPulling="2025-12-05 08:36:47.472751793 +0000 UTC m=+6108.001659064" observedRunningTime="2025-12-05 08:36:49.319262128 +0000 UTC m=+6109.848169399" watchObservedRunningTime="2025-12-05 08:36:49.323600585 +0000 UTC m=+6109.852507846" Dec 05 08:36:49 crc kubenswrapper[4997]: I1205 08:36:49.338553 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84496478f-jmzvf" podStartSLOduration=2.91996317 podStartE2EDuration="27.338534968s" podCreationTimestamp="2025-12-05 08:36:22 +0000 UTC" firstStartedPulling="2025-12-05 08:36:23.054560875 +0000 UTC m=+6083.583468136" lastFinishedPulling="2025-12-05 08:36:47.473132673 +0000 UTC m=+6108.002039934" observedRunningTime="2025-12-05 08:36:49.333832061 +0000 UTC m=+6109.862739332" watchObservedRunningTime="2025-12-05 08:36:49.338534968 +0000 UTC m=+6109.867442219" Dec 05 08:36:50 crc kubenswrapper[4997]: I1205 08:36:50.314706 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84496478f-jmzvf" Dec 05 08:36:51 crc kubenswrapper[4997]: I1205 08:36:51.326208 4997 generic.go:334] "Generic (PLEG): container finished" podID="108e77e9-9f51-4686-bac0-1e0397f33c28" containerID="a356d4c2ca4c957af74feedbfe5279244ae62dbd35600f079e3b610a039ba8c7" exitCode=0 Dec 05 08:36:51 crc kubenswrapper[4997]: I1205 08:36:51.326287 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"108e77e9-9f51-4686-bac0-1e0397f33c28","Type":"ContainerDied","Data":"a356d4c2ca4c957af74feedbfe5279244ae62dbd35600f079e3b610a039ba8c7"} Dec 05 08:36:51 crc kubenswrapper[4997]: I1205 08:36:51.328096 4997 generic.go:334] "Generic (PLEG): container finished" podID="351c29ac-7ec2-4c65-95a1-040707a8d01f" containerID="a9c7d4490fe1e7dfea116b79fcd56493b5509de6103c3f5bd6101d2711434128" exitCode=0 Dec 05 08:36:51 crc kubenswrapper[4997]: I1205 08:36:51.328169 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"351c29ac-7ec2-4c65-95a1-040707a8d01f","Type":"ContainerDied","Data":"a9c7d4490fe1e7dfea116b79fcd56493b5509de6103c3f5bd6101d2711434128"} Dec 05 08:36:52 crc kubenswrapper[4997]: I1205 08:36:52.336725 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"108e77e9-9f51-4686-bac0-1e0397f33c28","Type":"ContainerStarted","Data":"00fbca384e39069783eacc0c47f4f995b62b6fcddeb13ca1c073b20ac060eb72"} Dec 05 08:36:52 crc kubenswrapper[4997]: I1205 08:36:52.339035 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"351c29ac-7ec2-4c65-95a1-040707a8d01f","Type":"ContainerStarted","Data":"3787e35585e1806a3750b18dea14ba9a4cee62d0a5daa5c45b21f9740745aa38"} Dec 05 08:36:52 crc kubenswrapper[4997]: I1205 08:36:52.359844 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=9.060065403 podStartE2EDuration="29.359820655s" podCreationTimestamp="2025-12-05 08:36:23 +0000 UTC" firstStartedPulling="2025-12-05 08:36:27.142408806 +0000 UTC m=+6087.671316067" lastFinishedPulling="2025-12-05 08:36:47.442164058 +0000 UTC m=+6107.971071319" observedRunningTime="2025-12-05 08:36:52.355648642 +0000 UTC m=+6112.884555923" watchObservedRunningTime="2025-12-05 08:36:52.359820655 +0000 UTC m=+6112.888727916" Dec 05 08:36:52 crc kubenswrapper[4997]: I1205 08:36:52.381944 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=9.03326806 podStartE2EDuration="29.381926661s" podCreationTimestamp="2025-12-05 08:36:23 +0000 UTC" firstStartedPulling="2025-12-05 08:36:27.124130573 +0000 UTC m=+6087.653037844" lastFinishedPulling="2025-12-05 08:36:47.472789184 +0000 UTC m=+6108.001696445" observedRunningTime="2025-12-05 08:36:52.375292022 +0000 UTC m=+6112.904199303" watchObservedRunningTime="2025-12-05 08:36:52.381926661 +0000 UTC m=+6112.910833922" Dec 05 08:36:54 crc kubenswrapper[4997]: I1205 08:36:54.458325 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 05 08:36:54 crc kubenswrapper[4997]: I1205 08:36:54.458693 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 05 08:36:55 crc kubenswrapper[4997]: I1205 08:36:55.251692 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:55 crc kubenswrapper[4997]: I1205 08:36:55.251745 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:56 crc kubenswrapper[4997]: I1205 08:36:56.786770 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 05 08:36:56 crc kubenswrapper[4997]: I1205 08:36:56.864373 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 05 08:36:57 crc kubenswrapper[4997]: I1205 08:36:57.526867 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7688886755-zxr2l" Dec 05 08:36:57 crc kubenswrapper[4997]: I1205 08:36:57.774090 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:36:57 crc kubenswrapper[4997]: E1205 08:36:57.774354 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:36:57 crc kubenswrapper[4997]: I1205 08:36:57.786136 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84496478f-jmzvf" Dec 05 08:36:57 crc kubenswrapper[4997]: I1205 08:36:57.841962 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7688886755-zxr2l"] Dec 05 08:36:58 crc kubenswrapper[4997]: I1205 08:36:58.384415 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"2800a118-2688-4e45-8554-6d613c61d3ad","Type":"ContainerStarted","Data":"235d9ff636d34172cd7857d1224386a0dc0adf882009f4ed1ba005a72771017d"} Dec 05 08:36:58 crc kubenswrapper[4997]: I1205 08:36:58.385074 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 05 08:36:58 crc kubenswrapper[4997]: I1205 08:36:58.386190 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7688886755-zxr2l" podUID="69967565-7a54-4904-9cca-41dba04dbef1" containerName="dnsmasq-dns" containerID="cri-o://b2d7b05ce9cb4f564021005614067d2f47e0994386cf49d0fc3ff90cafffa6fc" gracePeriod=10 Dec 05 08:36:58 crc kubenswrapper[4997]: I1205 08:36:58.407268 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.022111106 podStartE2EDuration="35.407250532s" podCreationTimestamp="2025-12-05 08:36:23 +0000 UTC" firstStartedPulling="2025-12-05 08:36:24.614640373 +0000 UTC m=+6085.143547634" lastFinishedPulling="2025-12-05 08:36:57.999779799 +0000 UTC m=+6118.528687060" observedRunningTime="2025-12-05 08:36:58.401971389 +0000 UTC m=+6118.930878670" watchObservedRunningTime="2025-12-05 08:36:58.407250532 +0000 UTC m=+6118.936157793" Dec 05 08:36:58 crc kubenswrapper[4997]: I1205 08:36:58.775212 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7688886755-zxr2l" Dec 05 08:36:58 crc kubenswrapper[4997]: I1205 08:36:58.914375 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69967565-7a54-4904-9cca-41dba04dbef1-config\") pod \"69967565-7a54-4904-9cca-41dba04dbef1\" (UID: \"69967565-7a54-4904-9cca-41dba04dbef1\") " Dec 05 08:36:58 crc kubenswrapper[4997]: I1205 08:36:58.914498 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmghs\" (UniqueName: \"kubernetes.io/projected/69967565-7a54-4904-9cca-41dba04dbef1-kube-api-access-jmghs\") pod \"69967565-7a54-4904-9cca-41dba04dbef1\" (UID: \"69967565-7a54-4904-9cca-41dba04dbef1\") " Dec 05 08:36:58 crc kubenswrapper[4997]: I1205 08:36:58.914586 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69967565-7a54-4904-9cca-41dba04dbef1-dns-svc\") pod \"69967565-7a54-4904-9cca-41dba04dbef1\" (UID: \"69967565-7a54-4904-9cca-41dba04dbef1\") " Dec 05 08:36:58 crc kubenswrapper[4997]: I1205 08:36:58.919532 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69967565-7a54-4904-9cca-41dba04dbef1-kube-api-access-jmghs" (OuterVolumeSpecName: "kube-api-access-jmghs") pod "69967565-7a54-4904-9cca-41dba04dbef1" (UID: "69967565-7a54-4904-9cca-41dba04dbef1"). InnerVolumeSpecName "kube-api-access-jmghs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:36:58 crc kubenswrapper[4997]: I1205 08:36:58.950649 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69967565-7a54-4904-9cca-41dba04dbef1-config" (OuterVolumeSpecName: "config") pod "69967565-7a54-4904-9cca-41dba04dbef1" (UID: "69967565-7a54-4904-9cca-41dba04dbef1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:58 crc kubenswrapper[4997]: I1205 08:36:58.953525 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69967565-7a54-4904-9cca-41dba04dbef1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "69967565-7a54-4904-9cca-41dba04dbef1" (UID: "69967565-7a54-4904-9cca-41dba04dbef1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.016120 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69967565-7a54-4904-9cca-41dba04dbef1-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.016164 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmghs\" (UniqueName: \"kubernetes.io/projected/69967565-7a54-4904-9cca-41dba04dbef1-kube-api-access-jmghs\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.016186 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69967565-7a54-4904-9cca-41dba04dbef1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.325799 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.391913 4997 generic.go:334] "Generic (PLEG): container finished" podID="69967565-7a54-4904-9cca-41dba04dbef1" containerID="b2d7b05ce9cb4f564021005614067d2f47e0994386cf49d0fc3ff90cafffa6fc" exitCode=0 Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.392001 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7688886755-zxr2l" Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.392008 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7688886755-zxr2l" event={"ID":"69967565-7a54-4904-9cca-41dba04dbef1","Type":"ContainerDied","Data":"b2d7b05ce9cb4f564021005614067d2f47e0994386cf49d0fc3ff90cafffa6fc"} Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.392079 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7688886755-zxr2l" event={"ID":"69967565-7a54-4904-9cca-41dba04dbef1","Type":"ContainerDied","Data":"761cf8c0894921937646ded4b0cb69c3f454dd4917ffcddd3e376a30e1bf82b8"} Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.392108 4997 scope.go:117] "RemoveContainer" containerID="b2d7b05ce9cb4f564021005614067d2f47e0994386cf49d0fc3ff90cafffa6fc" Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.406266 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.407550 4997 scope.go:117] "RemoveContainer" containerID="de8e9b5efc1448aa1979e0ba3b5f1eac3d2d92464c143ede24fe0a28990d0b5f" Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.424802 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7688886755-zxr2l"] Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.428592 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7688886755-zxr2l"] Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.446294 4997 scope.go:117] "RemoveContainer" containerID="b2d7b05ce9cb4f564021005614067d2f47e0994386cf49d0fc3ff90cafffa6fc" Dec 05 08:36:59 crc kubenswrapper[4997]: E1205 08:36:59.446748 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2d7b05ce9cb4f564021005614067d2f47e0994386cf49d0fc3ff90cafffa6fc\": container with ID starting with b2d7b05ce9cb4f564021005614067d2f47e0994386cf49d0fc3ff90cafffa6fc not found: ID does not exist" containerID="b2d7b05ce9cb4f564021005614067d2f47e0994386cf49d0fc3ff90cafffa6fc" Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.446794 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2d7b05ce9cb4f564021005614067d2f47e0994386cf49d0fc3ff90cafffa6fc"} err="failed to get container status \"b2d7b05ce9cb4f564021005614067d2f47e0994386cf49d0fc3ff90cafffa6fc\": rpc error: code = NotFound desc = could not find container \"b2d7b05ce9cb4f564021005614067d2f47e0994386cf49d0fc3ff90cafffa6fc\": container with ID starting with b2d7b05ce9cb4f564021005614067d2f47e0994386cf49d0fc3ff90cafffa6fc not found: ID does not exist" Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.446829 4997 scope.go:117] "RemoveContainer" containerID="de8e9b5efc1448aa1979e0ba3b5f1eac3d2d92464c143ede24fe0a28990d0b5f" Dec 05 08:36:59 crc kubenswrapper[4997]: E1205 08:36:59.447069 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de8e9b5efc1448aa1979e0ba3b5f1eac3d2d92464c143ede24fe0a28990d0b5f\": container with ID starting with de8e9b5efc1448aa1979e0ba3b5f1eac3d2d92464c143ede24fe0a28990d0b5f not found: ID does not exist" containerID="de8e9b5efc1448aa1979e0ba3b5f1eac3d2d92464c143ede24fe0a28990d0b5f" Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.447110 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de8e9b5efc1448aa1979e0ba3b5f1eac3d2d92464c143ede24fe0a28990d0b5f"} err="failed to get container status \"de8e9b5efc1448aa1979e0ba3b5f1eac3d2d92464c143ede24fe0a28990d0b5f\": rpc error: code = NotFound desc = could not find container \"de8e9b5efc1448aa1979e0ba3b5f1eac3d2d92464c143ede24fe0a28990d0b5f\": container with ID starting with de8e9b5efc1448aa1979e0ba3b5f1eac3d2d92464c143ede24fe0a28990d0b5f not found: ID does not exist" Dec 05 08:36:59 crc kubenswrapper[4997]: I1205 08:36:59.758528 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69967565-7a54-4904-9cca-41dba04dbef1" path="/var/lib/kubelet/pods/69967565-7a54-4904-9cca-41dba04dbef1/volumes" Dec 05 08:37:04 crc kubenswrapper[4997]: I1205 08:37:04.333885 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 05 08:37:05 crc kubenswrapper[4997]: I1205 08:37:05.384162 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6dhq5"] Dec 05 08:37:05 crc kubenswrapper[4997]: E1205 08:37:05.384500 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69967565-7a54-4904-9cca-41dba04dbef1" containerName="init" Dec 05 08:37:05 crc kubenswrapper[4997]: I1205 08:37:05.384519 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="69967565-7a54-4904-9cca-41dba04dbef1" containerName="init" Dec 05 08:37:05 crc kubenswrapper[4997]: E1205 08:37:05.384536 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69967565-7a54-4904-9cca-41dba04dbef1" containerName="dnsmasq-dns" Dec 05 08:37:05 crc kubenswrapper[4997]: I1205 08:37:05.384544 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="69967565-7a54-4904-9cca-41dba04dbef1" containerName="dnsmasq-dns" Dec 05 08:37:05 crc kubenswrapper[4997]: I1205 08:37:05.384719 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="69967565-7a54-4904-9cca-41dba04dbef1" containerName="dnsmasq-dns" Dec 05 08:37:05 crc kubenswrapper[4997]: I1205 08:37:05.385765 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6dhq5" Dec 05 08:37:05 crc kubenswrapper[4997]: I1205 08:37:05.405029 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6dhq5"] Dec 05 08:37:05 crc kubenswrapper[4997]: I1205 08:37:05.511337 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g8qw\" (UniqueName: \"kubernetes.io/projected/af9739e2-7a99-48ca-9d42-cc1d33552e8c-kube-api-access-5g8qw\") pod \"community-operators-6dhq5\" (UID: \"af9739e2-7a99-48ca-9d42-cc1d33552e8c\") " pod="openshift-marketplace/community-operators-6dhq5" Dec 05 08:37:05 crc kubenswrapper[4997]: I1205 08:37:05.511402 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9739e2-7a99-48ca-9d42-cc1d33552e8c-catalog-content\") pod \"community-operators-6dhq5\" (UID: \"af9739e2-7a99-48ca-9d42-cc1d33552e8c\") " pod="openshift-marketplace/community-operators-6dhq5" Dec 05 08:37:05 crc kubenswrapper[4997]: I1205 08:37:05.511458 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9739e2-7a99-48ca-9d42-cc1d33552e8c-utilities\") pod \"community-operators-6dhq5\" (UID: \"af9739e2-7a99-48ca-9d42-cc1d33552e8c\") " pod="openshift-marketplace/community-operators-6dhq5" Dec 05 08:37:05 crc kubenswrapper[4997]: I1205 08:37:05.612325 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9739e2-7a99-48ca-9d42-cc1d33552e8c-utilities\") pod \"community-operators-6dhq5\" (UID: \"af9739e2-7a99-48ca-9d42-cc1d33552e8c\") " pod="openshift-marketplace/community-operators-6dhq5" Dec 05 08:37:05 crc kubenswrapper[4997]: I1205 08:37:05.612457 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g8qw\" (UniqueName: \"kubernetes.io/projected/af9739e2-7a99-48ca-9d42-cc1d33552e8c-kube-api-access-5g8qw\") pod \"community-operators-6dhq5\" (UID: \"af9739e2-7a99-48ca-9d42-cc1d33552e8c\") " pod="openshift-marketplace/community-operators-6dhq5" Dec 05 08:37:05 crc kubenswrapper[4997]: I1205 08:37:05.612503 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9739e2-7a99-48ca-9d42-cc1d33552e8c-catalog-content\") pod \"community-operators-6dhq5\" (UID: \"af9739e2-7a99-48ca-9d42-cc1d33552e8c\") " pod="openshift-marketplace/community-operators-6dhq5" Dec 05 08:37:05 crc kubenswrapper[4997]: I1205 08:37:05.612907 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9739e2-7a99-48ca-9d42-cc1d33552e8c-utilities\") pod \"community-operators-6dhq5\" (UID: \"af9739e2-7a99-48ca-9d42-cc1d33552e8c\") " pod="openshift-marketplace/community-operators-6dhq5" Dec 05 08:37:05 crc kubenswrapper[4997]: I1205 08:37:05.612937 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9739e2-7a99-48ca-9d42-cc1d33552e8c-catalog-content\") pod \"community-operators-6dhq5\" (UID: \"af9739e2-7a99-48ca-9d42-cc1d33552e8c\") " pod="openshift-marketplace/community-operators-6dhq5" Dec 05 08:37:05 crc kubenswrapper[4997]: I1205 08:37:05.637686 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g8qw\" (UniqueName: \"kubernetes.io/projected/af9739e2-7a99-48ca-9d42-cc1d33552e8c-kube-api-access-5g8qw\") pod \"community-operators-6dhq5\" (UID: \"af9739e2-7a99-48ca-9d42-cc1d33552e8c\") " pod="openshift-marketplace/community-operators-6dhq5" Dec 05 08:37:05 crc kubenswrapper[4997]: I1205 08:37:05.710763 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6dhq5" Dec 05 08:37:06 crc kubenswrapper[4997]: I1205 08:37:06.213740 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6dhq5"] Dec 05 08:37:06 crc kubenswrapper[4997]: W1205 08:37:06.214160 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf9739e2_7a99_48ca_9d42_cc1d33552e8c.slice/crio-cb5aa9b85dd3404985bed026ef39ea99ba4e4d6f08def885811d48e796db3987 WatchSource:0}: Error finding container cb5aa9b85dd3404985bed026ef39ea99ba4e4d6f08def885811d48e796db3987: Status 404 returned error can't find the container with id cb5aa9b85dd3404985bed026ef39ea99ba4e4d6f08def885811d48e796db3987 Dec 05 08:37:06 crc kubenswrapper[4997]: I1205 08:37:06.449265 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6dhq5" event={"ID":"af9739e2-7a99-48ca-9d42-cc1d33552e8c","Type":"ContainerStarted","Data":"cb5aa9b85dd3404985bed026ef39ea99ba4e4d6f08def885811d48e796db3987"} Dec 05 08:37:08 crc kubenswrapper[4997]: I1205 08:37:08.748564 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:37:08 crc kubenswrapper[4997]: E1205 08:37:08.749105 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:37:11 crc kubenswrapper[4997]: I1205 08:37:11.488741 4997 generic.go:334] "Generic (PLEG): container finished" podID="af9739e2-7a99-48ca-9d42-cc1d33552e8c" containerID="95cd736cc79df4c77c6e684c9752621d04e2ba71fa6e7d1df7f3203e54353acd" exitCode=0 Dec 05 08:37:11 crc kubenswrapper[4997]: I1205 08:37:11.488797 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6dhq5" event={"ID":"af9739e2-7a99-48ca-9d42-cc1d33552e8c","Type":"ContainerDied","Data":"95cd736cc79df4c77c6e684c9752621d04e2ba71fa6e7d1df7f3203e54353acd"} Dec 05 08:37:12 crc kubenswrapper[4997]: I1205 08:37:12.501491 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6dhq5" event={"ID":"af9739e2-7a99-48ca-9d42-cc1d33552e8c","Type":"ContainerStarted","Data":"f6313e945a0a49e676fbc15f023db6778ea5029e92e5b6d28c880c453c3dc11e"} Dec 05 08:37:13 crc kubenswrapper[4997]: I1205 08:37:13.509725 4997 generic.go:334] "Generic (PLEG): container finished" podID="af9739e2-7a99-48ca-9d42-cc1d33552e8c" containerID="f6313e945a0a49e676fbc15f023db6778ea5029e92e5b6d28c880c453c3dc11e" exitCode=0 Dec 05 08:37:13 crc kubenswrapper[4997]: I1205 08:37:13.509758 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6dhq5" event={"ID":"af9739e2-7a99-48ca-9d42-cc1d33552e8c","Type":"ContainerDied","Data":"f6313e945a0a49e676fbc15f023db6778ea5029e92e5b6d28c880c453c3dc11e"} Dec 05 08:37:14 crc kubenswrapper[4997]: I1205 08:37:14.519020 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6dhq5" event={"ID":"af9739e2-7a99-48ca-9d42-cc1d33552e8c","Type":"ContainerStarted","Data":"2dc9b9afa07ecb15a9afe9a68f448564c9aac270a8f589110e287ab6b37d66f9"} Dec 05 08:37:14 crc kubenswrapper[4997]: I1205 08:37:14.558116 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6dhq5" podStartSLOduration=7.146174457 podStartE2EDuration="9.558079824s" podCreationTimestamp="2025-12-05 08:37:05 +0000 UTC" firstStartedPulling="2025-12-05 08:37:11.490229931 +0000 UTC m=+6132.019137202" lastFinishedPulling="2025-12-05 08:37:13.902135308 +0000 UTC m=+6134.431042569" observedRunningTime="2025-12-05 08:37:14.538705821 +0000 UTC m=+6135.067613092" watchObservedRunningTime="2025-12-05 08:37:14.558079824 +0000 UTC m=+6135.086987075" Dec 05 08:37:15 crc kubenswrapper[4997]: I1205 08:37:15.711671 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6dhq5" Dec 05 08:37:15 crc kubenswrapper[4997]: I1205 08:37:15.711758 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6dhq5" Dec 05 08:37:16 crc kubenswrapper[4997]: I1205 08:37:16.761972 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-6dhq5" podUID="af9739e2-7a99-48ca-9d42-cc1d33552e8c" containerName="registry-server" probeResult="failure" output=< Dec 05 08:37:16 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 08:37:16 crc kubenswrapper[4997]: > Dec 05 08:37:21 crc kubenswrapper[4997]: I1205 08:37:21.573678 4997 generic.go:334] "Generic (PLEG): container finished" podID="235b6759-15f6-41b7-a3ed-151f56da12f6" containerID="03267fcc99286e2f62f3ae2c1d5fd66aef47dc6f7e6d0e6a11692dec3ebf3e72" exitCode=0 Dec 05 08:37:21 crc kubenswrapper[4997]: I1205 08:37:21.573787 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"235b6759-15f6-41b7-a3ed-151f56da12f6","Type":"ContainerDied","Data":"03267fcc99286e2f62f3ae2c1d5fd66aef47dc6f7e6d0e6a11692dec3ebf3e72"} Dec 05 08:37:21 crc kubenswrapper[4997]: I1205 08:37:21.575726 4997 generic.go:334] "Generic (PLEG): container finished" podID="c67f66b9-d729-463a-8a23-b24870c6ec50" containerID="1989c9041938c4d4c76b60e44b4e3378aa6fded73aa42770e050df62c35aed10" exitCode=0 Dec 05 08:37:21 crc kubenswrapper[4997]: I1205 08:37:21.575772 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c67f66b9-d729-463a-8a23-b24870c6ec50","Type":"ContainerDied","Data":"1989c9041938c4d4c76b60e44b4e3378aa6fded73aa42770e050df62c35aed10"} Dec 05 08:37:22 crc kubenswrapper[4997]: I1205 08:37:22.586274 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"235b6759-15f6-41b7-a3ed-151f56da12f6","Type":"ContainerStarted","Data":"a5c814d2cace6349b94bdb8779fc1b179ed24c1bb28d6c852ff2e0ba495d846d"} Dec 05 08:37:22 crc kubenswrapper[4997]: I1205 08:37:22.586871 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 05 08:37:22 crc kubenswrapper[4997]: I1205 08:37:22.589863 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c67f66b9-d729-463a-8a23-b24870c6ec50","Type":"ContainerStarted","Data":"4c58f4722a4ab72c2bc5edb4011ca1e950a13f92f0c2113fe4aeeb755e735e1d"} Dec 05 08:37:22 crc kubenswrapper[4997]: I1205 08:37:22.590055 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:22 crc kubenswrapper[4997]: I1205 08:37:22.610628 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.205019614 podStartE2EDuration="1m0.610580933s" podCreationTimestamp="2025-12-05 08:36:22 +0000 UTC" firstStartedPulling="2025-12-05 08:36:24.005916871 +0000 UTC m=+6084.534824132" lastFinishedPulling="2025-12-05 08:36:47.41147819 +0000 UTC m=+6107.940385451" observedRunningTime="2025-12-05 08:37:22.605803793 +0000 UTC m=+6143.134711074" watchObservedRunningTime="2025-12-05 08:37:22.610580933 +0000 UTC m=+6143.139488194" Dec 05 08:37:22 crc kubenswrapper[4997]: I1205 08:37:22.630113 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.764357064 podStartE2EDuration="1m0.630088089s" podCreationTimestamp="2025-12-05 08:36:22 +0000 UTC" firstStartedPulling="2025-12-05 08:36:24.500803702 +0000 UTC m=+6085.029710963" lastFinishedPulling="2025-12-05 08:36:47.366534727 +0000 UTC m=+6107.895441988" observedRunningTime="2025-12-05 08:37:22.624370165 +0000 UTC m=+6143.153277436" watchObservedRunningTime="2025-12-05 08:37:22.630088089 +0000 UTC m=+6143.158995370" Dec 05 08:37:22 crc kubenswrapper[4997]: I1205 08:37:22.749332 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:37:22 crc kubenswrapper[4997]: E1205 08:37:22.749538 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:37:25 crc kubenswrapper[4997]: I1205 08:37:25.779469 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6dhq5" Dec 05 08:37:25 crc kubenswrapper[4997]: I1205 08:37:25.822673 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6dhq5" Dec 05 08:37:26 crc kubenswrapper[4997]: I1205 08:37:26.012869 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6dhq5"] Dec 05 08:37:27 crc kubenswrapper[4997]: I1205 08:37:27.623682 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6dhq5" podUID="af9739e2-7a99-48ca-9d42-cc1d33552e8c" containerName="registry-server" containerID="cri-o://2dc9b9afa07ecb15a9afe9a68f448564c9aac270a8f589110e287ab6b37d66f9" gracePeriod=2 Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.053189 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6dhq5" Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.179001 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9739e2-7a99-48ca-9d42-cc1d33552e8c-utilities\") pod \"af9739e2-7a99-48ca-9d42-cc1d33552e8c\" (UID: \"af9739e2-7a99-48ca-9d42-cc1d33552e8c\") " Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.179102 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5g8qw\" (UniqueName: \"kubernetes.io/projected/af9739e2-7a99-48ca-9d42-cc1d33552e8c-kube-api-access-5g8qw\") pod \"af9739e2-7a99-48ca-9d42-cc1d33552e8c\" (UID: \"af9739e2-7a99-48ca-9d42-cc1d33552e8c\") " Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.179277 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9739e2-7a99-48ca-9d42-cc1d33552e8c-catalog-content\") pod \"af9739e2-7a99-48ca-9d42-cc1d33552e8c\" (UID: \"af9739e2-7a99-48ca-9d42-cc1d33552e8c\") " Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.179804 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af9739e2-7a99-48ca-9d42-cc1d33552e8c-utilities" (OuterVolumeSpecName: "utilities") pod "af9739e2-7a99-48ca-9d42-cc1d33552e8c" (UID: "af9739e2-7a99-48ca-9d42-cc1d33552e8c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.195464 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af9739e2-7a99-48ca-9d42-cc1d33552e8c-kube-api-access-5g8qw" (OuterVolumeSpecName: "kube-api-access-5g8qw") pod "af9739e2-7a99-48ca-9d42-cc1d33552e8c" (UID: "af9739e2-7a99-48ca-9d42-cc1d33552e8c"). InnerVolumeSpecName "kube-api-access-5g8qw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.230999 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af9739e2-7a99-48ca-9d42-cc1d33552e8c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af9739e2-7a99-48ca-9d42-cc1d33552e8c" (UID: "af9739e2-7a99-48ca-9d42-cc1d33552e8c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.281605 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9739e2-7a99-48ca-9d42-cc1d33552e8c-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.281667 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5g8qw\" (UniqueName: \"kubernetes.io/projected/af9739e2-7a99-48ca-9d42-cc1d33552e8c-kube-api-access-5g8qw\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.281679 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9739e2-7a99-48ca-9d42-cc1d33552e8c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.632914 4997 generic.go:334] "Generic (PLEG): container finished" podID="af9739e2-7a99-48ca-9d42-cc1d33552e8c" containerID="2dc9b9afa07ecb15a9afe9a68f448564c9aac270a8f589110e287ab6b37d66f9" exitCode=0 Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.632951 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6dhq5" Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.632968 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6dhq5" event={"ID":"af9739e2-7a99-48ca-9d42-cc1d33552e8c","Type":"ContainerDied","Data":"2dc9b9afa07ecb15a9afe9a68f448564c9aac270a8f589110e287ab6b37d66f9"} Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.634005 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6dhq5" event={"ID":"af9739e2-7a99-48ca-9d42-cc1d33552e8c","Type":"ContainerDied","Data":"cb5aa9b85dd3404985bed026ef39ea99ba4e4d6f08def885811d48e796db3987"} Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.634061 4997 scope.go:117] "RemoveContainer" containerID="2dc9b9afa07ecb15a9afe9a68f448564c9aac270a8f589110e287ab6b37d66f9" Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.657997 4997 scope.go:117] "RemoveContainer" containerID="f6313e945a0a49e676fbc15f023db6778ea5029e92e5b6d28c880c453c3dc11e" Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.680456 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6dhq5"] Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.682177 4997 scope.go:117] "RemoveContainer" containerID="95cd736cc79df4c77c6e684c9752621d04e2ba71fa6e7d1df7f3203e54353acd" Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.688335 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6dhq5"] Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.708404 4997 scope.go:117] "RemoveContainer" containerID="2dc9b9afa07ecb15a9afe9a68f448564c9aac270a8f589110e287ab6b37d66f9" Dec 05 08:37:28 crc kubenswrapper[4997]: E1205 08:37:28.708984 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dc9b9afa07ecb15a9afe9a68f448564c9aac270a8f589110e287ab6b37d66f9\": container with ID starting with 2dc9b9afa07ecb15a9afe9a68f448564c9aac270a8f589110e287ab6b37d66f9 not found: ID does not exist" containerID="2dc9b9afa07ecb15a9afe9a68f448564c9aac270a8f589110e287ab6b37d66f9" Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.709029 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc9b9afa07ecb15a9afe9a68f448564c9aac270a8f589110e287ab6b37d66f9"} err="failed to get container status \"2dc9b9afa07ecb15a9afe9a68f448564c9aac270a8f589110e287ab6b37d66f9\": rpc error: code = NotFound desc = could not find container \"2dc9b9afa07ecb15a9afe9a68f448564c9aac270a8f589110e287ab6b37d66f9\": container with ID starting with 2dc9b9afa07ecb15a9afe9a68f448564c9aac270a8f589110e287ab6b37d66f9 not found: ID does not exist" Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.709057 4997 scope.go:117] "RemoveContainer" containerID="f6313e945a0a49e676fbc15f023db6778ea5029e92e5b6d28c880c453c3dc11e" Dec 05 08:37:28 crc kubenswrapper[4997]: E1205 08:37:28.709589 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6313e945a0a49e676fbc15f023db6778ea5029e92e5b6d28c880c453c3dc11e\": container with ID starting with f6313e945a0a49e676fbc15f023db6778ea5029e92e5b6d28c880c453c3dc11e not found: ID does not exist" containerID="f6313e945a0a49e676fbc15f023db6778ea5029e92e5b6d28c880c453c3dc11e" Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.709632 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6313e945a0a49e676fbc15f023db6778ea5029e92e5b6d28c880c453c3dc11e"} err="failed to get container status \"f6313e945a0a49e676fbc15f023db6778ea5029e92e5b6d28c880c453c3dc11e\": rpc error: code = NotFound desc = could not find container \"f6313e945a0a49e676fbc15f023db6778ea5029e92e5b6d28c880c453c3dc11e\": container with ID starting with f6313e945a0a49e676fbc15f023db6778ea5029e92e5b6d28c880c453c3dc11e not found: ID does not exist" Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.709653 4997 scope.go:117] "RemoveContainer" containerID="95cd736cc79df4c77c6e684c9752621d04e2ba71fa6e7d1df7f3203e54353acd" Dec 05 08:37:28 crc kubenswrapper[4997]: E1205 08:37:28.709985 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95cd736cc79df4c77c6e684c9752621d04e2ba71fa6e7d1df7f3203e54353acd\": container with ID starting with 95cd736cc79df4c77c6e684c9752621d04e2ba71fa6e7d1df7f3203e54353acd not found: ID does not exist" containerID="95cd736cc79df4c77c6e684c9752621d04e2ba71fa6e7d1df7f3203e54353acd" Dec 05 08:37:28 crc kubenswrapper[4997]: I1205 08:37:28.710016 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95cd736cc79df4c77c6e684c9752621d04e2ba71fa6e7d1df7f3203e54353acd"} err="failed to get container status \"95cd736cc79df4c77c6e684c9752621d04e2ba71fa6e7d1df7f3203e54353acd\": rpc error: code = NotFound desc = could not find container \"95cd736cc79df4c77c6e684c9752621d04e2ba71fa6e7d1df7f3203e54353acd\": container with ID starting with 95cd736cc79df4c77c6e684c9752621d04e2ba71fa6e7d1df7f3203e54353acd not found: ID does not exist" Dec 05 08:37:28 crc kubenswrapper[4997]: E1205 08:37:28.736266 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf9739e2_7a99_48ca_9d42_cc1d33552e8c.slice\": RecentStats: unable to find data in memory cache]" Dec 05 08:37:29 crc kubenswrapper[4997]: I1205 08:37:29.758456 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af9739e2-7a99-48ca-9d42-cc1d33552e8c" path="/var/lib/kubelet/pods/af9739e2-7a99-48ca-9d42-cc1d33552e8c/volumes" Dec 05 08:37:33 crc kubenswrapper[4997]: I1205 08:37:33.651601 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 05 08:37:33 crc kubenswrapper[4997]: I1205 08:37:33.946751 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:34 crc kubenswrapper[4997]: I1205 08:37:34.749442 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:37:34 crc kubenswrapper[4997]: E1205 08:37:34.749672 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.176454 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-778d75ccf7-nv7g4"] Dec 05 08:37:39 crc kubenswrapper[4997]: E1205 08:37:39.177243 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af9739e2-7a99-48ca-9d42-cc1d33552e8c" containerName="registry-server" Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.177280 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="af9739e2-7a99-48ca-9d42-cc1d33552e8c" containerName="registry-server" Dec 05 08:37:39 crc kubenswrapper[4997]: E1205 08:37:39.177316 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af9739e2-7a99-48ca-9d42-cc1d33552e8c" containerName="extract-content" Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.177328 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="af9739e2-7a99-48ca-9d42-cc1d33552e8c" containerName="extract-content" Dec 05 08:37:39 crc kubenswrapper[4997]: E1205 08:37:39.177353 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af9739e2-7a99-48ca-9d42-cc1d33552e8c" containerName="extract-utilities" Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.177368 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="af9739e2-7a99-48ca-9d42-cc1d33552e8c" containerName="extract-utilities" Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.177881 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="af9739e2-7a99-48ca-9d42-cc1d33552e8c" containerName="registry-server" Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.178988 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.217591 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-778d75ccf7-nv7g4"] Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.252483 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9acdab75-6b5a-4252-a746-03f7d3b5493e-dns-svc\") pod \"dnsmasq-dns-778d75ccf7-nv7g4\" (UID: \"9acdab75-6b5a-4252-a746-03f7d3b5493e\") " pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.252627 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rf95\" (UniqueName: \"kubernetes.io/projected/9acdab75-6b5a-4252-a746-03f7d3b5493e-kube-api-access-7rf95\") pod \"dnsmasq-dns-778d75ccf7-nv7g4\" (UID: \"9acdab75-6b5a-4252-a746-03f7d3b5493e\") " pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.252687 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9acdab75-6b5a-4252-a746-03f7d3b5493e-config\") pod \"dnsmasq-dns-778d75ccf7-nv7g4\" (UID: \"9acdab75-6b5a-4252-a746-03f7d3b5493e\") " pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.354704 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rf95\" (UniqueName: \"kubernetes.io/projected/9acdab75-6b5a-4252-a746-03f7d3b5493e-kube-api-access-7rf95\") pod \"dnsmasq-dns-778d75ccf7-nv7g4\" (UID: \"9acdab75-6b5a-4252-a746-03f7d3b5493e\") " pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.355101 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9acdab75-6b5a-4252-a746-03f7d3b5493e-config\") pod \"dnsmasq-dns-778d75ccf7-nv7g4\" (UID: \"9acdab75-6b5a-4252-a746-03f7d3b5493e\") " pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.355145 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9acdab75-6b5a-4252-a746-03f7d3b5493e-dns-svc\") pod \"dnsmasq-dns-778d75ccf7-nv7g4\" (UID: \"9acdab75-6b5a-4252-a746-03f7d3b5493e\") " pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.356235 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9acdab75-6b5a-4252-a746-03f7d3b5493e-dns-svc\") pod \"dnsmasq-dns-778d75ccf7-nv7g4\" (UID: \"9acdab75-6b5a-4252-a746-03f7d3b5493e\") " pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.357897 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9acdab75-6b5a-4252-a746-03f7d3b5493e-config\") pod \"dnsmasq-dns-778d75ccf7-nv7g4\" (UID: \"9acdab75-6b5a-4252-a746-03f7d3b5493e\") " pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.389577 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rf95\" (UniqueName: \"kubernetes.io/projected/9acdab75-6b5a-4252-a746-03f7d3b5493e-kube-api-access-7rf95\") pod \"dnsmasq-dns-778d75ccf7-nv7g4\" (UID: \"9acdab75-6b5a-4252-a746-03f7d3b5493e\") " pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.495301 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.744034 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-778d75ccf7-nv7g4"] Dec 05 08:37:39 crc kubenswrapper[4997]: I1205 08:37:39.859480 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:37:40 crc kubenswrapper[4997]: I1205 08:37:40.472659 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:37:40 crc kubenswrapper[4997]: I1205 08:37:40.723949 4997 generic.go:334] "Generic (PLEG): container finished" podID="9acdab75-6b5a-4252-a746-03f7d3b5493e" containerID="a7fc8e5261025fc95708fd530d4697762c97e37594e8098b3548d3b1244d6816" exitCode=0 Dec 05 08:37:40 crc kubenswrapper[4997]: I1205 08:37:40.724019 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" event={"ID":"9acdab75-6b5a-4252-a746-03f7d3b5493e","Type":"ContainerDied","Data":"a7fc8e5261025fc95708fd530d4697762c97e37594e8098b3548d3b1244d6816"} Dec 05 08:37:40 crc kubenswrapper[4997]: I1205 08:37:40.724187 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" event={"ID":"9acdab75-6b5a-4252-a746-03f7d3b5493e","Type":"ContainerStarted","Data":"372bb81a7dcfd924f98e3d6332da8c67626d26660faec3b7882899235e820b7b"} Dec 05 08:37:41 crc kubenswrapper[4997]: I1205 08:37:41.614848 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="235b6759-15f6-41b7-a3ed-151f56da12f6" containerName="rabbitmq" containerID="cri-o://a5c814d2cace6349b94bdb8779fc1b179ed24c1bb28d6c852ff2e0ba495d846d" gracePeriod=604799 Dec 05 08:37:41 crc kubenswrapper[4997]: I1205 08:37:41.731875 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" event={"ID":"9acdab75-6b5a-4252-a746-03f7d3b5493e","Type":"ContainerStarted","Data":"4a19c7414bb6f4f547d19d92cae5e409eff005da06fbeed721ba71d19c0f112a"} Dec 05 08:37:41 crc kubenswrapper[4997]: I1205 08:37:41.731997 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" Dec 05 08:37:41 crc kubenswrapper[4997]: I1205 08:37:41.752403 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" podStartSLOduration=2.752383615 podStartE2EDuration="2.752383615s" podCreationTimestamp="2025-12-05 08:37:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:37:41.7470267 +0000 UTC m=+6162.275933961" watchObservedRunningTime="2025-12-05 08:37:41.752383615 +0000 UTC m=+6162.281290876" Dec 05 08:37:42 crc kubenswrapper[4997]: I1205 08:37:42.088220 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="c67f66b9-d729-463a-8a23-b24870c6ec50" containerName="rabbitmq" containerID="cri-o://4c58f4722a4ab72c2bc5edb4011ca1e950a13f92f0c2113fe4aeeb755e735e1d" gracePeriod=604799 Dec 05 08:37:43 crc kubenswrapper[4997]: I1205 08:37:43.649025 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="235b6759-15f6-41b7-a3ed-151f56da12f6" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.251:5672: connect: connection refused" Dec 05 08:37:43 crc kubenswrapper[4997]: I1205 08:37:43.944457 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="c67f66b9-d729-463a-8a23-b24870c6ec50" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.252:5672: connect: connection refused" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.162676 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cspmz"] Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.165123 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cspmz" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.174329 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cspmz"] Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.188466 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e23b6d26-af92-4d88-926c-a785fbfd7a03-utilities\") pod \"certified-operators-cspmz\" (UID: \"e23b6d26-af92-4d88-926c-a785fbfd7a03\") " pod="openshift-marketplace/certified-operators-cspmz" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.188556 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mwq5\" (UniqueName: \"kubernetes.io/projected/e23b6d26-af92-4d88-926c-a785fbfd7a03-kube-api-access-9mwq5\") pod \"certified-operators-cspmz\" (UID: \"e23b6d26-af92-4d88-926c-a785fbfd7a03\") " pod="openshift-marketplace/certified-operators-cspmz" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.188592 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e23b6d26-af92-4d88-926c-a785fbfd7a03-catalog-content\") pod \"certified-operators-cspmz\" (UID: \"e23b6d26-af92-4d88-926c-a785fbfd7a03\") " pod="openshift-marketplace/certified-operators-cspmz" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.290047 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e23b6d26-af92-4d88-926c-a785fbfd7a03-utilities\") pod \"certified-operators-cspmz\" (UID: \"e23b6d26-af92-4d88-926c-a785fbfd7a03\") " pod="openshift-marketplace/certified-operators-cspmz" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.290445 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mwq5\" (UniqueName: \"kubernetes.io/projected/e23b6d26-af92-4d88-926c-a785fbfd7a03-kube-api-access-9mwq5\") pod \"certified-operators-cspmz\" (UID: \"e23b6d26-af92-4d88-926c-a785fbfd7a03\") " pod="openshift-marketplace/certified-operators-cspmz" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.290480 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e23b6d26-af92-4d88-926c-a785fbfd7a03-catalog-content\") pod \"certified-operators-cspmz\" (UID: \"e23b6d26-af92-4d88-926c-a785fbfd7a03\") " pod="openshift-marketplace/certified-operators-cspmz" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.290722 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e23b6d26-af92-4d88-926c-a785fbfd7a03-utilities\") pod \"certified-operators-cspmz\" (UID: \"e23b6d26-af92-4d88-926c-a785fbfd7a03\") " pod="openshift-marketplace/certified-operators-cspmz" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.290829 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e23b6d26-af92-4d88-926c-a785fbfd7a03-catalog-content\") pod \"certified-operators-cspmz\" (UID: \"e23b6d26-af92-4d88-926c-a785fbfd7a03\") " pod="openshift-marketplace/certified-operators-cspmz" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.317882 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mwq5\" (UniqueName: \"kubernetes.io/projected/e23b6d26-af92-4d88-926c-a785fbfd7a03-kube-api-access-9mwq5\") pod \"certified-operators-cspmz\" (UID: \"e23b6d26-af92-4d88-926c-a785fbfd7a03\") " pod="openshift-marketplace/certified-operators-cspmz" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.378344 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.392027 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-erlang-cookie\") pod \"235b6759-15f6-41b7-a3ed-151f56da12f6\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.392101 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/235b6759-15f6-41b7-a3ed-151f56da12f6-server-conf\") pod \"235b6759-15f6-41b7-a3ed-151f56da12f6\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.392137 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/235b6759-15f6-41b7-a3ed-151f56da12f6-erlang-cookie-secret\") pod \"235b6759-15f6-41b7-a3ed-151f56da12f6\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.392158 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jrp2\" (UniqueName: \"kubernetes.io/projected/235b6759-15f6-41b7-a3ed-151f56da12f6-kube-api-access-2jrp2\") pod \"235b6759-15f6-41b7-a3ed-151f56da12f6\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.392258 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\") pod \"235b6759-15f6-41b7-a3ed-151f56da12f6\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.392318 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/235b6759-15f6-41b7-a3ed-151f56da12f6-pod-info\") pod \"235b6759-15f6-41b7-a3ed-151f56da12f6\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.392349 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/235b6759-15f6-41b7-a3ed-151f56da12f6-plugins-conf\") pod \"235b6759-15f6-41b7-a3ed-151f56da12f6\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.392390 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-plugins\") pod \"235b6759-15f6-41b7-a3ed-151f56da12f6\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.392442 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-confd\") pod \"235b6759-15f6-41b7-a3ed-151f56da12f6\" (UID: \"235b6759-15f6-41b7-a3ed-151f56da12f6\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.393169 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "235b6759-15f6-41b7-a3ed-151f56da12f6" (UID: "235b6759-15f6-41b7-a3ed-151f56da12f6"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.393303 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/235b6759-15f6-41b7-a3ed-151f56da12f6-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "235b6759-15f6-41b7-a3ed-151f56da12f6" (UID: "235b6759-15f6-41b7-a3ed-151f56da12f6"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.393576 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "235b6759-15f6-41b7-a3ed-151f56da12f6" (UID: "235b6759-15f6-41b7-a3ed-151f56da12f6"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.394023 4997 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.394049 4997 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/235b6759-15f6-41b7-a3ed-151f56da12f6-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.394059 4997 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.396607 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/235b6759-15f6-41b7-a3ed-151f56da12f6-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "235b6759-15f6-41b7-a3ed-151f56da12f6" (UID: "235b6759-15f6-41b7-a3ed-151f56da12f6"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.396714 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/235b6759-15f6-41b7-a3ed-151f56da12f6-kube-api-access-2jrp2" (OuterVolumeSpecName: "kube-api-access-2jrp2") pod "235b6759-15f6-41b7-a3ed-151f56da12f6" (UID: "235b6759-15f6-41b7-a3ed-151f56da12f6"). InnerVolumeSpecName "kube-api-access-2jrp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.396930 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/235b6759-15f6-41b7-a3ed-151f56da12f6-pod-info" (OuterVolumeSpecName: "pod-info") pod "235b6759-15f6-41b7-a3ed-151f56da12f6" (UID: "235b6759-15f6-41b7-a3ed-151f56da12f6"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.421300 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54bb559b-40e8-4284-9bda-bafced57b9d3" (OuterVolumeSpecName: "persistence") pod "235b6759-15f6-41b7-a3ed-151f56da12f6" (UID: "235b6759-15f6-41b7-a3ed-151f56da12f6"). InnerVolumeSpecName "pvc-54bb559b-40e8-4284-9bda-bafced57b9d3". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.453185 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/235b6759-15f6-41b7-a3ed-151f56da12f6-server-conf" (OuterVolumeSpecName: "server-conf") pod "235b6759-15f6-41b7-a3ed-151f56da12f6" (UID: "235b6759-15f6-41b7-a3ed-151f56da12f6"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.501989 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cspmz" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.502484 4997 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/235b6759-15f6-41b7-a3ed-151f56da12f6-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.502517 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jrp2\" (UniqueName: \"kubernetes.io/projected/235b6759-15f6-41b7-a3ed-151f56da12f6-kube-api-access-2jrp2\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.502556 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\") on node \"crc\" " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.502566 4997 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/235b6759-15f6-41b7-a3ed-151f56da12f6-pod-info\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.502575 4997 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/235b6759-15f6-41b7-a3ed-151f56da12f6-server-conf\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.516378 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "235b6759-15f6-41b7-a3ed-151f56da12f6" (UID: "235b6759-15f6-41b7-a3ed-151f56da12f6"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.552973 4997 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.553221 4997 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-54bb559b-40e8-4284-9bda-bafced57b9d3" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54bb559b-40e8-4284-9bda-bafced57b9d3") on node "crc" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.606168 4997 reconciler_common.go:293] "Volume detached for volume \"pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.606207 4997 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/235b6759-15f6-41b7-a3ed-151f56da12f6-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.746662 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.810676 4997 generic.go:334] "Generic (PLEG): container finished" podID="235b6759-15f6-41b7-a3ed-151f56da12f6" containerID="a5c814d2cace6349b94bdb8779fc1b179ed24c1bb28d6c852ff2e0ba495d846d" exitCode=0 Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.810966 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"235b6759-15f6-41b7-a3ed-151f56da12f6","Type":"ContainerDied","Data":"a5c814d2cace6349b94bdb8779fc1b179ed24c1bb28d6c852ff2e0ba495d846d"} Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.810993 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"235b6759-15f6-41b7-a3ed-151f56da12f6","Type":"ContainerDied","Data":"e8443b1838c57ae9f8129b83eb55ba1b6e6b9196a81de29d5740c2187e1726a0"} Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.811011 4997 scope.go:117] "RemoveContainer" containerID="a5c814d2cace6349b94bdb8779fc1b179ed24c1bb28d6c852ff2e0ba495d846d" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.811128 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.822816 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c67f66b9-d729-463a-8a23-b24870c6ec50-server-conf\") pod \"c67f66b9-d729-463a-8a23-b24870c6ec50\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.822901 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-erlang-cookie\") pod \"c67f66b9-d729-463a-8a23-b24870c6ec50\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.822930 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-confd\") pod \"c67f66b9-d729-463a-8a23-b24870c6ec50\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.822970 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c67f66b9-d729-463a-8a23-b24870c6ec50-pod-info\") pod \"c67f66b9-d729-463a-8a23-b24870c6ec50\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.822995 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c67f66b9-d729-463a-8a23-b24870c6ec50-erlang-cookie-secret\") pod \"c67f66b9-d729-463a-8a23-b24870c6ec50\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.823145 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5a76facf-8690-4b19-80a2-011b963e7f01\") pod \"c67f66b9-d729-463a-8a23-b24870c6ec50\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.823178 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wt9ps\" (UniqueName: \"kubernetes.io/projected/c67f66b9-d729-463a-8a23-b24870c6ec50-kube-api-access-wt9ps\") pod \"c67f66b9-d729-463a-8a23-b24870c6ec50\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.823211 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c67f66b9-d729-463a-8a23-b24870c6ec50-plugins-conf\") pod \"c67f66b9-d729-463a-8a23-b24870c6ec50\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.823253 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-plugins\") pod \"c67f66b9-d729-463a-8a23-b24870c6ec50\" (UID: \"c67f66b9-d729-463a-8a23-b24870c6ec50\") " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.823827 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c67f66b9-d729-463a-8a23-b24870c6ec50" (UID: "c67f66b9-d729-463a-8a23-b24870c6ec50"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.825220 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c67f66b9-d729-463a-8a23-b24870c6ec50-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c67f66b9-d729-463a-8a23-b24870c6ec50" (UID: "c67f66b9-d729-463a-8a23-b24870c6ec50"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.826407 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c67f66b9-d729-463a-8a23-b24870c6ec50" (UID: "c67f66b9-d729-463a-8a23-b24870c6ec50"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.827331 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c67f66b9-d729-463a-8a23-b24870c6ec50-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c67f66b9-d729-463a-8a23-b24870c6ec50" (UID: "c67f66b9-d729-463a-8a23-b24870c6ec50"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.831362 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c67f66b9-d729-463a-8a23-b24870c6ec50-kube-api-access-wt9ps" (OuterVolumeSpecName: "kube-api-access-wt9ps") pod "c67f66b9-d729-463a-8a23-b24870c6ec50" (UID: "c67f66b9-d729-463a-8a23-b24870c6ec50"). InnerVolumeSpecName "kube-api-access-wt9ps". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.833696 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c67f66b9-d729-463a-8a23-b24870c6ec50-pod-info" (OuterVolumeSpecName: "pod-info") pod "c67f66b9-d729-463a-8a23-b24870c6ec50" (UID: "c67f66b9-d729-463a-8a23-b24870c6ec50"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.840983 4997 generic.go:334] "Generic (PLEG): container finished" podID="c67f66b9-d729-463a-8a23-b24870c6ec50" containerID="4c58f4722a4ab72c2bc5edb4011ca1e950a13f92f0c2113fe4aeeb755e735e1d" exitCode=0 Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.841030 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c67f66b9-d729-463a-8a23-b24870c6ec50","Type":"ContainerDied","Data":"4c58f4722a4ab72c2bc5edb4011ca1e950a13f92f0c2113fe4aeeb755e735e1d"} Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.841064 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c67f66b9-d729-463a-8a23-b24870c6ec50","Type":"ContainerDied","Data":"0e98739ad027b512b84e7ff54a61b09797442c21641432f87499b73dbd0425db"} Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.841120 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.844460 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c67f66b9-d729-463a-8a23-b24870c6ec50-server-conf" (OuterVolumeSpecName: "server-conf") pod "c67f66b9-d729-463a-8a23-b24870c6ec50" (UID: "c67f66b9-d729-463a-8a23-b24870c6ec50"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.887947 4997 scope.go:117] "RemoveContainer" containerID="03267fcc99286e2f62f3ae2c1d5fd66aef47dc6f7e6d0e6a11692dec3ebf3e72" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.898821 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5a76facf-8690-4b19-80a2-011b963e7f01" (OuterVolumeSpecName: "persistence") pod "c67f66b9-d729-463a-8a23-b24870c6ec50" (UID: "c67f66b9-d729-463a-8a23-b24870c6ec50"). InnerVolumeSpecName "pvc-5a76facf-8690-4b19-80a2-011b963e7f01". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.917322 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.920149 4997 scope.go:117] "RemoveContainer" containerID="a5c814d2cace6349b94bdb8779fc1b179ed24c1bb28d6c852ff2e0ba495d846d" Dec 05 08:37:48 crc kubenswrapper[4997]: E1205 08:37:48.932116 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5c814d2cace6349b94bdb8779fc1b179ed24c1bb28d6c852ff2e0ba495d846d\": container with ID starting with a5c814d2cace6349b94bdb8779fc1b179ed24c1bb28d6c852ff2e0ba495d846d not found: ID does not exist" containerID="a5c814d2cace6349b94bdb8779fc1b179ed24c1bb28d6c852ff2e0ba495d846d" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.932170 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5c814d2cace6349b94bdb8779fc1b179ed24c1bb28d6c852ff2e0ba495d846d"} err="failed to get container status \"a5c814d2cace6349b94bdb8779fc1b179ed24c1bb28d6c852ff2e0ba495d846d\": rpc error: code = NotFound desc = could not find container \"a5c814d2cace6349b94bdb8779fc1b179ed24c1bb28d6c852ff2e0ba495d846d\": container with ID starting with a5c814d2cace6349b94bdb8779fc1b179ed24c1bb28d6c852ff2e0ba495d846d not found: ID does not exist" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.932200 4997 scope.go:117] "RemoveContainer" containerID="03267fcc99286e2f62f3ae2c1d5fd66aef47dc6f7e6d0e6a11692dec3ebf3e72" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.936440 4997 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.936480 4997 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c67f66b9-d729-463a-8a23-b24870c6ec50-server-conf\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.936525 4997 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.936541 4997 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c67f66b9-d729-463a-8a23-b24870c6ec50-pod-info\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.936553 4997 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c67f66b9-d729-463a-8a23-b24870c6ec50-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.936590 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-5a76facf-8690-4b19-80a2-011b963e7f01\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5a76facf-8690-4b19-80a2-011b963e7f01\") on node \"crc\" " Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.936606 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wt9ps\" (UniqueName: \"kubernetes.io/projected/c67f66b9-d729-463a-8a23-b24870c6ec50-kube-api-access-wt9ps\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.936667 4997 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c67f66b9-d729-463a-8a23-b24870c6ec50-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4997]: E1205 08:37:48.945795 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03267fcc99286e2f62f3ae2c1d5fd66aef47dc6f7e6d0e6a11692dec3ebf3e72\": container with ID starting with 03267fcc99286e2f62f3ae2c1d5fd66aef47dc6f7e6d0e6a11692dec3ebf3e72 not found: ID does not exist" containerID="03267fcc99286e2f62f3ae2c1d5fd66aef47dc6f7e6d0e6a11692dec3ebf3e72" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.945848 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03267fcc99286e2f62f3ae2c1d5fd66aef47dc6f7e6d0e6a11692dec3ebf3e72"} err="failed to get container status \"03267fcc99286e2f62f3ae2c1d5fd66aef47dc6f7e6d0e6a11692dec3ebf3e72\": rpc error: code = NotFound desc = could not find container \"03267fcc99286e2f62f3ae2c1d5fd66aef47dc6f7e6d0e6a11692dec3ebf3e72\": container with ID starting with 03267fcc99286e2f62f3ae2c1d5fd66aef47dc6f7e6d0e6a11692dec3ebf3e72 not found: ID does not exist" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.945872 4997 scope.go:117] "RemoveContainer" containerID="4c58f4722a4ab72c2bc5edb4011ca1e950a13f92f0c2113fe4aeeb755e735e1d" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.949305 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.958908 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:37:48 crc kubenswrapper[4997]: E1205 08:37:48.961416 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c67f66b9-d729-463a-8a23-b24870c6ec50" containerName="setup-container" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.961439 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c67f66b9-d729-463a-8a23-b24870c6ec50" containerName="setup-container" Dec 05 08:37:48 crc kubenswrapper[4997]: E1205 08:37:48.961473 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c67f66b9-d729-463a-8a23-b24870c6ec50" containerName="rabbitmq" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.961479 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c67f66b9-d729-463a-8a23-b24870c6ec50" containerName="rabbitmq" Dec 05 08:37:48 crc kubenswrapper[4997]: E1205 08:37:48.961491 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="235b6759-15f6-41b7-a3ed-151f56da12f6" containerName="rabbitmq" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.961497 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="235b6759-15f6-41b7-a3ed-151f56da12f6" containerName="rabbitmq" Dec 05 08:37:48 crc kubenswrapper[4997]: E1205 08:37:48.961507 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="235b6759-15f6-41b7-a3ed-151f56da12f6" containerName="setup-container" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.961512 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="235b6759-15f6-41b7-a3ed-151f56da12f6" containerName="setup-container" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.961716 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="235b6759-15f6-41b7-a3ed-151f56da12f6" containerName="rabbitmq" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.961751 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="c67f66b9-d729-463a-8a23-b24870c6ec50" containerName="rabbitmq" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.963309 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.969905 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.970213 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.970230 4997 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.970413 4997 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-5a76facf-8690-4b19-80a2-011b963e7f01" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5a76facf-8690-4b19-80a2-011b963e7f01") on node "crc" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.970421 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.970470 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-56hjt" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.970432 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 05 08:37:48 crc kubenswrapper[4997]: I1205 08:37:48.970711 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.021183 4997 scope.go:117] "RemoveContainer" containerID="1989c9041938c4d4c76b60e44b4e3378aa6fded73aa42770e050df62c35aed10" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.038204 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a3c52965-15ba-45b0-b0fc-b447ae56547e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.038252 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a3c52965-15ba-45b0-b0fc-b447ae56547e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.038276 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a3c52965-15ba-45b0-b0fc-b447ae56547e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.038307 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a3c52965-15ba-45b0-b0fc-b447ae56547e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.038337 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a3c52965-15ba-45b0-b0fc-b447ae56547e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.038358 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6htw\" (UniqueName: \"kubernetes.io/projected/a3c52965-15ba-45b0-b0fc-b447ae56547e-kube-api-access-z6htw\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.038401 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a3c52965-15ba-45b0-b0fc-b447ae56547e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.038426 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a3c52965-15ba-45b0-b0fc-b447ae56547e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.038450 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.038496 4997 reconciler_common.go:293] "Volume detached for volume \"pvc-5a76facf-8690-4b19-80a2-011b963e7f01\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5a76facf-8690-4b19-80a2-011b963e7f01\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.045430 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c67f66b9-d729-463a-8a23-b24870c6ec50" (UID: "c67f66b9-d729-463a-8a23-b24870c6ec50"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.049935 4997 scope.go:117] "RemoveContainer" containerID="4c58f4722a4ab72c2bc5edb4011ca1e950a13f92f0c2113fe4aeeb755e735e1d" Dec 05 08:37:49 crc kubenswrapper[4997]: E1205 08:37:49.050869 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c58f4722a4ab72c2bc5edb4011ca1e950a13f92f0c2113fe4aeeb755e735e1d\": container with ID starting with 4c58f4722a4ab72c2bc5edb4011ca1e950a13f92f0c2113fe4aeeb755e735e1d not found: ID does not exist" containerID="4c58f4722a4ab72c2bc5edb4011ca1e950a13f92f0c2113fe4aeeb755e735e1d" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.050904 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c58f4722a4ab72c2bc5edb4011ca1e950a13f92f0c2113fe4aeeb755e735e1d"} err="failed to get container status \"4c58f4722a4ab72c2bc5edb4011ca1e950a13f92f0c2113fe4aeeb755e735e1d\": rpc error: code = NotFound desc = could not find container \"4c58f4722a4ab72c2bc5edb4011ca1e950a13f92f0c2113fe4aeeb755e735e1d\": container with ID starting with 4c58f4722a4ab72c2bc5edb4011ca1e950a13f92f0c2113fe4aeeb755e735e1d not found: ID does not exist" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.050926 4997 scope.go:117] "RemoveContainer" containerID="1989c9041938c4d4c76b60e44b4e3378aa6fded73aa42770e050df62c35aed10" Dec 05 08:37:49 crc kubenswrapper[4997]: E1205 08:37:49.051285 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1989c9041938c4d4c76b60e44b4e3378aa6fded73aa42770e050df62c35aed10\": container with ID starting with 1989c9041938c4d4c76b60e44b4e3378aa6fded73aa42770e050df62c35aed10 not found: ID does not exist" containerID="1989c9041938c4d4c76b60e44b4e3378aa6fded73aa42770e050df62c35aed10" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.051311 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1989c9041938c4d4c76b60e44b4e3378aa6fded73aa42770e050df62c35aed10"} err="failed to get container status \"1989c9041938c4d4c76b60e44b4e3378aa6fded73aa42770e050df62c35aed10\": rpc error: code = NotFound desc = could not find container \"1989c9041938c4d4c76b60e44b4e3378aa6fded73aa42770e050df62c35aed10\": container with ID starting with 1989c9041938c4d4c76b60e44b4e3378aa6fded73aa42770e050df62c35aed10 not found: ID does not exist" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.102312 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cspmz"] Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.139799 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a3c52965-15ba-45b0-b0fc-b447ae56547e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.139908 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a3c52965-15ba-45b0-b0fc-b447ae56547e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.139958 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.140000 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a3c52965-15ba-45b0-b0fc-b447ae56547e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.140033 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a3c52965-15ba-45b0-b0fc-b447ae56547e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.140064 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a3c52965-15ba-45b0-b0fc-b447ae56547e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.140097 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a3c52965-15ba-45b0-b0fc-b447ae56547e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.140139 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a3c52965-15ba-45b0-b0fc-b447ae56547e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.140170 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6htw\" (UniqueName: \"kubernetes.io/projected/a3c52965-15ba-45b0-b0fc-b447ae56547e-kube-api-access-z6htw\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.140298 4997 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c67f66b9-d729-463a-8a23-b24870c6ec50-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.141047 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a3c52965-15ba-45b0-b0fc-b447ae56547e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.141405 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a3c52965-15ba-45b0-b0fc-b447ae56547e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.142540 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a3c52965-15ba-45b0-b0fc-b447ae56547e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.142862 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a3c52965-15ba-45b0-b0fc-b447ae56547e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.145075 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a3c52965-15ba-45b0-b0fc-b447ae56547e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.146465 4997 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.146527 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/cebfedb8bf3ab96dff23009c7dd389f6e95ad5aceed82f3636d5fe256275b59d/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.152520 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a3c52965-15ba-45b0-b0fc-b447ae56547e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.156080 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a3c52965-15ba-45b0-b0fc-b447ae56547e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.160198 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6htw\" (UniqueName: \"kubernetes.io/projected/a3c52965-15ba-45b0-b0fc-b447ae56547e-kube-api-access-z6htw\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.189019 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.199458 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54bb559b-40e8-4284-9bda-bafced57b9d3\") pod \"rabbitmq-server-0\" (UID: \"a3c52965-15ba-45b0-b0fc-b447ae56547e\") " pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.206913 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.219243 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.222730 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.226584 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.226797 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.227071 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.227115 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-khzbp" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.227200 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.229057 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.346935 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/522f749b-2397-472f-bde2-b0f8fb8419a8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.346983 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/522f749b-2397-472f-bde2-b0f8fb8419a8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.347104 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/522f749b-2397-472f-bde2-b0f8fb8419a8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.347180 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/522f749b-2397-472f-bde2-b0f8fb8419a8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.347263 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/522f749b-2397-472f-bde2-b0f8fb8419a8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.347321 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgb7d\" (UniqueName: \"kubernetes.io/projected/522f749b-2397-472f-bde2-b0f8fb8419a8-kube-api-access-qgb7d\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.347350 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5a76facf-8690-4b19-80a2-011b963e7f01\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5a76facf-8690-4b19-80a2-011b963e7f01\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.347382 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/522f749b-2397-472f-bde2-b0f8fb8419a8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.347402 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/522f749b-2397-472f-bde2-b0f8fb8419a8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.389847 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.449065 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/522f749b-2397-472f-bde2-b0f8fb8419a8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.449136 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/522f749b-2397-472f-bde2-b0f8fb8419a8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.449183 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/522f749b-2397-472f-bde2-b0f8fb8419a8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.449204 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/522f749b-2397-472f-bde2-b0f8fb8419a8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.449230 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/522f749b-2397-472f-bde2-b0f8fb8419a8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.449253 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/522f749b-2397-472f-bde2-b0f8fb8419a8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.449283 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/522f749b-2397-472f-bde2-b0f8fb8419a8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.449319 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgb7d\" (UniqueName: \"kubernetes.io/projected/522f749b-2397-472f-bde2-b0f8fb8419a8-kube-api-access-qgb7d\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.449350 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5a76facf-8690-4b19-80a2-011b963e7f01\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5a76facf-8690-4b19-80a2-011b963e7f01\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.450103 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/522f749b-2397-472f-bde2-b0f8fb8419a8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.450317 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/522f749b-2397-472f-bde2-b0f8fb8419a8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.450507 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/522f749b-2397-472f-bde2-b0f8fb8419a8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.452277 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/522f749b-2397-472f-bde2-b0f8fb8419a8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.452313 4997 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.452336 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5a76facf-8690-4b19-80a2-011b963e7f01\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5a76facf-8690-4b19-80a2-011b963e7f01\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/13f336a81097285c82b8991b303326fa943937a5f7cb23c86fbd23e7b1c598a3/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.455450 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/522f749b-2397-472f-bde2-b0f8fb8419a8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.455813 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/522f749b-2397-472f-bde2-b0f8fb8419a8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.456021 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/522f749b-2397-472f-bde2-b0f8fb8419a8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.466948 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgb7d\" (UniqueName: \"kubernetes.io/projected/522f749b-2397-472f-bde2-b0f8fb8419a8-kube-api-access-qgb7d\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.479666 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5a76facf-8690-4b19-80a2-011b963e7f01\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5a76facf-8690-4b19-80a2-011b963e7f01\") pod \"rabbitmq-cell1-server-0\" (UID: \"522f749b-2397-472f-bde2-b0f8fb8419a8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.498147 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.552122 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.554066 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84496478f-jmzvf"] Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.554285 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84496478f-jmzvf" podUID="b9355424-f0c2-4a5c-9fce-5a48bc52887a" containerName="dnsmasq-dns" containerID="cri-o://5091aa1c750377aa763a37282f33b5e3b01fdc56def3ca0ac73d6abf3e4004a7" gracePeriod=10 Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.755847 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:37:49 crc kubenswrapper[4997]: E1205 08:37:49.756113 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.767033 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="235b6759-15f6-41b7-a3ed-151f56da12f6" path="/var/lib/kubelet/pods/235b6759-15f6-41b7-a3ed-151f56da12f6/volumes" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.767773 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c67f66b9-d729-463a-8a23-b24870c6ec50" path="/var/lib/kubelet/pods/c67f66b9-d729-463a-8a23-b24870c6ec50/volumes" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.864848 4997 generic.go:334] "Generic (PLEG): container finished" podID="e23b6d26-af92-4d88-926c-a785fbfd7a03" containerID="392552facdbccfd22c1ba77f57ec4f2254458733254eb6f51e9da9b09dbe8265" exitCode=0 Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.865161 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cspmz" event={"ID":"e23b6d26-af92-4d88-926c-a785fbfd7a03","Type":"ContainerDied","Data":"392552facdbccfd22c1ba77f57ec4f2254458733254eb6f51e9da9b09dbe8265"} Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.865188 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cspmz" event={"ID":"e23b6d26-af92-4d88-926c-a785fbfd7a03","Type":"ContainerStarted","Data":"312e17638f1b3f486482a4feeee0bd1c5b1a6ac9d317fe32b0f52e254d66ae0d"} Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.876272 4997 generic.go:334] "Generic (PLEG): container finished" podID="b9355424-f0c2-4a5c-9fce-5a48bc52887a" containerID="5091aa1c750377aa763a37282f33b5e3b01fdc56def3ca0ac73d6abf3e4004a7" exitCode=0 Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.876386 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84496478f-jmzvf" event={"ID":"b9355424-f0c2-4a5c-9fce-5a48bc52887a","Type":"ContainerDied","Data":"5091aa1c750377aa763a37282f33b5e3b01fdc56def3ca0ac73d6abf3e4004a7"} Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.880965 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:37:49 crc kubenswrapper[4997]: W1205 08:37:49.883436 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3c52965_15ba_45b0_b0fc_b447ae56547e.slice/crio-0165672485c4318235b8eaaae4c89ce8f521a55e55012fbce67c468001c01ace WatchSource:0}: Error finding container 0165672485c4318235b8eaaae4c89ce8f521a55e55012fbce67c468001c01ace: Status 404 returned error can't find the container with id 0165672485c4318235b8eaaae4c89ce8f521a55e55012fbce67c468001c01ace Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.927963 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84496478f-jmzvf" Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.991368 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9355424-f0c2-4a5c-9fce-5a48bc52887a-config\") pod \"b9355424-f0c2-4a5c-9fce-5a48bc52887a\" (UID: \"b9355424-f0c2-4a5c-9fce-5a48bc52887a\") " Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.991464 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4s88h\" (UniqueName: \"kubernetes.io/projected/b9355424-f0c2-4a5c-9fce-5a48bc52887a-kube-api-access-4s88h\") pod \"b9355424-f0c2-4a5c-9fce-5a48bc52887a\" (UID: \"b9355424-f0c2-4a5c-9fce-5a48bc52887a\") " Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.991552 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9355424-f0c2-4a5c-9fce-5a48bc52887a-dns-svc\") pod \"b9355424-f0c2-4a5c-9fce-5a48bc52887a\" (UID: \"b9355424-f0c2-4a5c-9fce-5a48bc52887a\") " Dec 05 08:37:49 crc kubenswrapper[4997]: I1205 08:37:49.995121 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9355424-f0c2-4a5c-9fce-5a48bc52887a-kube-api-access-4s88h" (OuterVolumeSpecName: "kube-api-access-4s88h") pod "b9355424-f0c2-4a5c-9fce-5a48bc52887a" (UID: "b9355424-f0c2-4a5c-9fce-5a48bc52887a"). InnerVolumeSpecName "kube-api-access-4s88h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:50 crc kubenswrapper[4997]: I1205 08:37:50.034333 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9355424-f0c2-4a5c-9fce-5a48bc52887a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b9355424-f0c2-4a5c-9fce-5a48bc52887a" (UID: "b9355424-f0c2-4a5c-9fce-5a48bc52887a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:37:50 crc kubenswrapper[4997]: I1205 08:37:50.035151 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9355424-f0c2-4a5c-9fce-5a48bc52887a-config" (OuterVolumeSpecName: "config") pod "b9355424-f0c2-4a5c-9fce-5a48bc52887a" (UID: "b9355424-f0c2-4a5c-9fce-5a48bc52887a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:37:50 crc kubenswrapper[4997]: I1205 08:37:50.093595 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9355424-f0c2-4a5c-9fce-5a48bc52887a-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:50 crc kubenswrapper[4997]: I1205 08:37:50.093642 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4s88h\" (UniqueName: \"kubernetes.io/projected/b9355424-f0c2-4a5c-9fce-5a48bc52887a-kube-api-access-4s88h\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:50 crc kubenswrapper[4997]: I1205 08:37:50.093655 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9355424-f0c2-4a5c-9fce-5a48bc52887a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:50 crc kubenswrapper[4997]: I1205 08:37:50.141989 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:37:50 crc kubenswrapper[4997]: W1205 08:37:50.147318 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod522f749b_2397_472f_bde2_b0f8fb8419a8.slice/crio-fc349d682b8fc7963fcb13ef42591f692004ea8caa9b24351bc6a478e31e7411 WatchSource:0}: Error finding container fc349d682b8fc7963fcb13ef42591f692004ea8caa9b24351bc6a478e31e7411: Status 404 returned error can't find the container with id fc349d682b8fc7963fcb13ef42591f692004ea8caa9b24351bc6a478e31e7411 Dec 05 08:37:50 crc kubenswrapper[4997]: I1205 08:37:50.894450 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a3c52965-15ba-45b0-b0fc-b447ae56547e","Type":"ContainerStarted","Data":"0165672485c4318235b8eaaae4c89ce8f521a55e55012fbce67c468001c01ace"} Dec 05 08:37:50 crc kubenswrapper[4997]: I1205 08:37:50.895550 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"522f749b-2397-472f-bde2-b0f8fb8419a8","Type":"ContainerStarted","Data":"fc349d682b8fc7963fcb13ef42591f692004ea8caa9b24351bc6a478e31e7411"} Dec 05 08:37:50 crc kubenswrapper[4997]: I1205 08:37:50.897586 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84496478f-jmzvf" event={"ID":"b9355424-f0c2-4a5c-9fce-5a48bc52887a","Type":"ContainerDied","Data":"ff9407c03bc127c1e0066f3b6dde0a9f2b7280fd785fa3e818320658da8673f7"} Dec 05 08:37:50 crc kubenswrapper[4997]: I1205 08:37:50.897701 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84496478f-jmzvf" Dec 05 08:37:50 crc kubenswrapper[4997]: I1205 08:37:50.897795 4997 scope.go:117] "RemoveContainer" containerID="5091aa1c750377aa763a37282f33b5e3b01fdc56def3ca0ac73d6abf3e4004a7" Dec 05 08:37:50 crc kubenswrapper[4997]: I1205 08:37:50.917274 4997 scope.go:117] "RemoveContainer" containerID="38d158200b6b86401b2f099d58e007a6b1b337495674dd1adc42bc344ea8d4c4" Dec 05 08:37:50 crc kubenswrapper[4997]: I1205 08:37:50.936128 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84496478f-jmzvf"] Dec 05 08:37:50 crc kubenswrapper[4997]: I1205 08:37:50.941349 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84496478f-jmzvf"] Dec 05 08:37:51 crc kubenswrapper[4997]: I1205 08:37:51.759286 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9355424-f0c2-4a5c-9fce-5a48bc52887a" path="/var/lib/kubelet/pods/b9355424-f0c2-4a5c-9fce-5a48bc52887a/volumes" Dec 05 08:37:51 crc kubenswrapper[4997]: I1205 08:37:51.906300 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"522f749b-2397-472f-bde2-b0f8fb8419a8","Type":"ContainerStarted","Data":"3420a96bfb48d0e44b6a1cd26331da68c7de23c84021830d0c174d1081fb7142"} Dec 05 08:37:51 crc kubenswrapper[4997]: I1205 08:37:51.912749 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a3c52965-15ba-45b0-b0fc-b447ae56547e","Type":"ContainerStarted","Data":"19f5c41784aa4ae94435fc7a397b4868336a1a06687f38cd387c6b9bc5d9170b"} Dec 05 08:37:53 crc kubenswrapper[4997]: I1205 08:37:53.941103 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cspmz" event={"ID":"e23b6d26-af92-4d88-926c-a785fbfd7a03","Type":"ContainerStarted","Data":"b330bc6429d65cb0920fc02bafe218ac19f4442c0410fddd94205561a84ea580"} Dec 05 08:37:54 crc kubenswrapper[4997]: I1205 08:37:54.951673 4997 generic.go:334] "Generic (PLEG): container finished" podID="e23b6d26-af92-4d88-926c-a785fbfd7a03" containerID="b330bc6429d65cb0920fc02bafe218ac19f4442c0410fddd94205561a84ea580" exitCode=0 Dec 05 08:37:54 crc kubenswrapper[4997]: I1205 08:37:54.951720 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cspmz" event={"ID":"e23b6d26-af92-4d88-926c-a785fbfd7a03","Type":"ContainerDied","Data":"b330bc6429d65cb0920fc02bafe218ac19f4442c0410fddd94205561a84ea580"} Dec 05 08:37:55 crc kubenswrapper[4997]: I1205 08:37:55.960676 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cspmz" event={"ID":"e23b6d26-af92-4d88-926c-a785fbfd7a03","Type":"ContainerStarted","Data":"4959145b79f437481b01e417d847a2f7c4cda8593748339967898a7b60956e21"} Dec 05 08:37:55 crc kubenswrapper[4997]: I1205 08:37:55.979085 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cspmz" podStartSLOduration=2.468725311 podStartE2EDuration="7.979066678s" podCreationTimestamp="2025-12-05 08:37:48 +0000 UTC" firstStartedPulling="2025-12-05 08:37:49.867404379 +0000 UTC m=+6170.396311640" lastFinishedPulling="2025-12-05 08:37:55.377745746 +0000 UTC m=+6175.906653007" observedRunningTime="2025-12-05 08:37:55.974718961 +0000 UTC m=+6176.503626252" watchObservedRunningTime="2025-12-05 08:37:55.979066678 +0000 UTC m=+6176.507973939" Dec 05 08:37:58 crc kubenswrapper[4997]: I1205 08:37:58.502456 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cspmz" Dec 05 08:37:58 crc kubenswrapper[4997]: I1205 08:37:58.502783 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cspmz" Dec 05 08:37:58 crc kubenswrapper[4997]: I1205 08:37:58.543333 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cspmz" Dec 05 08:38:03 crc kubenswrapper[4997]: I1205 08:38:03.749518 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:38:03 crc kubenswrapper[4997]: E1205 08:38:03.750414 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:38:08 crc kubenswrapper[4997]: I1205 08:38:08.546837 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cspmz" Dec 05 08:38:08 crc kubenswrapper[4997]: I1205 08:38:08.627855 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cspmz"] Dec 05 08:38:08 crc kubenswrapper[4997]: I1205 08:38:08.675889 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qxz27"] Dec 05 08:38:08 crc kubenswrapper[4997]: I1205 08:38:08.676443 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qxz27" podUID="016b7cd0-efde-4d2c-940c-181108006a33" containerName="registry-server" containerID="cri-o://487fd2b294c68db72810d2d7c43a20ef768624e86a9529c833861fdd3b696b6c" gracePeriod=2 Dec 05 08:38:09 crc kubenswrapper[4997]: I1205 08:38:09.606567 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qxz27" Dec 05 08:38:09 crc kubenswrapper[4997]: I1205 08:38:09.695403 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/016b7cd0-efde-4d2c-940c-181108006a33-catalog-content\") pod \"016b7cd0-efde-4d2c-940c-181108006a33\" (UID: \"016b7cd0-efde-4d2c-940c-181108006a33\") " Dec 05 08:38:09 crc kubenswrapper[4997]: I1205 08:38:09.696465 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85m26\" (UniqueName: \"kubernetes.io/projected/016b7cd0-efde-4d2c-940c-181108006a33-kube-api-access-85m26\") pod \"016b7cd0-efde-4d2c-940c-181108006a33\" (UID: \"016b7cd0-efde-4d2c-940c-181108006a33\") " Dec 05 08:38:09 crc kubenswrapper[4997]: I1205 08:38:09.696691 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/016b7cd0-efde-4d2c-940c-181108006a33-utilities\") pod \"016b7cd0-efde-4d2c-940c-181108006a33\" (UID: \"016b7cd0-efde-4d2c-940c-181108006a33\") " Dec 05 08:38:09 crc kubenswrapper[4997]: I1205 08:38:09.698709 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/016b7cd0-efde-4d2c-940c-181108006a33-utilities" (OuterVolumeSpecName: "utilities") pod "016b7cd0-efde-4d2c-940c-181108006a33" (UID: "016b7cd0-efde-4d2c-940c-181108006a33"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:38:09 crc kubenswrapper[4997]: I1205 08:38:09.704144 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/016b7cd0-efde-4d2c-940c-181108006a33-kube-api-access-85m26" (OuterVolumeSpecName: "kube-api-access-85m26") pod "016b7cd0-efde-4d2c-940c-181108006a33" (UID: "016b7cd0-efde-4d2c-940c-181108006a33"). InnerVolumeSpecName "kube-api-access-85m26". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:09 crc kubenswrapper[4997]: I1205 08:38:09.753519 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/016b7cd0-efde-4d2c-940c-181108006a33-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "016b7cd0-efde-4d2c-940c-181108006a33" (UID: "016b7cd0-efde-4d2c-940c-181108006a33"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:38:09 crc kubenswrapper[4997]: I1205 08:38:09.798895 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/016b7cd0-efde-4d2c-940c-181108006a33-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:09 crc kubenswrapper[4997]: I1205 08:38:09.798941 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85m26\" (UniqueName: \"kubernetes.io/projected/016b7cd0-efde-4d2c-940c-181108006a33-kube-api-access-85m26\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:09 crc kubenswrapper[4997]: I1205 08:38:09.798956 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/016b7cd0-efde-4d2c-940c-181108006a33-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:10 crc kubenswrapper[4997]: I1205 08:38:10.083222 4997 generic.go:334] "Generic (PLEG): container finished" podID="016b7cd0-efde-4d2c-940c-181108006a33" containerID="487fd2b294c68db72810d2d7c43a20ef768624e86a9529c833861fdd3b696b6c" exitCode=0 Dec 05 08:38:10 crc kubenswrapper[4997]: I1205 08:38:10.083285 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxz27" event={"ID":"016b7cd0-efde-4d2c-940c-181108006a33","Type":"ContainerDied","Data":"487fd2b294c68db72810d2d7c43a20ef768624e86a9529c833861fdd3b696b6c"} Dec 05 08:38:10 crc kubenswrapper[4997]: I1205 08:38:10.083344 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxz27" event={"ID":"016b7cd0-efde-4d2c-940c-181108006a33","Type":"ContainerDied","Data":"6b5f884e76f29d3b65abca91b8b5e5ffe97e044651cacf681877d8d4e514ba6c"} Dec 05 08:38:10 crc kubenswrapper[4997]: I1205 08:38:10.083370 4997 scope.go:117] "RemoveContainer" containerID="487fd2b294c68db72810d2d7c43a20ef768624e86a9529c833861fdd3b696b6c" Dec 05 08:38:10 crc kubenswrapper[4997]: I1205 08:38:10.083568 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qxz27" Dec 05 08:38:10 crc kubenswrapper[4997]: I1205 08:38:10.113796 4997 scope.go:117] "RemoveContainer" containerID="75938ac1bf709a5faa682693493ce26c1217ed8e5b6fef11bb387a7a4b04ba74" Dec 05 08:38:10 crc kubenswrapper[4997]: I1205 08:38:10.119332 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qxz27"] Dec 05 08:38:10 crc kubenswrapper[4997]: I1205 08:38:10.129383 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qxz27"] Dec 05 08:38:10 crc kubenswrapper[4997]: I1205 08:38:10.142493 4997 scope.go:117] "RemoveContainer" containerID="a9c68c2270409df9305fea61bfb523736272b3d4a5aff279aefe99e07a9aaf93" Dec 05 08:38:10 crc kubenswrapper[4997]: I1205 08:38:10.170340 4997 scope.go:117] "RemoveContainer" containerID="487fd2b294c68db72810d2d7c43a20ef768624e86a9529c833861fdd3b696b6c" Dec 05 08:38:10 crc kubenswrapper[4997]: E1205 08:38:10.170860 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"487fd2b294c68db72810d2d7c43a20ef768624e86a9529c833861fdd3b696b6c\": container with ID starting with 487fd2b294c68db72810d2d7c43a20ef768624e86a9529c833861fdd3b696b6c not found: ID does not exist" containerID="487fd2b294c68db72810d2d7c43a20ef768624e86a9529c833861fdd3b696b6c" Dec 05 08:38:10 crc kubenswrapper[4997]: I1205 08:38:10.170912 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"487fd2b294c68db72810d2d7c43a20ef768624e86a9529c833861fdd3b696b6c"} err="failed to get container status \"487fd2b294c68db72810d2d7c43a20ef768624e86a9529c833861fdd3b696b6c\": rpc error: code = NotFound desc = could not find container \"487fd2b294c68db72810d2d7c43a20ef768624e86a9529c833861fdd3b696b6c\": container with ID starting with 487fd2b294c68db72810d2d7c43a20ef768624e86a9529c833861fdd3b696b6c not found: ID does not exist" Dec 05 08:38:10 crc kubenswrapper[4997]: I1205 08:38:10.170939 4997 scope.go:117] "RemoveContainer" containerID="75938ac1bf709a5faa682693493ce26c1217ed8e5b6fef11bb387a7a4b04ba74" Dec 05 08:38:10 crc kubenswrapper[4997]: E1205 08:38:10.171397 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75938ac1bf709a5faa682693493ce26c1217ed8e5b6fef11bb387a7a4b04ba74\": container with ID starting with 75938ac1bf709a5faa682693493ce26c1217ed8e5b6fef11bb387a7a4b04ba74 not found: ID does not exist" containerID="75938ac1bf709a5faa682693493ce26c1217ed8e5b6fef11bb387a7a4b04ba74" Dec 05 08:38:10 crc kubenswrapper[4997]: I1205 08:38:10.171428 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75938ac1bf709a5faa682693493ce26c1217ed8e5b6fef11bb387a7a4b04ba74"} err="failed to get container status \"75938ac1bf709a5faa682693493ce26c1217ed8e5b6fef11bb387a7a4b04ba74\": rpc error: code = NotFound desc = could not find container \"75938ac1bf709a5faa682693493ce26c1217ed8e5b6fef11bb387a7a4b04ba74\": container with ID starting with 75938ac1bf709a5faa682693493ce26c1217ed8e5b6fef11bb387a7a4b04ba74 not found: ID does not exist" Dec 05 08:38:10 crc kubenswrapper[4997]: I1205 08:38:10.171447 4997 scope.go:117] "RemoveContainer" containerID="a9c68c2270409df9305fea61bfb523736272b3d4a5aff279aefe99e07a9aaf93" Dec 05 08:38:10 crc kubenswrapper[4997]: E1205 08:38:10.171865 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9c68c2270409df9305fea61bfb523736272b3d4a5aff279aefe99e07a9aaf93\": container with ID starting with a9c68c2270409df9305fea61bfb523736272b3d4a5aff279aefe99e07a9aaf93 not found: ID does not exist" containerID="a9c68c2270409df9305fea61bfb523736272b3d4a5aff279aefe99e07a9aaf93" Dec 05 08:38:10 crc kubenswrapper[4997]: I1205 08:38:10.171932 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9c68c2270409df9305fea61bfb523736272b3d4a5aff279aefe99e07a9aaf93"} err="failed to get container status \"a9c68c2270409df9305fea61bfb523736272b3d4a5aff279aefe99e07a9aaf93\": rpc error: code = NotFound desc = could not find container \"a9c68c2270409df9305fea61bfb523736272b3d4a5aff279aefe99e07a9aaf93\": container with ID starting with a9c68c2270409df9305fea61bfb523736272b3d4a5aff279aefe99e07a9aaf93 not found: ID does not exist" Dec 05 08:38:11 crc kubenswrapper[4997]: I1205 08:38:11.767594 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="016b7cd0-efde-4d2c-940c-181108006a33" path="/var/lib/kubelet/pods/016b7cd0-efde-4d2c-940c-181108006a33/volumes" Dec 05 08:38:17 crc kubenswrapper[4997]: I1205 08:38:17.749040 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:38:17 crc kubenswrapper[4997]: E1205 08:38:17.749861 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:38:23 crc kubenswrapper[4997]: I1205 08:38:23.202605 4997 generic.go:334] "Generic (PLEG): container finished" podID="a3c52965-15ba-45b0-b0fc-b447ae56547e" containerID="19f5c41784aa4ae94435fc7a397b4868336a1a06687f38cd387c6b9bc5d9170b" exitCode=0 Dec 05 08:38:23 crc kubenswrapper[4997]: I1205 08:38:23.202689 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a3c52965-15ba-45b0-b0fc-b447ae56547e","Type":"ContainerDied","Data":"19f5c41784aa4ae94435fc7a397b4868336a1a06687f38cd387c6b9bc5d9170b"} Dec 05 08:38:24 crc kubenswrapper[4997]: I1205 08:38:24.213596 4997 generic.go:334] "Generic (PLEG): container finished" podID="522f749b-2397-472f-bde2-b0f8fb8419a8" containerID="3420a96bfb48d0e44b6a1cd26331da68c7de23c84021830d0c174d1081fb7142" exitCode=0 Dec 05 08:38:24 crc kubenswrapper[4997]: I1205 08:38:24.213873 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"522f749b-2397-472f-bde2-b0f8fb8419a8","Type":"ContainerDied","Data":"3420a96bfb48d0e44b6a1cd26331da68c7de23c84021830d0c174d1081fb7142"} Dec 05 08:38:24 crc kubenswrapper[4997]: I1205 08:38:24.216693 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a3c52965-15ba-45b0-b0fc-b447ae56547e","Type":"ContainerStarted","Data":"1a93318c76ac6785ef6f95f4c3aec02d5cd0f70e81caff99bd4466a2b9865bb2"} Dec 05 08:38:24 crc kubenswrapper[4997]: I1205 08:38:24.216900 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 05 08:38:24 crc kubenswrapper[4997]: I1205 08:38:24.271233 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.271197422 podStartE2EDuration="36.271197422s" podCreationTimestamp="2025-12-05 08:37:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:38:24.260533483 +0000 UTC m=+6204.789440764" watchObservedRunningTime="2025-12-05 08:38:24.271197422 +0000 UTC m=+6204.800104683" Dec 05 08:38:25 crc kubenswrapper[4997]: I1205 08:38:25.226145 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"522f749b-2397-472f-bde2-b0f8fb8419a8","Type":"ContainerStarted","Data":"fcfd6f93ab58b9143052e0224fe0a6120e7d56173fae09a6d184df6d0f06a9b7"} Dec 05 08:38:25 crc kubenswrapper[4997]: I1205 08:38:25.227122 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:38:25 crc kubenswrapper[4997]: I1205 08:38:25.262098 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.262078223 podStartE2EDuration="36.262078223s" podCreationTimestamp="2025-12-05 08:37:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:38:25.256007219 +0000 UTC m=+6205.784914550" watchObservedRunningTime="2025-12-05 08:38:25.262078223 +0000 UTC m=+6205.790985484" Dec 05 08:38:28 crc kubenswrapper[4997]: I1205 08:38:28.749369 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:38:28 crc kubenswrapper[4997]: E1205 08:38:28.750321 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:38:39 crc kubenswrapper[4997]: I1205 08:38:39.393576 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 05 08:38:39 crc kubenswrapper[4997]: I1205 08:38:39.556550 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:38:42 crc kubenswrapper[4997]: I1205 08:38:42.749271 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:38:42 crc kubenswrapper[4997]: E1205 08:38:42.749776 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:38:50 crc kubenswrapper[4997]: I1205 08:38:50.853177 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Dec 05 08:38:50 crc kubenswrapper[4997]: E1205 08:38:50.854205 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="016b7cd0-efde-4d2c-940c-181108006a33" containerName="extract-utilities" Dec 05 08:38:50 crc kubenswrapper[4997]: I1205 08:38:50.854225 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="016b7cd0-efde-4d2c-940c-181108006a33" containerName="extract-utilities" Dec 05 08:38:50 crc kubenswrapper[4997]: E1205 08:38:50.854255 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9355424-f0c2-4a5c-9fce-5a48bc52887a" containerName="dnsmasq-dns" Dec 05 08:38:50 crc kubenswrapper[4997]: I1205 08:38:50.854263 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9355424-f0c2-4a5c-9fce-5a48bc52887a" containerName="dnsmasq-dns" Dec 05 08:38:50 crc kubenswrapper[4997]: E1205 08:38:50.854274 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="016b7cd0-efde-4d2c-940c-181108006a33" containerName="extract-content" Dec 05 08:38:50 crc kubenswrapper[4997]: I1205 08:38:50.854283 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="016b7cd0-efde-4d2c-940c-181108006a33" containerName="extract-content" Dec 05 08:38:50 crc kubenswrapper[4997]: E1205 08:38:50.854295 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9355424-f0c2-4a5c-9fce-5a48bc52887a" containerName="init" Dec 05 08:38:50 crc kubenswrapper[4997]: I1205 08:38:50.854302 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9355424-f0c2-4a5c-9fce-5a48bc52887a" containerName="init" Dec 05 08:38:50 crc kubenswrapper[4997]: E1205 08:38:50.854321 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="016b7cd0-efde-4d2c-940c-181108006a33" containerName="registry-server" Dec 05 08:38:50 crc kubenswrapper[4997]: I1205 08:38:50.854328 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="016b7cd0-efde-4d2c-940c-181108006a33" containerName="registry-server" Dec 05 08:38:50 crc kubenswrapper[4997]: I1205 08:38:50.854507 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="016b7cd0-efde-4d2c-940c-181108006a33" containerName="registry-server" Dec 05 08:38:50 crc kubenswrapper[4997]: I1205 08:38:50.854528 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9355424-f0c2-4a5c-9fce-5a48bc52887a" containerName="dnsmasq-dns" Dec 05 08:38:50 crc kubenswrapper[4997]: I1205 08:38:50.855205 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 05 08:38:50 crc kubenswrapper[4997]: I1205 08:38:50.856869 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-qsjdm" Dec 05 08:38:50 crc kubenswrapper[4997]: I1205 08:38:50.861920 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 05 08:38:50 crc kubenswrapper[4997]: I1205 08:38:50.964213 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svl8b\" (UniqueName: \"kubernetes.io/projected/ae541989-f021-4a96-8a1e-d5392bcc47b9-kube-api-access-svl8b\") pod \"mariadb-client-1-default\" (UID: \"ae541989-f021-4a96-8a1e-d5392bcc47b9\") " pod="openstack/mariadb-client-1-default" Dec 05 08:38:51 crc kubenswrapper[4997]: I1205 08:38:51.066215 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svl8b\" (UniqueName: \"kubernetes.io/projected/ae541989-f021-4a96-8a1e-d5392bcc47b9-kube-api-access-svl8b\") pod \"mariadb-client-1-default\" (UID: \"ae541989-f021-4a96-8a1e-d5392bcc47b9\") " pod="openstack/mariadb-client-1-default" Dec 05 08:38:51 crc kubenswrapper[4997]: I1205 08:38:51.085728 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svl8b\" (UniqueName: \"kubernetes.io/projected/ae541989-f021-4a96-8a1e-d5392bcc47b9-kube-api-access-svl8b\") pod \"mariadb-client-1-default\" (UID: \"ae541989-f021-4a96-8a1e-d5392bcc47b9\") " pod="openstack/mariadb-client-1-default" Dec 05 08:38:51 crc kubenswrapper[4997]: I1205 08:38:51.176146 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 05 08:38:51 crc kubenswrapper[4997]: I1205 08:38:51.460063 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 05 08:38:52 crc kubenswrapper[4997]: I1205 08:38:52.428407 4997 generic.go:334] "Generic (PLEG): container finished" podID="ae541989-f021-4a96-8a1e-d5392bcc47b9" containerID="716aacb225bde0a355384d8290185e03a4c2a8da45cd47231f685f2e7dbf45f5" exitCode=0 Dec 05 08:38:52 crc kubenswrapper[4997]: I1205 08:38:52.428471 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"ae541989-f021-4a96-8a1e-d5392bcc47b9","Type":"ContainerDied","Data":"716aacb225bde0a355384d8290185e03a4c2a8da45cd47231f685f2e7dbf45f5"} Dec 05 08:38:52 crc kubenswrapper[4997]: I1205 08:38:52.428703 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"ae541989-f021-4a96-8a1e-d5392bcc47b9","Type":"ContainerStarted","Data":"cddfe1d9ab2c88bdf3a3ba15fd83c804850bc5b0ad9077b82489fb7d196453e3"} Dec 05 08:38:53 crc kubenswrapper[4997]: I1205 08:38:53.766157 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 05 08:38:53 crc kubenswrapper[4997]: I1205 08:38:53.801806 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_ae541989-f021-4a96-8a1e-d5392bcc47b9/mariadb-client-1-default/0.log" Dec 05 08:38:53 crc kubenswrapper[4997]: I1205 08:38:53.830095 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 05 08:38:53 crc kubenswrapper[4997]: I1205 08:38:53.837320 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 05 08:38:53 crc kubenswrapper[4997]: I1205 08:38:53.888887 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svl8b\" (UniqueName: \"kubernetes.io/projected/ae541989-f021-4a96-8a1e-d5392bcc47b9-kube-api-access-svl8b\") pod \"ae541989-f021-4a96-8a1e-d5392bcc47b9\" (UID: \"ae541989-f021-4a96-8a1e-d5392bcc47b9\") " Dec 05 08:38:53 crc kubenswrapper[4997]: I1205 08:38:53.894139 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae541989-f021-4a96-8a1e-d5392bcc47b9-kube-api-access-svl8b" (OuterVolumeSpecName: "kube-api-access-svl8b") pod "ae541989-f021-4a96-8a1e-d5392bcc47b9" (UID: "ae541989-f021-4a96-8a1e-d5392bcc47b9"). InnerVolumeSpecName "kube-api-access-svl8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:53 crc kubenswrapper[4997]: I1205 08:38:53.991102 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svl8b\" (UniqueName: \"kubernetes.io/projected/ae541989-f021-4a96-8a1e-d5392bcc47b9-kube-api-access-svl8b\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:54 crc kubenswrapper[4997]: I1205 08:38:54.240340 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Dec 05 08:38:54 crc kubenswrapper[4997]: E1205 08:38:54.240769 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae541989-f021-4a96-8a1e-d5392bcc47b9" containerName="mariadb-client-1-default" Dec 05 08:38:54 crc kubenswrapper[4997]: I1205 08:38:54.240793 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae541989-f021-4a96-8a1e-d5392bcc47b9" containerName="mariadb-client-1-default" Dec 05 08:38:54 crc kubenswrapper[4997]: I1205 08:38:54.240992 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae541989-f021-4a96-8a1e-d5392bcc47b9" containerName="mariadb-client-1-default" Dec 05 08:38:54 crc kubenswrapper[4997]: I1205 08:38:54.241665 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 05 08:38:54 crc kubenswrapper[4997]: I1205 08:38:54.250761 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 05 08:38:54 crc kubenswrapper[4997]: I1205 08:38:54.296083 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72zzc\" (UniqueName: \"kubernetes.io/projected/8d8595bf-505e-45d5-ac2a-21e1fd3feebf-kube-api-access-72zzc\") pod \"mariadb-client-2-default\" (UID: \"8d8595bf-505e-45d5-ac2a-21e1fd3feebf\") " pod="openstack/mariadb-client-2-default" Dec 05 08:38:54 crc kubenswrapper[4997]: I1205 08:38:54.396971 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72zzc\" (UniqueName: \"kubernetes.io/projected/8d8595bf-505e-45d5-ac2a-21e1fd3feebf-kube-api-access-72zzc\") pod \"mariadb-client-2-default\" (UID: \"8d8595bf-505e-45d5-ac2a-21e1fd3feebf\") " pod="openstack/mariadb-client-2-default" Dec 05 08:38:54 crc kubenswrapper[4997]: I1205 08:38:54.416736 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72zzc\" (UniqueName: \"kubernetes.io/projected/8d8595bf-505e-45d5-ac2a-21e1fd3feebf-kube-api-access-72zzc\") pod \"mariadb-client-2-default\" (UID: \"8d8595bf-505e-45d5-ac2a-21e1fd3feebf\") " pod="openstack/mariadb-client-2-default" Dec 05 08:38:54 crc kubenswrapper[4997]: I1205 08:38:54.441328 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cddfe1d9ab2c88bdf3a3ba15fd83c804850bc5b0ad9077b82489fb7d196453e3" Dec 05 08:38:54 crc kubenswrapper[4997]: I1205 08:38:54.441382 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 05 08:38:54 crc kubenswrapper[4997]: I1205 08:38:54.563362 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 05 08:38:54 crc kubenswrapper[4997]: I1205 08:38:54.855581 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 05 08:38:55 crc kubenswrapper[4997]: I1205 08:38:55.450413 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"8d8595bf-505e-45d5-ac2a-21e1fd3feebf","Type":"ContainerStarted","Data":"5b77441652cf2b8510f7cfae5350a2349d44ca0a2b910f2c54bb3e8a35ea3b30"} Dec 05 08:38:55 crc kubenswrapper[4997]: I1205 08:38:55.450760 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"8d8595bf-505e-45d5-ac2a-21e1fd3feebf","Type":"ContainerStarted","Data":"4d05d4762d922f7ce4e65ce4fbe4dd890d1d486082996254e9b043d5cc59c5a4"} Dec 05 08:38:55 crc kubenswrapper[4997]: I1205 08:38:55.467234 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-2-default" podStartSLOduration=1.467209337 podStartE2EDuration="1.467209337s" podCreationTimestamp="2025-12-05 08:38:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:38:55.465696497 +0000 UTC m=+6235.994603778" watchObservedRunningTime="2025-12-05 08:38:55.467209337 +0000 UTC m=+6235.996116618" Dec 05 08:38:55 crc kubenswrapper[4997]: I1205 08:38:55.751152 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:38:55 crc kubenswrapper[4997]: E1205 08:38:55.751403 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:38:55 crc kubenswrapper[4997]: I1205 08:38:55.759145 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae541989-f021-4a96-8a1e-d5392bcc47b9" path="/var/lib/kubelet/pods/ae541989-f021-4a96-8a1e-d5392bcc47b9/volumes" Dec 05 08:38:56 crc kubenswrapper[4997]: I1205 08:38:56.459125 4997 generic.go:334] "Generic (PLEG): container finished" podID="8d8595bf-505e-45d5-ac2a-21e1fd3feebf" containerID="5b77441652cf2b8510f7cfae5350a2349d44ca0a2b910f2c54bb3e8a35ea3b30" exitCode=1 Dec 05 08:38:56 crc kubenswrapper[4997]: I1205 08:38:56.459196 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"8d8595bf-505e-45d5-ac2a-21e1fd3feebf","Type":"ContainerDied","Data":"5b77441652cf2b8510f7cfae5350a2349d44ca0a2b910f2c54bb3e8a35ea3b30"} Dec 05 08:38:57 crc kubenswrapper[4997]: I1205 08:38:57.884857 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 05 08:38:57 crc kubenswrapper[4997]: I1205 08:38:57.926344 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 05 08:38:57 crc kubenswrapper[4997]: I1205 08:38:57.931701 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 05 08:38:58 crc kubenswrapper[4997]: I1205 08:38:58.074393 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72zzc\" (UniqueName: \"kubernetes.io/projected/8d8595bf-505e-45d5-ac2a-21e1fd3feebf-kube-api-access-72zzc\") pod \"8d8595bf-505e-45d5-ac2a-21e1fd3feebf\" (UID: \"8d8595bf-505e-45d5-ac2a-21e1fd3feebf\") " Dec 05 08:38:58 crc kubenswrapper[4997]: I1205 08:38:58.080105 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d8595bf-505e-45d5-ac2a-21e1fd3feebf-kube-api-access-72zzc" (OuterVolumeSpecName: "kube-api-access-72zzc") pod "8d8595bf-505e-45d5-ac2a-21e1fd3feebf" (UID: "8d8595bf-505e-45d5-ac2a-21e1fd3feebf"). InnerVolumeSpecName "kube-api-access-72zzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:58 crc kubenswrapper[4997]: I1205 08:38:58.176669 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72zzc\" (UniqueName: \"kubernetes.io/projected/8d8595bf-505e-45d5-ac2a-21e1fd3feebf-kube-api-access-72zzc\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:58 crc kubenswrapper[4997]: I1205 08:38:58.323727 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Dec 05 08:38:58 crc kubenswrapper[4997]: E1205 08:38:58.324107 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d8595bf-505e-45d5-ac2a-21e1fd3feebf" containerName="mariadb-client-2-default" Dec 05 08:38:58 crc kubenswrapper[4997]: I1205 08:38:58.324125 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d8595bf-505e-45d5-ac2a-21e1fd3feebf" containerName="mariadb-client-2-default" Dec 05 08:38:58 crc kubenswrapper[4997]: I1205 08:38:58.324272 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d8595bf-505e-45d5-ac2a-21e1fd3feebf" containerName="mariadb-client-2-default" Dec 05 08:38:58 crc kubenswrapper[4997]: I1205 08:38:58.324808 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 05 08:38:58 crc kubenswrapper[4997]: I1205 08:38:58.330306 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Dec 05 08:38:58 crc kubenswrapper[4997]: I1205 08:38:58.379636 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6tpz\" (UniqueName: \"kubernetes.io/projected/26a9571a-c91c-4c79-b2d1-624f36ff880f-kube-api-access-x6tpz\") pod \"mariadb-client-1\" (UID: \"26a9571a-c91c-4c79-b2d1-624f36ff880f\") " pod="openstack/mariadb-client-1" Dec 05 08:38:58 crc kubenswrapper[4997]: I1205 08:38:58.480685 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6tpz\" (UniqueName: \"kubernetes.io/projected/26a9571a-c91c-4c79-b2d1-624f36ff880f-kube-api-access-x6tpz\") pod \"mariadb-client-1\" (UID: \"26a9571a-c91c-4c79-b2d1-624f36ff880f\") " pod="openstack/mariadb-client-1" Dec 05 08:38:58 crc kubenswrapper[4997]: I1205 08:38:58.482124 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d05d4762d922f7ce4e65ce4fbe4dd890d1d486082996254e9b043d5cc59c5a4" Dec 05 08:38:58 crc kubenswrapper[4997]: I1205 08:38:58.482141 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 05 08:38:58 crc kubenswrapper[4997]: I1205 08:38:58.499425 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6tpz\" (UniqueName: \"kubernetes.io/projected/26a9571a-c91c-4c79-b2d1-624f36ff880f-kube-api-access-x6tpz\") pod \"mariadb-client-1\" (UID: \"26a9571a-c91c-4c79-b2d1-624f36ff880f\") " pod="openstack/mariadb-client-1" Dec 05 08:38:58 crc kubenswrapper[4997]: I1205 08:38:58.648345 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 05 08:38:59 crc kubenswrapper[4997]: I1205 08:38:59.026377 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Dec 05 08:38:59 crc kubenswrapper[4997]: I1205 08:38:59.492687 4997 generic.go:334] "Generic (PLEG): container finished" podID="26a9571a-c91c-4c79-b2d1-624f36ff880f" containerID="9b12990500a504f1a48a30db2d1509ad56e1e898de80ba25b1e0e7e5318293e7" exitCode=0 Dec 05 08:38:59 crc kubenswrapper[4997]: I1205 08:38:59.492813 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"26a9571a-c91c-4c79-b2d1-624f36ff880f","Type":"ContainerDied","Data":"9b12990500a504f1a48a30db2d1509ad56e1e898de80ba25b1e0e7e5318293e7"} Dec 05 08:38:59 crc kubenswrapper[4997]: I1205 08:38:59.492991 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"26a9571a-c91c-4c79-b2d1-624f36ff880f","Type":"ContainerStarted","Data":"170558edb91c0d1605ef7c11adfe5aedc8dd8cb7e114913498df065993e3fdf3"} Dec 05 08:38:59 crc kubenswrapper[4997]: I1205 08:38:59.761364 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d8595bf-505e-45d5-ac2a-21e1fd3feebf" path="/var/lib/kubelet/pods/8d8595bf-505e-45d5-ac2a-21e1fd3feebf/volumes" Dec 05 08:39:00 crc kubenswrapper[4997]: I1205 08:39:00.843439 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 05 08:39:00 crc kubenswrapper[4997]: I1205 08:39:00.862582 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_26a9571a-c91c-4c79-b2d1-624f36ff880f/mariadb-client-1/0.log" Dec 05 08:39:00 crc kubenswrapper[4997]: I1205 08:39:00.891727 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Dec 05 08:39:00 crc kubenswrapper[4997]: I1205 08:39:00.897957 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Dec 05 08:39:01 crc kubenswrapper[4997]: I1205 08:39:01.026872 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6tpz\" (UniqueName: \"kubernetes.io/projected/26a9571a-c91c-4c79-b2d1-624f36ff880f-kube-api-access-x6tpz\") pod \"26a9571a-c91c-4c79-b2d1-624f36ff880f\" (UID: \"26a9571a-c91c-4c79-b2d1-624f36ff880f\") " Dec 05 08:39:01 crc kubenswrapper[4997]: I1205 08:39:01.032847 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26a9571a-c91c-4c79-b2d1-624f36ff880f-kube-api-access-x6tpz" (OuterVolumeSpecName: "kube-api-access-x6tpz") pod "26a9571a-c91c-4c79-b2d1-624f36ff880f" (UID: "26a9571a-c91c-4c79-b2d1-624f36ff880f"). InnerVolumeSpecName "kube-api-access-x6tpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:39:01 crc kubenswrapper[4997]: I1205 08:39:01.129825 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6tpz\" (UniqueName: \"kubernetes.io/projected/26a9571a-c91c-4c79-b2d1-624f36ff880f-kube-api-access-x6tpz\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:01 crc kubenswrapper[4997]: I1205 08:39:01.303688 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Dec 05 08:39:01 crc kubenswrapper[4997]: E1205 08:39:01.304105 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a9571a-c91c-4c79-b2d1-624f36ff880f" containerName="mariadb-client-1" Dec 05 08:39:01 crc kubenswrapper[4997]: I1205 08:39:01.304126 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a9571a-c91c-4c79-b2d1-624f36ff880f" containerName="mariadb-client-1" Dec 05 08:39:01 crc kubenswrapper[4997]: I1205 08:39:01.304320 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="26a9571a-c91c-4c79-b2d1-624f36ff880f" containerName="mariadb-client-1" Dec 05 08:39:01 crc kubenswrapper[4997]: I1205 08:39:01.304882 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 05 08:39:01 crc kubenswrapper[4997]: I1205 08:39:01.311991 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 05 08:39:01 crc kubenswrapper[4997]: I1205 08:39:01.434002 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nmgt\" (UniqueName: \"kubernetes.io/projected/448368f2-ee28-434f-9082-45578fb23b77-kube-api-access-7nmgt\") pod \"mariadb-client-4-default\" (UID: \"448368f2-ee28-434f-9082-45578fb23b77\") " pod="openstack/mariadb-client-4-default" Dec 05 08:39:01 crc kubenswrapper[4997]: I1205 08:39:01.509535 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="170558edb91c0d1605ef7c11adfe5aedc8dd8cb7e114913498df065993e3fdf3" Dec 05 08:39:01 crc kubenswrapper[4997]: I1205 08:39:01.509640 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 05 08:39:01 crc kubenswrapper[4997]: I1205 08:39:01.535506 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nmgt\" (UniqueName: \"kubernetes.io/projected/448368f2-ee28-434f-9082-45578fb23b77-kube-api-access-7nmgt\") pod \"mariadb-client-4-default\" (UID: \"448368f2-ee28-434f-9082-45578fb23b77\") " pod="openstack/mariadb-client-4-default" Dec 05 08:39:01 crc kubenswrapper[4997]: I1205 08:39:01.553247 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nmgt\" (UniqueName: \"kubernetes.io/projected/448368f2-ee28-434f-9082-45578fb23b77-kube-api-access-7nmgt\") pod \"mariadb-client-4-default\" (UID: \"448368f2-ee28-434f-9082-45578fb23b77\") " pod="openstack/mariadb-client-4-default" Dec 05 08:39:01 crc kubenswrapper[4997]: I1205 08:39:01.624216 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 05 08:39:01 crc kubenswrapper[4997]: I1205 08:39:01.761694 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26a9571a-c91c-4c79-b2d1-624f36ff880f" path="/var/lib/kubelet/pods/26a9571a-c91c-4c79-b2d1-624f36ff880f/volumes" Dec 05 08:39:01 crc kubenswrapper[4997]: I1205 08:39:01.870992 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 05 08:39:02 crc kubenswrapper[4997]: I1205 08:39:02.519487 4997 generic.go:334] "Generic (PLEG): container finished" podID="448368f2-ee28-434f-9082-45578fb23b77" containerID="b127ee1cca7e4caf274e99d2f311479d7237ca9d71238ac8ea3153222de6e7c3" exitCode=0 Dec 05 08:39:02 crc kubenswrapper[4997]: I1205 08:39:02.519588 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"448368f2-ee28-434f-9082-45578fb23b77","Type":"ContainerDied","Data":"b127ee1cca7e4caf274e99d2f311479d7237ca9d71238ac8ea3153222de6e7c3"} Dec 05 08:39:02 crc kubenswrapper[4997]: I1205 08:39:02.519895 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"448368f2-ee28-434f-9082-45578fb23b77","Type":"ContainerStarted","Data":"d903f09e2ef1ce94c41f3dc6007e4efd3ee8a3fadde544345a49e91ed662bf64"} Dec 05 08:39:03 crc kubenswrapper[4997]: I1205 08:39:03.874859 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 05 08:39:03 crc kubenswrapper[4997]: I1205 08:39:03.892962 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_448368f2-ee28-434f-9082-45578fb23b77/mariadb-client-4-default/0.log" Dec 05 08:39:03 crc kubenswrapper[4997]: I1205 08:39:03.916525 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 05 08:39:03 crc kubenswrapper[4997]: I1205 08:39:03.920632 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 05 08:39:03 crc kubenswrapper[4997]: I1205 08:39:03.974347 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nmgt\" (UniqueName: \"kubernetes.io/projected/448368f2-ee28-434f-9082-45578fb23b77-kube-api-access-7nmgt\") pod \"448368f2-ee28-434f-9082-45578fb23b77\" (UID: \"448368f2-ee28-434f-9082-45578fb23b77\") " Dec 05 08:39:03 crc kubenswrapper[4997]: I1205 08:39:03.981167 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/448368f2-ee28-434f-9082-45578fb23b77-kube-api-access-7nmgt" (OuterVolumeSpecName: "kube-api-access-7nmgt") pod "448368f2-ee28-434f-9082-45578fb23b77" (UID: "448368f2-ee28-434f-9082-45578fb23b77"). InnerVolumeSpecName "kube-api-access-7nmgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:39:04 crc kubenswrapper[4997]: I1205 08:39:04.076196 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nmgt\" (UniqueName: \"kubernetes.io/projected/448368f2-ee28-434f-9082-45578fb23b77-kube-api-access-7nmgt\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:04 crc kubenswrapper[4997]: I1205 08:39:04.533860 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d903f09e2ef1ce94c41f3dc6007e4efd3ee8a3fadde544345a49e91ed662bf64" Dec 05 08:39:04 crc kubenswrapper[4997]: I1205 08:39:04.533897 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 05 08:39:05 crc kubenswrapper[4997]: I1205 08:39:05.763695 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="448368f2-ee28-434f-9082-45578fb23b77" path="/var/lib/kubelet/pods/448368f2-ee28-434f-9082-45578fb23b77/volumes" Dec 05 08:39:07 crc kubenswrapper[4997]: I1205 08:39:07.606623 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Dec 05 08:39:07 crc kubenswrapper[4997]: E1205 08:39:07.606904 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="448368f2-ee28-434f-9082-45578fb23b77" containerName="mariadb-client-4-default" Dec 05 08:39:07 crc kubenswrapper[4997]: I1205 08:39:07.606915 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="448368f2-ee28-434f-9082-45578fb23b77" containerName="mariadb-client-4-default" Dec 05 08:39:07 crc kubenswrapper[4997]: I1205 08:39:07.607054 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="448368f2-ee28-434f-9082-45578fb23b77" containerName="mariadb-client-4-default" Dec 05 08:39:07 crc kubenswrapper[4997]: I1205 08:39:07.607511 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 05 08:39:07 crc kubenswrapper[4997]: I1205 08:39:07.610435 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-qsjdm" Dec 05 08:39:07 crc kubenswrapper[4997]: I1205 08:39:07.614024 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 05 08:39:07 crc kubenswrapper[4997]: I1205 08:39:07.728032 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlbvq\" (UniqueName: \"kubernetes.io/projected/6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2-kube-api-access-nlbvq\") pod \"mariadb-client-5-default\" (UID: \"6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2\") " pod="openstack/mariadb-client-5-default" Dec 05 08:39:07 crc kubenswrapper[4997]: I1205 08:39:07.829516 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlbvq\" (UniqueName: \"kubernetes.io/projected/6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2-kube-api-access-nlbvq\") pod \"mariadb-client-5-default\" (UID: \"6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2\") " pod="openstack/mariadb-client-5-default" Dec 05 08:39:07 crc kubenswrapper[4997]: I1205 08:39:07.848859 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlbvq\" (UniqueName: \"kubernetes.io/projected/6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2-kube-api-access-nlbvq\") pod \"mariadb-client-5-default\" (UID: \"6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2\") " pod="openstack/mariadb-client-5-default" Dec 05 08:39:07 crc kubenswrapper[4997]: I1205 08:39:07.928260 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 05 08:39:08 crc kubenswrapper[4997]: I1205 08:39:08.411499 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 05 08:39:09 crc kubenswrapper[4997]: I1205 08:39:08.562584 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2","Type":"ContainerStarted","Data":"ee7870455e6c6569850cb1964e127e39989f97c756b3d5e792623955e9febea1"} Dec 05 08:39:09 crc kubenswrapper[4997]: I1205 08:39:09.572191 4997 generic.go:334] "Generic (PLEG): container finished" podID="6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2" containerID="5cd05b3261d8f22e530138087a11b10af4ceb446f5634a76f8a2dee49b01abe1" exitCode=0 Dec 05 08:39:09 crc kubenswrapper[4997]: I1205 08:39:09.572313 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2","Type":"ContainerDied","Data":"5cd05b3261d8f22e530138087a11b10af4ceb446f5634a76f8a2dee49b01abe1"} Dec 05 08:39:10 crc kubenswrapper[4997]: I1205 08:39:10.748543 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:39:10 crc kubenswrapper[4997]: E1205 08:39:10.748821 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:39:10 crc kubenswrapper[4997]: I1205 08:39:10.939515 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 05 08:39:10 crc kubenswrapper[4997]: I1205 08:39:10.960783 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2/mariadb-client-5-default/0.log" Dec 05 08:39:10 crc kubenswrapper[4997]: I1205 08:39:10.991970 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 05 08:39:10 crc kubenswrapper[4997]: I1205 08:39:10.999651 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 05 08:39:11 crc kubenswrapper[4997]: I1205 08:39:11.075794 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlbvq\" (UniqueName: \"kubernetes.io/projected/6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2-kube-api-access-nlbvq\") pod \"6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2\" (UID: \"6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2\") " Dec 05 08:39:11 crc kubenswrapper[4997]: I1205 08:39:11.081057 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2-kube-api-access-nlbvq" (OuterVolumeSpecName: "kube-api-access-nlbvq") pod "6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2" (UID: "6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2"). InnerVolumeSpecName "kube-api-access-nlbvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:39:11 crc kubenswrapper[4997]: I1205 08:39:11.140643 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Dec 05 08:39:11 crc kubenswrapper[4997]: E1205 08:39:11.141120 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2" containerName="mariadb-client-5-default" Dec 05 08:39:11 crc kubenswrapper[4997]: I1205 08:39:11.141152 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2" containerName="mariadb-client-5-default" Dec 05 08:39:11 crc kubenswrapper[4997]: I1205 08:39:11.141425 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2" containerName="mariadb-client-5-default" Dec 05 08:39:11 crc kubenswrapper[4997]: I1205 08:39:11.142218 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 05 08:39:11 crc kubenswrapper[4997]: I1205 08:39:11.150294 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 05 08:39:11 crc kubenswrapper[4997]: I1205 08:39:11.178188 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlbvq\" (UniqueName: \"kubernetes.io/projected/6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2-kube-api-access-nlbvq\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:11 crc kubenswrapper[4997]: I1205 08:39:11.279270 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwwf5\" (UniqueName: \"kubernetes.io/projected/b1f784d8-43e9-4f7b-89f7-5d89e215aea3-kube-api-access-lwwf5\") pod \"mariadb-client-6-default\" (UID: \"b1f784d8-43e9-4f7b-89f7-5d89e215aea3\") " pod="openstack/mariadb-client-6-default" Dec 05 08:39:11 crc kubenswrapper[4997]: I1205 08:39:11.380555 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwwf5\" (UniqueName: \"kubernetes.io/projected/b1f784d8-43e9-4f7b-89f7-5d89e215aea3-kube-api-access-lwwf5\") pod \"mariadb-client-6-default\" (UID: \"b1f784d8-43e9-4f7b-89f7-5d89e215aea3\") " pod="openstack/mariadb-client-6-default" Dec 05 08:39:11 crc kubenswrapper[4997]: I1205 08:39:11.400384 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwwf5\" (UniqueName: \"kubernetes.io/projected/b1f784d8-43e9-4f7b-89f7-5d89e215aea3-kube-api-access-lwwf5\") pod \"mariadb-client-6-default\" (UID: \"b1f784d8-43e9-4f7b-89f7-5d89e215aea3\") " pod="openstack/mariadb-client-6-default" Dec 05 08:39:11 crc kubenswrapper[4997]: I1205 08:39:11.465505 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 05 08:39:11 crc kubenswrapper[4997]: I1205 08:39:11.595561 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee7870455e6c6569850cb1964e127e39989f97c756b3d5e792623955e9febea1" Dec 05 08:39:11 crc kubenswrapper[4997]: I1205 08:39:11.595934 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 05 08:39:11 crc kubenswrapper[4997]: I1205 08:39:11.760650 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2" path="/var/lib/kubelet/pods/6ce2dfa8-a468-40db-8c6e-dde6c34d0ab2/volumes" Dec 05 08:39:11 crc kubenswrapper[4997]: I1205 08:39:11.782805 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 05 08:39:11 crc kubenswrapper[4997]: W1205 08:39:11.786139 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1f784d8_43e9_4f7b_89f7_5d89e215aea3.slice/crio-41de006e36686a9bcfaac90c53b999bef5b0b32aa8c7f808a87e32c71bc2709c WatchSource:0}: Error finding container 41de006e36686a9bcfaac90c53b999bef5b0b32aa8c7f808a87e32c71bc2709c: Status 404 returned error can't find the container with id 41de006e36686a9bcfaac90c53b999bef5b0b32aa8c7f808a87e32c71bc2709c Dec 05 08:39:12 crc kubenswrapper[4997]: I1205 08:39:12.606143 4997 generic.go:334] "Generic (PLEG): container finished" podID="b1f784d8-43e9-4f7b-89f7-5d89e215aea3" containerID="3b1b5ab4c8c61e83c90d2b4ca5180354d81cf61588a5775a1fb750bd58bc29cc" exitCode=1 Dec 05 08:39:12 crc kubenswrapper[4997]: I1205 08:39:12.606246 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"b1f784d8-43e9-4f7b-89f7-5d89e215aea3","Type":"ContainerDied","Data":"3b1b5ab4c8c61e83c90d2b4ca5180354d81cf61588a5775a1fb750bd58bc29cc"} Dec 05 08:39:12 crc kubenswrapper[4997]: I1205 08:39:12.606716 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"b1f784d8-43e9-4f7b-89f7-5d89e215aea3","Type":"ContainerStarted","Data":"41de006e36686a9bcfaac90c53b999bef5b0b32aa8c7f808a87e32c71bc2709c"} Dec 05 08:39:13 crc kubenswrapper[4997]: I1205 08:39:13.948736 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 05 08:39:13 crc kubenswrapper[4997]: I1205 08:39:13.967512 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-6-default_b1f784d8-43e9-4f7b-89f7-5d89e215aea3/mariadb-client-6-default/0.log" Dec 05 08:39:13 crc kubenswrapper[4997]: I1205 08:39:13.988548 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 05 08:39:13 crc kubenswrapper[4997]: I1205 08:39:13.997479 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 05 08:39:14 crc kubenswrapper[4997]: I1205 08:39:14.121837 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwwf5\" (UniqueName: \"kubernetes.io/projected/b1f784d8-43e9-4f7b-89f7-5d89e215aea3-kube-api-access-lwwf5\") pod \"b1f784d8-43e9-4f7b-89f7-5d89e215aea3\" (UID: \"b1f784d8-43e9-4f7b-89f7-5d89e215aea3\") " Dec 05 08:39:14 crc kubenswrapper[4997]: I1205 08:39:14.126541 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Dec 05 08:39:14 crc kubenswrapper[4997]: E1205 08:39:14.126955 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1f784d8-43e9-4f7b-89f7-5d89e215aea3" containerName="mariadb-client-6-default" Dec 05 08:39:14 crc kubenswrapper[4997]: I1205 08:39:14.126982 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1f784d8-43e9-4f7b-89f7-5d89e215aea3" containerName="mariadb-client-6-default" Dec 05 08:39:14 crc kubenswrapper[4997]: I1205 08:39:14.127248 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1f784d8-43e9-4f7b-89f7-5d89e215aea3" containerName="mariadb-client-6-default" Dec 05 08:39:14 crc kubenswrapper[4997]: I1205 08:39:14.127954 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 05 08:39:14 crc kubenswrapper[4997]: I1205 08:39:14.129518 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1f784d8-43e9-4f7b-89f7-5d89e215aea3-kube-api-access-lwwf5" (OuterVolumeSpecName: "kube-api-access-lwwf5") pod "b1f784d8-43e9-4f7b-89f7-5d89e215aea3" (UID: "b1f784d8-43e9-4f7b-89f7-5d89e215aea3"). InnerVolumeSpecName "kube-api-access-lwwf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:39:14 crc kubenswrapper[4997]: I1205 08:39:14.140915 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 05 08:39:14 crc kubenswrapper[4997]: I1205 08:39:14.223646 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwwf5\" (UniqueName: \"kubernetes.io/projected/b1f784d8-43e9-4f7b-89f7-5d89e215aea3-kube-api-access-lwwf5\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:14 crc kubenswrapper[4997]: I1205 08:39:14.324870 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh6tx\" (UniqueName: \"kubernetes.io/projected/3239239f-8c7a-4398-8468-a13f4282bff5-kube-api-access-hh6tx\") pod \"mariadb-client-7-default\" (UID: \"3239239f-8c7a-4398-8468-a13f4282bff5\") " pod="openstack/mariadb-client-7-default" Dec 05 08:39:14 crc kubenswrapper[4997]: I1205 08:39:14.426277 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh6tx\" (UniqueName: \"kubernetes.io/projected/3239239f-8c7a-4398-8468-a13f4282bff5-kube-api-access-hh6tx\") pod \"mariadb-client-7-default\" (UID: \"3239239f-8c7a-4398-8468-a13f4282bff5\") " pod="openstack/mariadb-client-7-default" Dec 05 08:39:14 crc kubenswrapper[4997]: I1205 08:39:14.442791 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh6tx\" (UniqueName: \"kubernetes.io/projected/3239239f-8c7a-4398-8468-a13f4282bff5-kube-api-access-hh6tx\") pod \"mariadb-client-7-default\" (UID: \"3239239f-8c7a-4398-8468-a13f4282bff5\") " pod="openstack/mariadb-client-7-default" Dec 05 08:39:14 crc kubenswrapper[4997]: I1205 08:39:14.471997 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 05 08:39:14 crc kubenswrapper[4997]: I1205 08:39:14.625989 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41de006e36686a9bcfaac90c53b999bef5b0b32aa8c7f808a87e32c71bc2709c" Dec 05 08:39:14 crc kubenswrapper[4997]: I1205 08:39:14.626224 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 05 08:39:14 crc kubenswrapper[4997]: I1205 08:39:14.754092 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 05 08:39:14 crc kubenswrapper[4997]: W1205 08:39:14.762853 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3239239f_8c7a_4398_8468_a13f4282bff5.slice/crio-9a70b8648742b1fc82c4228ed8f0af5878422708d5c5fcfc0607b45eaaee395e WatchSource:0}: Error finding container 9a70b8648742b1fc82c4228ed8f0af5878422708d5c5fcfc0607b45eaaee395e: Status 404 returned error can't find the container with id 9a70b8648742b1fc82c4228ed8f0af5878422708d5c5fcfc0607b45eaaee395e Dec 05 08:39:15 crc kubenswrapper[4997]: I1205 08:39:15.636311 4997 generic.go:334] "Generic (PLEG): container finished" podID="3239239f-8c7a-4398-8468-a13f4282bff5" containerID="1460f98105d83e46357e1b6281611d947507cf84c1ea698c1cf876940cd5eaee" exitCode=0 Dec 05 08:39:15 crc kubenswrapper[4997]: I1205 08:39:15.636435 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"3239239f-8c7a-4398-8468-a13f4282bff5","Type":"ContainerDied","Data":"1460f98105d83e46357e1b6281611d947507cf84c1ea698c1cf876940cd5eaee"} Dec 05 08:39:15 crc kubenswrapper[4997]: I1205 08:39:15.636692 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"3239239f-8c7a-4398-8468-a13f4282bff5","Type":"ContainerStarted","Data":"9a70b8648742b1fc82c4228ed8f0af5878422708d5c5fcfc0607b45eaaee395e"} Dec 05 08:39:15 crc kubenswrapper[4997]: I1205 08:39:15.761758 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1f784d8-43e9-4f7b-89f7-5d89e215aea3" path="/var/lib/kubelet/pods/b1f784d8-43e9-4f7b-89f7-5d89e215aea3/volumes" Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.022343 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.047509 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_3239239f-8c7a-4398-8468-a13f4282bff5/mariadb-client-7-default/0.log" Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.075236 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hh6tx\" (UniqueName: \"kubernetes.io/projected/3239239f-8c7a-4398-8468-a13f4282bff5-kube-api-access-hh6tx\") pod \"3239239f-8c7a-4398-8468-a13f4282bff5\" (UID: \"3239239f-8c7a-4398-8468-a13f4282bff5\") " Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.082679 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.093562 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.097439 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3239239f-8c7a-4398-8468-a13f4282bff5-kube-api-access-hh6tx" (OuterVolumeSpecName: "kube-api-access-hh6tx") pod "3239239f-8c7a-4398-8468-a13f4282bff5" (UID: "3239239f-8c7a-4398-8468-a13f4282bff5"). InnerVolumeSpecName "kube-api-access-hh6tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.176436 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hh6tx\" (UniqueName: \"kubernetes.io/projected/3239239f-8c7a-4398-8468-a13f4282bff5-kube-api-access-hh6tx\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.205879 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Dec 05 08:39:17 crc kubenswrapper[4997]: E1205 08:39:17.206193 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3239239f-8c7a-4398-8468-a13f4282bff5" containerName="mariadb-client-7-default" Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.206211 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3239239f-8c7a-4398-8468-a13f4282bff5" containerName="mariadb-client-7-default" Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.206366 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="3239239f-8c7a-4398-8468-a13f4282bff5" containerName="mariadb-client-7-default" Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.206890 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.221585 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.379130 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xfjk\" (UniqueName: \"kubernetes.io/projected/63c586b5-df00-49cd-8fa1-ebcbd779ffcd-kube-api-access-7xfjk\") pod \"mariadb-client-2\" (UID: \"63c586b5-df00-49cd-8fa1-ebcbd779ffcd\") " pod="openstack/mariadb-client-2" Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.480401 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xfjk\" (UniqueName: \"kubernetes.io/projected/63c586b5-df00-49cd-8fa1-ebcbd779ffcd-kube-api-access-7xfjk\") pod \"mariadb-client-2\" (UID: \"63c586b5-df00-49cd-8fa1-ebcbd779ffcd\") " pod="openstack/mariadb-client-2" Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.500888 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xfjk\" (UniqueName: \"kubernetes.io/projected/63c586b5-df00-49cd-8fa1-ebcbd779ffcd-kube-api-access-7xfjk\") pod \"mariadb-client-2\" (UID: \"63c586b5-df00-49cd-8fa1-ebcbd779ffcd\") " pod="openstack/mariadb-client-2" Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.533857 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.653827 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a70b8648742b1fc82c4228ed8f0af5878422708d5c5fcfc0607b45eaaee395e" Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.653856 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.760634 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3239239f-8c7a-4398-8468-a13f4282bff5" path="/var/lib/kubelet/pods/3239239f-8c7a-4398-8468-a13f4282bff5/volumes" Dec 05 08:39:17 crc kubenswrapper[4997]: I1205 08:39:17.843931 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Dec 05 08:39:17 crc kubenswrapper[4997]: W1205 08:39:17.850984 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63c586b5_df00_49cd_8fa1_ebcbd779ffcd.slice/crio-c58e7974c9d35a3671b3ae985d7019f521196d5793f4ea760e4babe708f236d1 WatchSource:0}: Error finding container c58e7974c9d35a3671b3ae985d7019f521196d5793f4ea760e4babe708f236d1: Status 404 returned error can't find the container with id c58e7974c9d35a3671b3ae985d7019f521196d5793f4ea760e4babe708f236d1 Dec 05 08:39:18 crc kubenswrapper[4997]: I1205 08:39:18.661876 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"63c586b5-df00-49cd-8fa1-ebcbd779ffcd","Type":"ContainerStarted","Data":"4943f8b18da4d09c7e6dc46140e2c94f29b236eb1a134e9343efecdc42c6179f"} Dec 05 08:39:18 crc kubenswrapper[4997]: I1205 08:39:18.662091 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"63c586b5-df00-49cd-8fa1-ebcbd779ffcd","Type":"ContainerStarted","Data":"c58e7974c9d35a3671b3ae985d7019f521196d5793f4ea760e4babe708f236d1"} Dec 05 08:39:19 crc kubenswrapper[4997]: I1205 08:39:19.670243 4997 generic.go:334] "Generic (PLEG): container finished" podID="63c586b5-df00-49cd-8fa1-ebcbd779ffcd" containerID="4943f8b18da4d09c7e6dc46140e2c94f29b236eb1a134e9343efecdc42c6179f" exitCode=0 Dec 05 08:39:19 crc kubenswrapper[4997]: I1205 08:39:19.670287 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"63c586b5-df00-49cd-8fa1-ebcbd779ffcd","Type":"ContainerDied","Data":"4943f8b18da4d09c7e6dc46140e2c94f29b236eb1a134e9343efecdc42c6179f"} Dec 05 08:39:21 crc kubenswrapper[4997]: I1205 08:39:21.006532 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 05 08:39:21 crc kubenswrapper[4997]: I1205 08:39:21.024410 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_63c586b5-df00-49cd-8fa1-ebcbd779ffcd/mariadb-client-2/0.log" Dec 05 08:39:21 crc kubenswrapper[4997]: I1205 08:39:21.048598 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Dec 05 08:39:21 crc kubenswrapper[4997]: I1205 08:39:21.055357 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Dec 05 08:39:21 crc kubenswrapper[4997]: I1205 08:39:21.137277 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xfjk\" (UniqueName: \"kubernetes.io/projected/63c586b5-df00-49cd-8fa1-ebcbd779ffcd-kube-api-access-7xfjk\") pod \"63c586b5-df00-49cd-8fa1-ebcbd779ffcd\" (UID: \"63c586b5-df00-49cd-8fa1-ebcbd779ffcd\") " Dec 05 08:39:21 crc kubenswrapper[4997]: I1205 08:39:21.143271 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63c586b5-df00-49cd-8fa1-ebcbd779ffcd-kube-api-access-7xfjk" (OuterVolumeSpecName: "kube-api-access-7xfjk") pod "63c586b5-df00-49cd-8fa1-ebcbd779ffcd" (UID: "63c586b5-df00-49cd-8fa1-ebcbd779ffcd"). InnerVolumeSpecName "kube-api-access-7xfjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:39:21 crc kubenswrapper[4997]: I1205 08:39:21.239258 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xfjk\" (UniqueName: \"kubernetes.io/projected/63c586b5-df00-49cd-8fa1-ebcbd779ffcd-kube-api-access-7xfjk\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:21 crc kubenswrapper[4997]: I1205 08:39:21.685469 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c58e7974c9d35a3671b3ae985d7019f521196d5793f4ea760e4babe708f236d1" Dec 05 08:39:21 crc kubenswrapper[4997]: I1205 08:39:21.685530 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 05 08:39:21 crc kubenswrapper[4997]: I1205 08:39:21.759450 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63c586b5-df00-49cd-8fa1-ebcbd779ffcd" path="/var/lib/kubelet/pods/63c586b5-df00-49cd-8fa1-ebcbd779ffcd/volumes" Dec 05 08:39:23 crc kubenswrapper[4997]: I1205 08:39:23.750005 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:39:23 crc kubenswrapper[4997]: E1205 08:39:23.750300 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:39:38 crc kubenswrapper[4997]: I1205 08:39:38.749340 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:39:38 crc kubenswrapper[4997]: E1205 08:39:38.750420 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:39:50 crc kubenswrapper[4997]: I1205 08:39:50.749954 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:39:51 crc kubenswrapper[4997]: I1205 08:39:51.923475 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"1ae3926813fc853136a11e68578c56580649609079883104b1edf72cf4d67701"} Dec 05 08:40:12 crc kubenswrapper[4997]: I1205 08:40:12.733576 4997 scope.go:117] "RemoveContainer" containerID="cce72cf13bf12378384ca3a19d45d45b633e16a929d088ee083df1ab78f57c6c" Dec 05 08:40:30 crc kubenswrapper[4997]: I1205 08:40:30.768203 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9fqxz"] Dec 05 08:40:30 crc kubenswrapper[4997]: E1205 08:40:30.769184 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63c586b5-df00-49cd-8fa1-ebcbd779ffcd" containerName="mariadb-client-2" Dec 05 08:40:30 crc kubenswrapper[4997]: I1205 08:40:30.769199 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="63c586b5-df00-49cd-8fa1-ebcbd779ffcd" containerName="mariadb-client-2" Dec 05 08:40:30 crc kubenswrapper[4997]: I1205 08:40:30.769372 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="63c586b5-df00-49cd-8fa1-ebcbd779ffcd" containerName="mariadb-client-2" Dec 05 08:40:30 crc kubenswrapper[4997]: I1205 08:40:30.770709 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9fqxz" Dec 05 08:40:30 crc kubenswrapper[4997]: I1205 08:40:30.775331 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9fqxz"] Dec 05 08:40:30 crc kubenswrapper[4997]: I1205 08:40:30.949231 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05e72d6a-37da-4550-96f3-0e7de464115d-catalog-content\") pod \"redhat-operators-9fqxz\" (UID: \"05e72d6a-37da-4550-96f3-0e7de464115d\") " pod="openshift-marketplace/redhat-operators-9fqxz" Dec 05 08:40:30 crc kubenswrapper[4997]: I1205 08:40:30.949304 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05e72d6a-37da-4550-96f3-0e7de464115d-utilities\") pod \"redhat-operators-9fqxz\" (UID: \"05e72d6a-37da-4550-96f3-0e7de464115d\") " pod="openshift-marketplace/redhat-operators-9fqxz" Dec 05 08:40:30 crc kubenswrapper[4997]: I1205 08:40:30.949335 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsqzl\" (UniqueName: \"kubernetes.io/projected/05e72d6a-37da-4550-96f3-0e7de464115d-kube-api-access-nsqzl\") pod \"redhat-operators-9fqxz\" (UID: \"05e72d6a-37da-4550-96f3-0e7de464115d\") " pod="openshift-marketplace/redhat-operators-9fqxz" Dec 05 08:40:31 crc kubenswrapper[4997]: I1205 08:40:31.050943 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05e72d6a-37da-4550-96f3-0e7de464115d-catalog-content\") pod \"redhat-operators-9fqxz\" (UID: \"05e72d6a-37da-4550-96f3-0e7de464115d\") " pod="openshift-marketplace/redhat-operators-9fqxz" Dec 05 08:40:31 crc kubenswrapper[4997]: I1205 08:40:31.051029 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05e72d6a-37da-4550-96f3-0e7de464115d-utilities\") pod \"redhat-operators-9fqxz\" (UID: \"05e72d6a-37da-4550-96f3-0e7de464115d\") " pod="openshift-marketplace/redhat-operators-9fqxz" Dec 05 08:40:31 crc kubenswrapper[4997]: I1205 08:40:31.051070 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsqzl\" (UniqueName: \"kubernetes.io/projected/05e72d6a-37da-4550-96f3-0e7de464115d-kube-api-access-nsqzl\") pod \"redhat-operators-9fqxz\" (UID: \"05e72d6a-37da-4550-96f3-0e7de464115d\") " pod="openshift-marketplace/redhat-operators-9fqxz" Dec 05 08:40:31 crc kubenswrapper[4997]: I1205 08:40:31.051602 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05e72d6a-37da-4550-96f3-0e7de464115d-catalog-content\") pod \"redhat-operators-9fqxz\" (UID: \"05e72d6a-37da-4550-96f3-0e7de464115d\") " pod="openshift-marketplace/redhat-operators-9fqxz" Dec 05 08:40:31 crc kubenswrapper[4997]: I1205 08:40:31.051703 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05e72d6a-37da-4550-96f3-0e7de464115d-utilities\") pod \"redhat-operators-9fqxz\" (UID: \"05e72d6a-37da-4550-96f3-0e7de464115d\") " pod="openshift-marketplace/redhat-operators-9fqxz" Dec 05 08:40:31 crc kubenswrapper[4997]: I1205 08:40:31.071652 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsqzl\" (UniqueName: \"kubernetes.io/projected/05e72d6a-37da-4550-96f3-0e7de464115d-kube-api-access-nsqzl\") pod \"redhat-operators-9fqxz\" (UID: \"05e72d6a-37da-4550-96f3-0e7de464115d\") " pod="openshift-marketplace/redhat-operators-9fqxz" Dec 05 08:40:31 crc kubenswrapper[4997]: I1205 08:40:31.090107 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9fqxz" Dec 05 08:40:31 crc kubenswrapper[4997]: I1205 08:40:31.324600 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9fqxz"] Dec 05 08:40:32 crc kubenswrapper[4997]: I1205 08:40:32.265747 4997 generic.go:334] "Generic (PLEG): container finished" podID="05e72d6a-37da-4550-96f3-0e7de464115d" containerID="1bad53d248fd2bf0871d9282d47f3c9e5fada488fd0dfa1355ac8fe905b9525a" exitCode=0 Dec 05 08:40:32 crc kubenswrapper[4997]: I1205 08:40:32.265818 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fqxz" event={"ID":"05e72d6a-37da-4550-96f3-0e7de464115d","Type":"ContainerDied","Data":"1bad53d248fd2bf0871d9282d47f3c9e5fada488fd0dfa1355ac8fe905b9525a"} Dec 05 08:40:32 crc kubenswrapper[4997]: I1205 08:40:32.266109 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fqxz" event={"ID":"05e72d6a-37da-4550-96f3-0e7de464115d","Type":"ContainerStarted","Data":"bc1c18b1ec80f858bd127d9f46228f9ab63a6f89e9c9496ee7bf612df7e9ced7"} Dec 05 08:40:32 crc kubenswrapper[4997]: I1205 08:40:32.267855 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 08:40:33 crc kubenswrapper[4997]: I1205 08:40:33.275146 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fqxz" event={"ID":"05e72d6a-37da-4550-96f3-0e7de464115d","Type":"ContainerStarted","Data":"6a6db60bcc6a5d9b223f901192f785c59ceff4302d9b2817aeacea46bf5548fb"} Dec 05 08:40:34 crc kubenswrapper[4997]: I1205 08:40:34.283504 4997 generic.go:334] "Generic (PLEG): container finished" podID="05e72d6a-37da-4550-96f3-0e7de464115d" containerID="6a6db60bcc6a5d9b223f901192f785c59ceff4302d9b2817aeacea46bf5548fb" exitCode=0 Dec 05 08:40:34 crc kubenswrapper[4997]: I1205 08:40:34.283864 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fqxz" event={"ID":"05e72d6a-37da-4550-96f3-0e7de464115d","Type":"ContainerDied","Data":"6a6db60bcc6a5d9b223f901192f785c59ceff4302d9b2817aeacea46bf5548fb"} Dec 05 08:40:35 crc kubenswrapper[4997]: I1205 08:40:35.294059 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fqxz" event={"ID":"05e72d6a-37da-4550-96f3-0e7de464115d","Type":"ContainerStarted","Data":"c1905c137d1308671be588bcc8357af1091f5c9d79e17ea7f8e165f2a09c6e60"} Dec 05 08:40:35 crc kubenswrapper[4997]: I1205 08:40:35.314325 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9fqxz" podStartSLOduration=2.890978247 podStartE2EDuration="5.314307734s" podCreationTimestamp="2025-12-05 08:40:30 +0000 UTC" firstStartedPulling="2025-12-05 08:40:32.267515238 +0000 UTC m=+6332.796422499" lastFinishedPulling="2025-12-05 08:40:34.690844725 +0000 UTC m=+6335.219751986" observedRunningTime="2025-12-05 08:40:35.311437747 +0000 UTC m=+6335.840345018" watchObservedRunningTime="2025-12-05 08:40:35.314307734 +0000 UTC m=+6335.843214995" Dec 05 08:40:41 crc kubenswrapper[4997]: I1205 08:40:41.090335 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9fqxz" Dec 05 08:40:41 crc kubenswrapper[4997]: I1205 08:40:41.090927 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9fqxz" Dec 05 08:40:41 crc kubenswrapper[4997]: I1205 08:40:41.139180 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9fqxz" Dec 05 08:40:41 crc kubenswrapper[4997]: I1205 08:40:41.390713 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9fqxz" Dec 05 08:40:41 crc kubenswrapper[4997]: I1205 08:40:41.435712 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9fqxz"] Dec 05 08:40:43 crc kubenswrapper[4997]: I1205 08:40:43.364411 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9fqxz" podUID="05e72d6a-37da-4550-96f3-0e7de464115d" containerName="registry-server" containerID="cri-o://c1905c137d1308671be588bcc8357af1091f5c9d79e17ea7f8e165f2a09c6e60" gracePeriod=2 Dec 05 08:40:46 crc kubenswrapper[4997]: I1205 08:40:46.386473 4997 generic.go:334] "Generic (PLEG): container finished" podID="05e72d6a-37da-4550-96f3-0e7de464115d" containerID="c1905c137d1308671be588bcc8357af1091f5c9d79e17ea7f8e165f2a09c6e60" exitCode=0 Dec 05 08:40:46 crc kubenswrapper[4997]: I1205 08:40:46.386554 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fqxz" event={"ID":"05e72d6a-37da-4550-96f3-0e7de464115d","Type":"ContainerDied","Data":"c1905c137d1308671be588bcc8357af1091f5c9d79e17ea7f8e165f2a09c6e60"} Dec 05 08:40:46 crc kubenswrapper[4997]: I1205 08:40:46.475801 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9fqxz" Dec 05 08:40:46 crc kubenswrapper[4997]: I1205 08:40:46.602257 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsqzl\" (UniqueName: \"kubernetes.io/projected/05e72d6a-37da-4550-96f3-0e7de464115d-kube-api-access-nsqzl\") pod \"05e72d6a-37da-4550-96f3-0e7de464115d\" (UID: \"05e72d6a-37da-4550-96f3-0e7de464115d\") " Dec 05 08:40:46 crc kubenswrapper[4997]: I1205 08:40:46.602441 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05e72d6a-37da-4550-96f3-0e7de464115d-utilities\") pod \"05e72d6a-37da-4550-96f3-0e7de464115d\" (UID: \"05e72d6a-37da-4550-96f3-0e7de464115d\") " Dec 05 08:40:46 crc kubenswrapper[4997]: I1205 08:40:46.602497 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05e72d6a-37da-4550-96f3-0e7de464115d-catalog-content\") pod \"05e72d6a-37da-4550-96f3-0e7de464115d\" (UID: \"05e72d6a-37da-4550-96f3-0e7de464115d\") " Dec 05 08:40:46 crc kubenswrapper[4997]: I1205 08:40:46.603792 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05e72d6a-37da-4550-96f3-0e7de464115d-utilities" (OuterVolumeSpecName: "utilities") pod "05e72d6a-37da-4550-96f3-0e7de464115d" (UID: "05e72d6a-37da-4550-96f3-0e7de464115d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:40:46 crc kubenswrapper[4997]: I1205 08:40:46.609363 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05e72d6a-37da-4550-96f3-0e7de464115d-kube-api-access-nsqzl" (OuterVolumeSpecName: "kube-api-access-nsqzl") pod "05e72d6a-37da-4550-96f3-0e7de464115d" (UID: "05e72d6a-37da-4550-96f3-0e7de464115d"). InnerVolumeSpecName "kube-api-access-nsqzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:40:46 crc kubenswrapper[4997]: I1205 08:40:46.703257 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05e72d6a-37da-4550-96f3-0e7de464115d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05e72d6a-37da-4550-96f3-0e7de464115d" (UID: "05e72d6a-37da-4550-96f3-0e7de464115d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:40:46 crc kubenswrapper[4997]: I1205 08:40:46.704059 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05e72d6a-37da-4550-96f3-0e7de464115d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:46 crc kubenswrapper[4997]: I1205 08:40:46.704087 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05e72d6a-37da-4550-96f3-0e7de464115d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:46 crc kubenswrapper[4997]: I1205 08:40:46.704137 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsqzl\" (UniqueName: \"kubernetes.io/projected/05e72d6a-37da-4550-96f3-0e7de464115d-kube-api-access-nsqzl\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:47 crc kubenswrapper[4997]: I1205 08:40:47.396850 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fqxz" event={"ID":"05e72d6a-37da-4550-96f3-0e7de464115d","Type":"ContainerDied","Data":"bc1c18b1ec80f858bd127d9f46228f9ab63a6f89e9c9496ee7bf612df7e9ced7"} Dec 05 08:40:47 crc kubenswrapper[4997]: I1205 08:40:47.396906 4997 scope.go:117] "RemoveContainer" containerID="c1905c137d1308671be588bcc8357af1091f5c9d79e17ea7f8e165f2a09c6e60" Dec 05 08:40:47 crc kubenswrapper[4997]: I1205 08:40:47.396905 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9fqxz" Dec 05 08:40:47 crc kubenswrapper[4997]: I1205 08:40:47.420981 4997 scope.go:117] "RemoveContainer" containerID="6a6db60bcc6a5d9b223f901192f785c59ceff4302d9b2817aeacea46bf5548fb" Dec 05 08:40:47 crc kubenswrapper[4997]: I1205 08:40:47.439913 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9fqxz"] Dec 05 08:40:47 crc kubenswrapper[4997]: I1205 08:40:47.447060 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9fqxz"] Dec 05 08:40:47 crc kubenswrapper[4997]: I1205 08:40:47.459347 4997 scope.go:117] "RemoveContainer" containerID="1bad53d248fd2bf0871d9282d47f3c9e5fada488fd0dfa1355ac8fe905b9525a" Dec 05 08:40:47 crc kubenswrapper[4997]: I1205 08:40:47.759520 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05e72d6a-37da-4550-96f3-0e7de464115d" path="/var/lib/kubelet/pods/05e72d6a-37da-4550-96f3-0e7de464115d/volumes" Dec 05 08:42:19 crc kubenswrapper[4997]: I1205 08:42:19.769945 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:42:19 crc kubenswrapper[4997]: I1205 08:42:19.770492 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:42:49 crc kubenswrapper[4997]: I1205 08:42:49.770579 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:42:49 crc kubenswrapper[4997]: I1205 08:42:49.771220 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:43:19 crc kubenswrapper[4997]: I1205 08:43:19.770128 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:43:19 crc kubenswrapper[4997]: I1205 08:43:19.770704 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:43:19 crc kubenswrapper[4997]: I1205 08:43:19.770750 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 08:43:19 crc kubenswrapper[4997]: I1205 08:43:19.771452 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1ae3926813fc853136a11e68578c56580649609079883104b1edf72cf4d67701"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:43:19 crc kubenswrapper[4997]: I1205 08:43:19.771513 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://1ae3926813fc853136a11e68578c56580649609079883104b1edf72cf4d67701" gracePeriod=600 Dec 05 08:43:19 crc kubenswrapper[4997]: I1205 08:43:19.929136 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="1ae3926813fc853136a11e68578c56580649609079883104b1edf72cf4d67701" exitCode=0 Dec 05 08:43:19 crc kubenswrapper[4997]: I1205 08:43:19.929227 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"1ae3926813fc853136a11e68578c56580649609079883104b1edf72cf4d67701"} Dec 05 08:43:19 crc kubenswrapper[4997]: I1205 08:43:19.929517 4997 scope.go:117] "RemoveContainer" containerID="457ac8e68e06a5a7e531c4aa9ebb5fca1efe2f95d81d10690aa805368104334b" Dec 05 08:43:20 crc kubenswrapper[4997]: I1205 08:43:20.939045 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9"} Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.143489 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw"] Dec 05 08:45:00 crc kubenswrapper[4997]: E1205 08:45:00.144208 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05e72d6a-37da-4550-96f3-0e7de464115d" containerName="extract-utilities" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.144224 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="05e72d6a-37da-4550-96f3-0e7de464115d" containerName="extract-utilities" Dec 05 08:45:00 crc kubenswrapper[4997]: E1205 08:45:00.144239 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05e72d6a-37da-4550-96f3-0e7de464115d" containerName="extract-content" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.144245 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="05e72d6a-37da-4550-96f3-0e7de464115d" containerName="extract-content" Dec 05 08:45:00 crc kubenswrapper[4997]: E1205 08:45:00.144262 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05e72d6a-37da-4550-96f3-0e7de464115d" containerName="registry-server" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.144268 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="05e72d6a-37da-4550-96f3-0e7de464115d" containerName="registry-server" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.144403 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="05e72d6a-37da-4550-96f3-0e7de464115d" containerName="registry-server" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.145387 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.151343 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.151705 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.162562 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw"] Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.320063 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcsnt\" (UniqueName: \"kubernetes.io/projected/8bb99647-f591-4b3a-88ab-85ef7d051bbc-kube-api-access-hcsnt\") pod \"collect-profiles-29415405-pvsdw\" (UID: \"8bb99647-f591-4b3a-88ab-85ef7d051bbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.320178 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bb99647-f591-4b3a-88ab-85ef7d051bbc-config-volume\") pod \"collect-profiles-29415405-pvsdw\" (UID: \"8bb99647-f591-4b3a-88ab-85ef7d051bbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.320272 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bb99647-f591-4b3a-88ab-85ef7d051bbc-secret-volume\") pod \"collect-profiles-29415405-pvsdw\" (UID: \"8bb99647-f591-4b3a-88ab-85ef7d051bbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.421428 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bb99647-f591-4b3a-88ab-85ef7d051bbc-secret-volume\") pod \"collect-profiles-29415405-pvsdw\" (UID: \"8bb99647-f591-4b3a-88ab-85ef7d051bbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.421529 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcsnt\" (UniqueName: \"kubernetes.io/projected/8bb99647-f591-4b3a-88ab-85ef7d051bbc-kube-api-access-hcsnt\") pod \"collect-profiles-29415405-pvsdw\" (UID: \"8bb99647-f591-4b3a-88ab-85ef7d051bbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.421578 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bb99647-f591-4b3a-88ab-85ef7d051bbc-config-volume\") pod \"collect-profiles-29415405-pvsdw\" (UID: \"8bb99647-f591-4b3a-88ab-85ef7d051bbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.422540 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bb99647-f591-4b3a-88ab-85ef7d051bbc-config-volume\") pod \"collect-profiles-29415405-pvsdw\" (UID: \"8bb99647-f591-4b3a-88ab-85ef7d051bbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.428335 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bb99647-f591-4b3a-88ab-85ef7d051bbc-secret-volume\") pod \"collect-profiles-29415405-pvsdw\" (UID: \"8bb99647-f591-4b3a-88ab-85ef7d051bbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.447977 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcsnt\" (UniqueName: \"kubernetes.io/projected/8bb99647-f591-4b3a-88ab-85ef7d051bbc-kube-api-access-hcsnt\") pod \"collect-profiles-29415405-pvsdw\" (UID: \"8bb99647-f591-4b3a-88ab-85ef7d051bbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.472646 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw" Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.697694 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw"] Dec 05 08:45:00 crc kubenswrapper[4997]: I1205 08:45:00.759039 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw" event={"ID":"8bb99647-f591-4b3a-88ab-85ef7d051bbc","Type":"ContainerStarted","Data":"178a53e4b24a12181d9f97618f8c2c5a630a396a6118ec2b1a6ba1a7a5c73066"} Dec 05 08:45:01 crc kubenswrapper[4997]: I1205 08:45:01.772121 4997 generic.go:334] "Generic (PLEG): container finished" podID="8bb99647-f591-4b3a-88ab-85ef7d051bbc" containerID="b0567b7982a7fdafe0cf620caadc10fd9e07419fe4b71fa066dad17a6481181c" exitCode=0 Dec 05 08:45:01 crc kubenswrapper[4997]: I1205 08:45:01.772190 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw" event={"ID":"8bb99647-f591-4b3a-88ab-85ef7d051bbc","Type":"ContainerDied","Data":"b0567b7982a7fdafe0cf620caadc10fd9e07419fe4b71fa066dad17a6481181c"} Dec 05 08:45:03 crc kubenswrapper[4997]: I1205 08:45:03.061464 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw" Dec 05 08:45:03 crc kubenswrapper[4997]: I1205 08:45:03.260417 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcsnt\" (UniqueName: \"kubernetes.io/projected/8bb99647-f591-4b3a-88ab-85ef7d051bbc-kube-api-access-hcsnt\") pod \"8bb99647-f591-4b3a-88ab-85ef7d051bbc\" (UID: \"8bb99647-f591-4b3a-88ab-85ef7d051bbc\") " Dec 05 08:45:03 crc kubenswrapper[4997]: I1205 08:45:03.260784 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bb99647-f591-4b3a-88ab-85ef7d051bbc-config-volume\") pod \"8bb99647-f591-4b3a-88ab-85ef7d051bbc\" (UID: \"8bb99647-f591-4b3a-88ab-85ef7d051bbc\") " Dec 05 08:45:03 crc kubenswrapper[4997]: I1205 08:45:03.260922 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bb99647-f591-4b3a-88ab-85ef7d051bbc-secret-volume\") pod \"8bb99647-f591-4b3a-88ab-85ef7d051bbc\" (UID: \"8bb99647-f591-4b3a-88ab-85ef7d051bbc\") " Dec 05 08:45:03 crc kubenswrapper[4997]: I1205 08:45:03.261722 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bb99647-f591-4b3a-88ab-85ef7d051bbc-config-volume" (OuterVolumeSpecName: "config-volume") pod "8bb99647-f591-4b3a-88ab-85ef7d051bbc" (UID: "8bb99647-f591-4b3a-88ab-85ef7d051bbc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:45:03 crc kubenswrapper[4997]: I1205 08:45:03.262185 4997 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bb99647-f591-4b3a-88ab-85ef7d051bbc-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 08:45:03 crc kubenswrapper[4997]: I1205 08:45:03.266011 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bb99647-f591-4b3a-88ab-85ef7d051bbc-kube-api-access-hcsnt" (OuterVolumeSpecName: "kube-api-access-hcsnt") pod "8bb99647-f591-4b3a-88ab-85ef7d051bbc" (UID: "8bb99647-f591-4b3a-88ab-85ef7d051bbc"). InnerVolumeSpecName "kube-api-access-hcsnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:45:03 crc kubenswrapper[4997]: I1205 08:45:03.266042 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bb99647-f591-4b3a-88ab-85ef7d051bbc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8bb99647-f591-4b3a-88ab-85ef7d051bbc" (UID: "8bb99647-f591-4b3a-88ab-85ef7d051bbc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:45:03 crc kubenswrapper[4997]: I1205 08:45:03.363332 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcsnt\" (UniqueName: \"kubernetes.io/projected/8bb99647-f591-4b3a-88ab-85ef7d051bbc-kube-api-access-hcsnt\") on node \"crc\" DevicePath \"\"" Dec 05 08:45:03 crc kubenswrapper[4997]: I1205 08:45:03.363373 4997 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bb99647-f591-4b3a-88ab-85ef7d051bbc-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 08:45:03 crc kubenswrapper[4997]: I1205 08:45:03.788634 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw" event={"ID":"8bb99647-f591-4b3a-88ab-85ef7d051bbc","Type":"ContainerDied","Data":"178a53e4b24a12181d9f97618f8c2c5a630a396a6118ec2b1a6ba1a7a5c73066"} Dec 05 08:45:03 crc kubenswrapper[4997]: I1205 08:45:03.788677 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="178a53e4b24a12181d9f97618f8c2c5a630a396a6118ec2b1a6ba1a7a5c73066" Dec 05 08:45:03 crc kubenswrapper[4997]: I1205 08:45:03.789011 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw" Dec 05 08:45:04 crc kubenswrapper[4997]: I1205 08:45:04.138296 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx"] Dec 05 08:45:04 crc kubenswrapper[4997]: I1205 08:45:04.147994 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415360-l9xdx"] Dec 05 08:45:05 crc kubenswrapper[4997]: I1205 08:45:05.759978 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e6e989f-8977-4f2f-aff8-e9ca326b8fdf" path="/var/lib/kubelet/pods/3e6e989f-8977-4f2f-aff8-e9ca326b8fdf/volumes" Dec 05 08:45:12 crc kubenswrapper[4997]: I1205 08:45:12.876493 4997 scope.go:117] "RemoveContainer" containerID="5cd05b3261d8f22e530138087a11b10af4ceb446f5634a76f8a2dee49b01abe1" Dec 05 08:45:12 crc kubenswrapper[4997]: I1205 08:45:12.910928 4997 scope.go:117] "RemoveContainer" containerID="b127ee1cca7e4caf274e99d2f311479d7237ca9d71238ac8ea3153222de6e7c3" Dec 05 08:45:12 crc kubenswrapper[4997]: I1205 08:45:12.944050 4997 scope.go:117] "RemoveContainer" containerID="716aacb225bde0a355384d8290185e03a4c2a8da45cd47231f685f2e7dbf45f5" Dec 05 08:45:12 crc kubenswrapper[4997]: I1205 08:45:12.986490 4997 scope.go:117] "RemoveContainer" containerID="4f0dc440d0ee08775c9046a60c51145b81e29935ac0f67b442dcccae9eedd3f6" Dec 05 08:45:13 crc kubenswrapper[4997]: I1205 08:45:13.012700 4997 scope.go:117] "RemoveContainer" containerID="9b12990500a504f1a48a30db2d1509ad56e1e898de80ba25b1e0e7e5318293e7" Dec 05 08:45:13 crc kubenswrapper[4997]: I1205 08:45:13.041867 4997 scope.go:117] "RemoveContainer" containerID="3b1b5ab4c8c61e83c90d2b4ca5180354d81cf61588a5775a1fb750bd58bc29cc" Dec 05 08:45:13 crc kubenswrapper[4997]: I1205 08:45:13.075256 4997 scope.go:117] "RemoveContainer" containerID="5b77441652cf2b8510f7cfae5350a2349d44ca0a2b910f2c54bb3e8a35ea3b30" Dec 05 08:45:49 crc kubenswrapper[4997]: I1205 08:45:49.770354 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:45:49 crc kubenswrapper[4997]: I1205 08:45:49.770949 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:46:13 crc kubenswrapper[4997]: I1205 08:46:13.196720 4997 scope.go:117] "RemoveContainer" containerID="1460f98105d83e46357e1b6281611d947507cf84c1ea698c1cf876940cd5eaee" Dec 05 08:46:13 crc kubenswrapper[4997]: I1205 08:46:13.231562 4997 scope.go:117] "RemoveContainer" containerID="4943f8b18da4d09c7e6dc46140e2c94f29b236eb1a134e9343efecdc42c6179f" Dec 05 08:46:19 crc kubenswrapper[4997]: I1205 08:46:19.770373 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:46:19 crc kubenswrapper[4997]: I1205 08:46:19.771214 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:46:49 crc kubenswrapper[4997]: I1205 08:46:49.770344 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:46:49 crc kubenswrapper[4997]: I1205 08:46:49.771079 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:46:49 crc kubenswrapper[4997]: I1205 08:46:49.771145 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 08:46:49 crc kubenswrapper[4997]: I1205 08:46:49.772078 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:46:49 crc kubenswrapper[4997]: I1205 08:46:49.772181 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" gracePeriod=600 Dec 05 08:46:49 crc kubenswrapper[4997]: E1205 08:46:49.893292 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:46:50 crc kubenswrapper[4997]: I1205 08:46:50.679764 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" exitCode=0 Dec 05 08:46:50 crc kubenswrapper[4997]: I1205 08:46:50.680023 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9"} Dec 05 08:46:50 crc kubenswrapper[4997]: I1205 08:46:50.680219 4997 scope.go:117] "RemoveContainer" containerID="1ae3926813fc853136a11e68578c56580649609079883104b1edf72cf4d67701" Dec 05 08:46:50 crc kubenswrapper[4997]: I1205 08:46:50.680966 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:46:50 crc kubenswrapper[4997]: E1205 08:46:50.681335 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:47:02 crc kubenswrapper[4997]: I1205 08:47:02.750081 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:47:02 crc kubenswrapper[4997]: E1205 08:47:02.750945 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:47:15 crc kubenswrapper[4997]: I1205 08:47:15.749942 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:47:15 crc kubenswrapper[4997]: E1205 08:47:15.750783 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:47:27 crc kubenswrapper[4997]: I1205 08:47:27.098585 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zfwtp"] Dec 05 08:47:27 crc kubenswrapper[4997]: E1205 08:47:27.101734 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bb99647-f591-4b3a-88ab-85ef7d051bbc" containerName="collect-profiles" Dec 05 08:47:27 crc kubenswrapper[4997]: I1205 08:47:27.101785 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bb99647-f591-4b3a-88ab-85ef7d051bbc" containerName="collect-profiles" Dec 05 08:47:27 crc kubenswrapper[4997]: I1205 08:47:27.102073 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bb99647-f591-4b3a-88ab-85ef7d051bbc" containerName="collect-profiles" Dec 05 08:47:27 crc kubenswrapper[4997]: I1205 08:47:27.103909 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfwtp" Dec 05 08:47:27 crc kubenswrapper[4997]: I1205 08:47:27.113257 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zfwtp"] Dec 05 08:47:27 crc kubenswrapper[4997]: I1205 08:47:27.191491 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-utilities\") pod \"community-operators-zfwtp\" (UID: \"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0\") " pod="openshift-marketplace/community-operators-zfwtp" Dec 05 08:47:27 crc kubenswrapper[4997]: I1205 08:47:27.191574 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phl7k\" (UniqueName: \"kubernetes.io/projected/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-kube-api-access-phl7k\") pod \"community-operators-zfwtp\" (UID: \"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0\") " pod="openshift-marketplace/community-operators-zfwtp" Dec 05 08:47:27 crc kubenswrapper[4997]: I1205 08:47:27.191639 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-catalog-content\") pod \"community-operators-zfwtp\" (UID: \"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0\") " pod="openshift-marketplace/community-operators-zfwtp" Dec 05 08:47:27 crc kubenswrapper[4997]: I1205 08:47:27.292751 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-utilities\") pod \"community-operators-zfwtp\" (UID: \"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0\") " pod="openshift-marketplace/community-operators-zfwtp" Dec 05 08:47:27 crc kubenswrapper[4997]: I1205 08:47:27.292819 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phl7k\" (UniqueName: \"kubernetes.io/projected/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-kube-api-access-phl7k\") pod \"community-operators-zfwtp\" (UID: \"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0\") " pod="openshift-marketplace/community-operators-zfwtp" Dec 05 08:47:27 crc kubenswrapper[4997]: I1205 08:47:27.292871 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-catalog-content\") pod \"community-operators-zfwtp\" (UID: \"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0\") " pod="openshift-marketplace/community-operators-zfwtp" Dec 05 08:47:27 crc kubenswrapper[4997]: I1205 08:47:27.293223 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-utilities\") pod \"community-operators-zfwtp\" (UID: \"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0\") " pod="openshift-marketplace/community-operators-zfwtp" Dec 05 08:47:27 crc kubenswrapper[4997]: I1205 08:47:27.293387 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-catalog-content\") pod \"community-operators-zfwtp\" (UID: \"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0\") " pod="openshift-marketplace/community-operators-zfwtp" Dec 05 08:47:27 crc kubenswrapper[4997]: I1205 08:47:27.316201 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phl7k\" (UniqueName: \"kubernetes.io/projected/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-kube-api-access-phl7k\") pod \"community-operators-zfwtp\" (UID: \"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0\") " pod="openshift-marketplace/community-operators-zfwtp" Dec 05 08:47:27 crc kubenswrapper[4997]: I1205 08:47:27.436182 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfwtp" Dec 05 08:47:27 crc kubenswrapper[4997]: I1205 08:47:27.766503 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zfwtp"] Dec 05 08:47:28 crc kubenswrapper[4997]: I1205 08:47:28.025402 4997 generic.go:334] "Generic (PLEG): container finished" podID="a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0" containerID="3374ab57acc59c30f25604b00db065381a74687dd254852a0540cbd59515124c" exitCode=0 Dec 05 08:47:28 crc kubenswrapper[4997]: I1205 08:47:28.025474 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfwtp" event={"ID":"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0","Type":"ContainerDied","Data":"3374ab57acc59c30f25604b00db065381a74687dd254852a0540cbd59515124c"} Dec 05 08:47:28 crc kubenswrapper[4997]: I1205 08:47:28.025530 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfwtp" event={"ID":"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0","Type":"ContainerStarted","Data":"4714076d789a2706597cd39806ef6bd7d2c4a716639e83aa8dabc3dcf51b11a1"} Dec 05 08:47:28 crc kubenswrapper[4997]: I1205 08:47:28.028192 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 08:47:29 crc kubenswrapper[4997]: I1205 08:47:29.034634 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfwtp" event={"ID":"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0","Type":"ContainerStarted","Data":"4454c1bdba3cc117c6821120e73a0493d5a7a32a80cbf755a07e9f3294144e07"} Dec 05 08:47:29 crc kubenswrapper[4997]: I1205 08:47:29.754169 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:47:29 crc kubenswrapper[4997]: E1205 08:47:29.754454 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:47:30 crc kubenswrapper[4997]: I1205 08:47:30.045685 4997 generic.go:334] "Generic (PLEG): container finished" podID="a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0" containerID="4454c1bdba3cc117c6821120e73a0493d5a7a32a80cbf755a07e9f3294144e07" exitCode=0 Dec 05 08:47:30 crc kubenswrapper[4997]: I1205 08:47:30.045751 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfwtp" event={"ID":"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0","Type":"ContainerDied","Data":"4454c1bdba3cc117c6821120e73a0493d5a7a32a80cbf755a07e9f3294144e07"} Dec 05 08:47:31 crc kubenswrapper[4997]: I1205 08:47:31.057230 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfwtp" event={"ID":"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0","Type":"ContainerStarted","Data":"cce30da570821b03c1f7048d3fc204a1bc06d073c1f7804d6543489ad980893e"} Dec 05 08:47:31 crc kubenswrapper[4997]: I1205 08:47:31.081806 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zfwtp" podStartSLOduration=1.66976138 podStartE2EDuration="4.081783258s" podCreationTimestamp="2025-12-05 08:47:27 +0000 UTC" firstStartedPulling="2025-12-05 08:47:28.027937147 +0000 UTC m=+6748.556844408" lastFinishedPulling="2025-12-05 08:47:30.439959025 +0000 UTC m=+6750.968866286" observedRunningTime="2025-12-05 08:47:31.073738272 +0000 UTC m=+6751.602645553" watchObservedRunningTime="2025-12-05 08:47:31.081783258 +0000 UTC m=+6751.610690519" Dec 05 08:47:37 crc kubenswrapper[4997]: I1205 08:47:37.436801 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zfwtp" Dec 05 08:47:37 crc kubenswrapper[4997]: I1205 08:47:37.437369 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zfwtp" Dec 05 08:47:37 crc kubenswrapper[4997]: I1205 08:47:37.491135 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zfwtp" Dec 05 08:47:38 crc kubenswrapper[4997]: I1205 08:47:38.164999 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zfwtp" Dec 05 08:47:38 crc kubenswrapper[4997]: I1205 08:47:38.210599 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zfwtp"] Dec 05 08:47:40 crc kubenswrapper[4997]: I1205 08:47:40.129738 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zfwtp" podUID="a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0" containerName="registry-server" containerID="cri-o://cce30da570821b03c1f7048d3fc204a1bc06d073c1f7804d6543489ad980893e" gracePeriod=2 Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.143399 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfwtp" event={"ID":"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0","Type":"ContainerDied","Data":"cce30da570821b03c1f7048d3fc204a1bc06d073c1f7804d6543489ad980893e"} Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.143385 4997 generic.go:334] "Generic (PLEG): container finished" podID="a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0" containerID="cce30da570821b03c1f7048d3fc204a1bc06d073c1f7804d6543489ad980893e" exitCode=0 Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.497588 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wg794"] Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.503804 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wg794" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.506868 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wg794"] Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.540814 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2359db0c-741c-4c21-9d49-0aa4b1f85bff-utilities\") pod \"redhat-marketplace-wg794\" (UID: \"2359db0c-741c-4c21-9d49-0aa4b1f85bff\") " pod="openshift-marketplace/redhat-marketplace-wg794" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.540869 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2359db0c-741c-4c21-9d49-0aa4b1f85bff-catalog-content\") pod \"redhat-marketplace-wg794\" (UID: \"2359db0c-741c-4c21-9d49-0aa4b1f85bff\") " pod="openshift-marketplace/redhat-marketplace-wg794" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.540930 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22pb4\" (UniqueName: \"kubernetes.io/projected/2359db0c-741c-4c21-9d49-0aa4b1f85bff-kube-api-access-22pb4\") pod \"redhat-marketplace-wg794\" (UID: \"2359db0c-741c-4c21-9d49-0aa4b1f85bff\") " pod="openshift-marketplace/redhat-marketplace-wg794" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.641881 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2359db0c-741c-4c21-9d49-0aa4b1f85bff-utilities\") pod \"redhat-marketplace-wg794\" (UID: \"2359db0c-741c-4c21-9d49-0aa4b1f85bff\") " pod="openshift-marketplace/redhat-marketplace-wg794" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.641924 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2359db0c-741c-4c21-9d49-0aa4b1f85bff-catalog-content\") pod \"redhat-marketplace-wg794\" (UID: \"2359db0c-741c-4c21-9d49-0aa4b1f85bff\") " pod="openshift-marketplace/redhat-marketplace-wg794" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.641977 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22pb4\" (UniqueName: \"kubernetes.io/projected/2359db0c-741c-4c21-9d49-0aa4b1f85bff-kube-api-access-22pb4\") pod \"redhat-marketplace-wg794\" (UID: \"2359db0c-741c-4c21-9d49-0aa4b1f85bff\") " pod="openshift-marketplace/redhat-marketplace-wg794" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.642337 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2359db0c-741c-4c21-9d49-0aa4b1f85bff-utilities\") pod \"redhat-marketplace-wg794\" (UID: \"2359db0c-741c-4c21-9d49-0aa4b1f85bff\") " pod="openshift-marketplace/redhat-marketplace-wg794" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.642653 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfwtp" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.642799 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2359db0c-741c-4c21-9d49-0aa4b1f85bff-catalog-content\") pod \"redhat-marketplace-wg794\" (UID: \"2359db0c-741c-4c21-9d49-0aa4b1f85bff\") " pod="openshift-marketplace/redhat-marketplace-wg794" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.670680 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22pb4\" (UniqueName: \"kubernetes.io/projected/2359db0c-741c-4c21-9d49-0aa4b1f85bff-kube-api-access-22pb4\") pod \"redhat-marketplace-wg794\" (UID: \"2359db0c-741c-4c21-9d49-0aa4b1f85bff\") " pod="openshift-marketplace/redhat-marketplace-wg794" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.742715 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phl7k\" (UniqueName: \"kubernetes.io/projected/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-kube-api-access-phl7k\") pod \"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0\" (UID: \"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0\") " Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.742821 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-catalog-content\") pod \"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0\" (UID: \"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0\") " Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.742857 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-utilities\") pod \"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0\" (UID: \"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0\") " Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.744112 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-utilities" (OuterVolumeSpecName: "utilities") pod "a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0" (UID: "a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.753828 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-kube-api-access-phl7k" (OuterVolumeSpecName: "kube-api-access-phl7k") pod "a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0" (UID: "a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0"). InnerVolumeSpecName "kube-api-access-phl7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.754602 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:47:41 crc kubenswrapper[4997]: E1205 08:47:41.755031 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.802020 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0" (UID: "a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.831837 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wg794" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.845061 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phl7k\" (UniqueName: \"kubernetes.io/projected/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-kube-api-access-phl7k\") on node \"crc\" DevicePath \"\"" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.845098 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:47:41 crc kubenswrapper[4997]: I1205 08:47:41.845111 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:47:42 crc kubenswrapper[4997]: I1205 08:47:42.040178 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wg794"] Dec 05 08:47:42 crc kubenswrapper[4997]: I1205 08:47:42.160454 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wg794" event={"ID":"2359db0c-741c-4c21-9d49-0aa4b1f85bff","Type":"ContainerStarted","Data":"ce43165a5cc44f7a7a0bfa41b2b3f417bd51bb005c4fdd86e90eb9ae3edba1c6"} Dec 05 08:47:42 crc kubenswrapper[4997]: I1205 08:47:42.166394 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfwtp" event={"ID":"a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0","Type":"ContainerDied","Data":"4714076d789a2706597cd39806ef6bd7d2c4a716639e83aa8dabc3dcf51b11a1"} Dec 05 08:47:42 crc kubenswrapper[4997]: I1205 08:47:42.166440 4997 scope.go:117] "RemoveContainer" containerID="cce30da570821b03c1f7048d3fc204a1bc06d073c1f7804d6543489ad980893e" Dec 05 08:47:42 crc kubenswrapper[4997]: I1205 08:47:42.166487 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfwtp" Dec 05 08:47:42 crc kubenswrapper[4997]: I1205 08:47:42.185973 4997 scope.go:117] "RemoveContainer" containerID="4454c1bdba3cc117c6821120e73a0493d5a7a32a80cbf755a07e9f3294144e07" Dec 05 08:47:42 crc kubenswrapper[4997]: I1205 08:47:42.219147 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zfwtp"] Dec 05 08:47:42 crc kubenswrapper[4997]: I1205 08:47:42.226101 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zfwtp"] Dec 05 08:47:42 crc kubenswrapper[4997]: I1205 08:47:42.231843 4997 scope.go:117] "RemoveContainer" containerID="3374ab57acc59c30f25604b00db065381a74687dd254852a0540cbd59515124c" Dec 05 08:47:43 crc kubenswrapper[4997]: I1205 08:47:43.174956 4997 generic.go:334] "Generic (PLEG): container finished" podID="2359db0c-741c-4c21-9d49-0aa4b1f85bff" containerID="bc9973742bc1014d2b5eebd59d826a44ad9c64df4d5d570f8d797d54e27df82b" exitCode=0 Dec 05 08:47:43 crc kubenswrapper[4997]: I1205 08:47:43.175028 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wg794" event={"ID":"2359db0c-741c-4c21-9d49-0aa4b1f85bff","Type":"ContainerDied","Data":"bc9973742bc1014d2b5eebd59d826a44ad9c64df4d5d570f8d797d54e27df82b"} Dec 05 08:47:43 crc kubenswrapper[4997]: I1205 08:47:43.760215 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0" path="/var/lib/kubelet/pods/a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0/volumes" Dec 05 08:47:44 crc kubenswrapper[4997]: I1205 08:47:44.186571 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wg794" event={"ID":"2359db0c-741c-4c21-9d49-0aa4b1f85bff","Type":"ContainerStarted","Data":"9e7b1d8e562e72d2316ea419f736b65afbed51924d06b7f059615332c4b48763"} Dec 05 08:47:45 crc kubenswrapper[4997]: I1205 08:47:45.199996 4997 generic.go:334] "Generic (PLEG): container finished" podID="2359db0c-741c-4c21-9d49-0aa4b1f85bff" containerID="9e7b1d8e562e72d2316ea419f736b65afbed51924d06b7f059615332c4b48763" exitCode=0 Dec 05 08:47:45 crc kubenswrapper[4997]: I1205 08:47:45.200074 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wg794" event={"ID":"2359db0c-741c-4c21-9d49-0aa4b1f85bff","Type":"ContainerDied","Data":"9e7b1d8e562e72d2316ea419f736b65afbed51924d06b7f059615332c4b48763"} Dec 05 08:47:46 crc kubenswrapper[4997]: I1205 08:47:46.209123 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wg794" event={"ID":"2359db0c-741c-4c21-9d49-0aa4b1f85bff","Type":"ContainerStarted","Data":"7a53e7cf1073e83f0aea1b29bc5c89ab3575989be7342de63f28c619cf191583"} Dec 05 08:47:51 crc kubenswrapper[4997]: I1205 08:47:51.831972 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wg794" Dec 05 08:47:51 crc kubenswrapper[4997]: I1205 08:47:51.832039 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wg794" Dec 05 08:47:51 crc kubenswrapper[4997]: I1205 08:47:51.877645 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wg794" Dec 05 08:47:51 crc kubenswrapper[4997]: I1205 08:47:51.899185 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wg794" podStartSLOduration=8.4716718 podStartE2EDuration="10.899167393s" podCreationTimestamp="2025-12-05 08:47:41 +0000 UTC" firstStartedPulling="2025-12-05 08:47:43.178704281 +0000 UTC m=+6763.707611552" lastFinishedPulling="2025-12-05 08:47:45.606199874 +0000 UTC m=+6766.135107145" observedRunningTime="2025-12-05 08:47:46.230682901 +0000 UTC m=+6766.759590172" watchObservedRunningTime="2025-12-05 08:47:51.899167393 +0000 UTC m=+6772.428074654" Dec 05 08:47:52 crc kubenswrapper[4997]: I1205 08:47:52.340214 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wg794" Dec 05 08:47:52 crc kubenswrapper[4997]: I1205 08:47:52.387895 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wg794"] Dec 05 08:47:52 crc kubenswrapper[4997]: I1205 08:47:52.748995 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:47:52 crc kubenswrapper[4997]: E1205 08:47:52.749394 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:47:54 crc kubenswrapper[4997]: I1205 08:47:54.275081 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wg794" podUID="2359db0c-741c-4c21-9d49-0aa4b1f85bff" containerName="registry-server" containerID="cri-o://7a53e7cf1073e83f0aea1b29bc5c89ab3575989be7342de63f28c619cf191583" gracePeriod=2 Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.234692 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wg794" Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.285219 4997 generic.go:334] "Generic (PLEG): container finished" podID="2359db0c-741c-4c21-9d49-0aa4b1f85bff" containerID="7a53e7cf1073e83f0aea1b29bc5c89ab3575989be7342de63f28c619cf191583" exitCode=0 Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.285260 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wg794" event={"ID":"2359db0c-741c-4c21-9d49-0aa4b1f85bff","Type":"ContainerDied","Data":"7a53e7cf1073e83f0aea1b29bc5c89ab3575989be7342de63f28c619cf191583"} Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.285288 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wg794" event={"ID":"2359db0c-741c-4c21-9d49-0aa4b1f85bff","Type":"ContainerDied","Data":"ce43165a5cc44f7a7a0bfa41b2b3f417bd51bb005c4fdd86e90eb9ae3edba1c6"} Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.285305 4997 scope.go:117] "RemoveContainer" containerID="7a53e7cf1073e83f0aea1b29bc5c89ab3575989be7342de63f28c619cf191583" Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.285434 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wg794" Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.291549 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2359db0c-741c-4c21-9d49-0aa4b1f85bff-catalog-content\") pod \"2359db0c-741c-4c21-9d49-0aa4b1f85bff\" (UID: \"2359db0c-741c-4c21-9d49-0aa4b1f85bff\") " Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.291732 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22pb4\" (UniqueName: \"kubernetes.io/projected/2359db0c-741c-4c21-9d49-0aa4b1f85bff-kube-api-access-22pb4\") pod \"2359db0c-741c-4c21-9d49-0aa4b1f85bff\" (UID: \"2359db0c-741c-4c21-9d49-0aa4b1f85bff\") " Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.291788 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2359db0c-741c-4c21-9d49-0aa4b1f85bff-utilities\") pod \"2359db0c-741c-4c21-9d49-0aa4b1f85bff\" (UID: \"2359db0c-741c-4c21-9d49-0aa4b1f85bff\") " Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.296056 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2359db0c-741c-4c21-9d49-0aa4b1f85bff-utilities" (OuterVolumeSpecName: "utilities") pod "2359db0c-741c-4c21-9d49-0aa4b1f85bff" (UID: "2359db0c-741c-4c21-9d49-0aa4b1f85bff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.300096 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2359db0c-741c-4c21-9d49-0aa4b1f85bff-kube-api-access-22pb4" (OuterVolumeSpecName: "kube-api-access-22pb4") pod "2359db0c-741c-4c21-9d49-0aa4b1f85bff" (UID: "2359db0c-741c-4c21-9d49-0aa4b1f85bff"). InnerVolumeSpecName "kube-api-access-22pb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.314605 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2359db0c-741c-4c21-9d49-0aa4b1f85bff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2359db0c-741c-4c21-9d49-0aa4b1f85bff" (UID: "2359db0c-741c-4c21-9d49-0aa4b1f85bff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.317420 4997 scope.go:117] "RemoveContainer" containerID="9e7b1d8e562e72d2316ea419f736b65afbed51924d06b7f059615332c4b48763" Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.346959 4997 scope.go:117] "RemoveContainer" containerID="bc9973742bc1014d2b5eebd59d826a44ad9c64df4d5d570f8d797d54e27df82b" Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.381076 4997 scope.go:117] "RemoveContainer" containerID="7a53e7cf1073e83f0aea1b29bc5c89ab3575989be7342de63f28c619cf191583" Dec 05 08:47:55 crc kubenswrapper[4997]: E1205 08:47:55.381814 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a53e7cf1073e83f0aea1b29bc5c89ab3575989be7342de63f28c619cf191583\": container with ID starting with 7a53e7cf1073e83f0aea1b29bc5c89ab3575989be7342de63f28c619cf191583 not found: ID does not exist" containerID="7a53e7cf1073e83f0aea1b29bc5c89ab3575989be7342de63f28c619cf191583" Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.381876 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a53e7cf1073e83f0aea1b29bc5c89ab3575989be7342de63f28c619cf191583"} err="failed to get container status \"7a53e7cf1073e83f0aea1b29bc5c89ab3575989be7342de63f28c619cf191583\": rpc error: code = NotFound desc = could not find container \"7a53e7cf1073e83f0aea1b29bc5c89ab3575989be7342de63f28c619cf191583\": container with ID starting with 7a53e7cf1073e83f0aea1b29bc5c89ab3575989be7342de63f28c619cf191583 not found: ID does not exist" Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.381916 4997 scope.go:117] "RemoveContainer" containerID="9e7b1d8e562e72d2316ea419f736b65afbed51924d06b7f059615332c4b48763" Dec 05 08:47:55 crc kubenswrapper[4997]: E1205 08:47:55.382429 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e7b1d8e562e72d2316ea419f736b65afbed51924d06b7f059615332c4b48763\": container with ID starting with 9e7b1d8e562e72d2316ea419f736b65afbed51924d06b7f059615332c4b48763 not found: ID does not exist" containerID="9e7b1d8e562e72d2316ea419f736b65afbed51924d06b7f059615332c4b48763" Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.382474 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e7b1d8e562e72d2316ea419f736b65afbed51924d06b7f059615332c4b48763"} err="failed to get container status \"9e7b1d8e562e72d2316ea419f736b65afbed51924d06b7f059615332c4b48763\": rpc error: code = NotFound desc = could not find container \"9e7b1d8e562e72d2316ea419f736b65afbed51924d06b7f059615332c4b48763\": container with ID starting with 9e7b1d8e562e72d2316ea419f736b65afbed51924d06b7f059615332c4b48763 not found: ID does not exist" Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.382502 4997 scope.go:117] "RemoveContainer" containerID="bc9973742bc1014d2b5eebd59d826a44ad9c64df4d5d570f8d797d54e27df82b" Dec 05 08:47:55 crc kubenswrapper[4997]: E1205 08:47:55.382921 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc9973742bc1014d2b5eebd59d826a44ad9c64df4d5d570f8d797d54e27df82b\": container with ID starting with bc9973742bc1014d2b5eebd59d826a44ad9c64df4d5d570f8d797d54e27df82b not found: ID does not exist" containerID="bc9973742bc1014d2b5eebd59d826a44ad9c64df4d5d570f8d797d54e27df82b" Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.382950 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc9973742bc1014d2b5eebd59d826a44ad9c64df4d5d570f8d797d54e27df82b"} err="failed to get container status \"bc9973742bc1014d2b5eebd59d826a44ad9c64df4d5d570f8d797d54e27df82b\": rpc error: code = NotFound desc = could not find container \"bc9973742bc1014d2b5eebd59d826a44ad9c64df4d5d570f8d797d54e27df82b\": container with ID starting with bc9973742bc1014d2b5eebd59d826a44ad9c64df4d5d570f8d797d54e27df82b not found: ID does not exist" Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.393882 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2359db0c-741c-4c21-9d49-0aa4b1f85bff-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.393928 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22pb4\" (UniqueName: \"kubernetes.io/projected/2359db0c-741c-4c21-9d49-0aa4b1f85bff-kube-api-access-22pb4\") on node \"crc\" DevicePath \"\"" Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.393950 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2359db0c-741c-4c21-9d49-0aa4b1f85bff-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.614426 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wg794"] Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.622010 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wg794"] Dec 05 08:47:55 crc kubenswrapper[4997]: I1205 08:47:55.758900 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2359db0c-741c-4c21-9d49-0aa4b1f85bff" path="/var/lib/kubelet/pods/2359db0c-741c-4c21-9d49-0aa4b1f85bff/volumes" Dec 05 08:48:07 crc kubenswrapper[4997]: I1205 08:48:07.750305 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:48:07 crc kubenswrapper[4997]: E1205 08:48:07.750868 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:48:20 crc kubenswrapper[4997]: I1205 08:48:20.749330 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:48:20 crc kubenswrapper[4997]: E1205 08:48:20.750199 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.123553 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7d9mb"] Dec 05 08:48:29 crc kubenswrapper[4997]: E1205 08:48:29.124701 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0" containerName="extract-utilities" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.124728 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0" containerName="extract-utilities" Dec 05 08:48:29 crc kubenswrapper[4997]: E1205 08:48:29.124752 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0" containerName="extract-content" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.124768 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0" containerName="extract-content" Dec 05 08:48:29 crc kubenswrapper[4997]: E1205 08:48:29.124792 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2359db0c-741c-4c21-9d49-0aa4b1f85bff" containerName="extract-content" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.124803 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2359db0c-741c-4c21-9d49-0aa4b1f85bff" containerName="extract-content" Dec 05 08:48:29 crc kubenswrapper[4997]: E1205 08:48:29.124822 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2359db0c-741c-4c21-9d49-0aa4b1f85bff" containerName="extract-utilities" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.124834 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2359db0c-741c-4c21-9d49-0aa4b1f85bff" containerName="extract-utilities" Dec 05 08:48:29 crc kubenswrapper[4997]: E1205 08:48:29.124854 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0" containerName="registry-server" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.124862 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0" containerName="registry-server" Dec 05 08:48:29 crc kubenswrapper[4997]: E1205 08:48:29.124908 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2359db0c-741c-4c21-9d49-0aa4b1f85bff" containerName="registry-server" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.124919 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2359db0c-741c-4c21-9d49-0aa4b1f85bff" containerName="registry-server" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.125164 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a01bd2c7-7c0d-4e0e-a849-ccb61fbfc6f0" containerName="registry-server" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.125197 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2359db0c-741c-4c21-9d49-0aa4b1f85bff" containerName="registry-server" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.127072 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7d9mb" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.133300 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7d9mb"] Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.239636 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfddp\" (UniqueName: \"kubernetes.io/projected/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-kube-api-access-xfddp\") pod \"certified-operators-7d9mb\" (UID: \"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e\") " pod="openshift-marketplace/certified-operators-7d9mb" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.239747 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-utilities\") pod \"certified-operators-7d9mb\" (UID: \"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e\") " pod="openshift-marketplace/certified-operators-7d9mb" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.239878 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-catalog-content\") pod \"certified-operators-7d9mb\" (UID: \"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e\") " pod="openshift-marketplace/certified-operators-7d9mb" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.342024 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-utilities\") pod \"certified-operators-7d9mb\" (UID: \"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e\") " pod="openshift-marketplace/certified-operators-7d9mb" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.342136 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-catalog-content\") pod \"certified-operators-7d9mb\" (UID: \"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e\") " pod="openshift-marketplace/certified-operators-7d9mb" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.342179 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfddp\" (UniqueName: \"kubernetes.io/projected/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-kube-api-access-xfddp\") pod \"certified-operators-7d9mb\" (UID: \"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e\") " pod="openshift-marketplace/certified-operators-7d9mb" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.342716 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-utilities\") pod \"certified-operators-7d9mb\" (UID: \"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e\") " pod="openshift-marketplace/certified-operators-7d9mb" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.342870 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-catalog-content\") pod \"certified-operators-7d9mb\" (UID: \"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e\") " pod="openshift-marketplace/certified-operators-7d9mb" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.366095 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfddp\" (UniqueName: \"kubernetes.io/projected/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-kube-api-access-xfddp\") pod \"certified-operators-7d9mb\" (UID: \"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e\") " pod="openshift-marketplace/certified-operators-7d9mb" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.459416 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7d9mb" Dec 05 08:48:29 crc kubenswrapper[4997]: I1205 08:48:29.712350 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7d9mb"] Dec 05 08:48:30 crc kubenswrapper[4997]: I1205 08:48:30.554740 4997 generic.go:334] "Generic (PLEG): container finished" podID="ef7d64a6-69e2-4bf1-bf71-a19aca9af20e" containerID="ae009c91a57e4b5a5bbca8fda7eaed3090c3dbec9aa8c5ee3103c86159f34a45" exitCode=0 Dec 05 08:48:30 crc kubenswrapper[4997]: I1205 08:48:30.554804 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7d9mb" event={"ID":"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e","Type":"ContainerDied","Data":"ae009c91a57e4b5a5bbca8fda7eaed3090c3dbec9aa8c5ee3103c86159f34a45"} Dec 05 08:48:30 crc kubenswrapper[4997]: I1205 08:48:30.555031 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7d9mb" event={"ID":"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e","Type":"ContainerStarted","Data":"37108dc36ff4f16a405d9bf29fbf85ed908e1ea14cb6a8ebe6c47976e34c399f"} Dec 05 08:48:31 crc kubenswrapper[4997]: I1205 08:48:31.564262 4997 generic.go:334] "Generic (PLEG): container finished" podID="ef7d64a6-69e2-4bf1-bf71-a19aca9af20e" containerID="02e00b48287235601c0407246daa81b3ad738cd32a1527c3eb7bf3130c0e5497" exitCode=0 Dec 05 08:48:31 crc kubenswrapper[4997]: I1205 08:48:31.564305 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7d9mb" event={"ID":"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e","Type":"ContainerDied","Data":"02e00b48287235601c0407246daa81b3ad738cd32a1527c3eb7bf3130c0e5497"} Dec 05 08:48:32 crc kubenswrapper[4997]: I1205 08:48:32.749480 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:48:32 crc kubenswrapper[4997]: E1205 08:48:32.751295 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:48:33 crc kubenswrapper[4997]: I1205 08:48:33.580687 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7d9mb" event={"ID":"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e","Type":"ContainerStarted","Data":"62f2d24b22487df6f97797460824c6058f8ad738c902248f1a5a99d421fa36fa"} Dec 05 08:48:33 crc kubenswrapper[4997]: I1205 08:48:33.601873 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7d9mb" podStartSLOduration=2.664481432 podStartE2EDuration="4.601854689s" podCreationTimestamp="2025-12-05 08:48:29 +0000 UTC" firstStartedPulling="2025-12-05 08:48:30.556071366 +0000 UTC m=+6811.084978627" lastFinishedPulling="2025-12-05 08:48:32.493444623 +0000 UTC m=+6813.022351884" observedRunningTime="2025-12-05 08:48:33.597237515 +0000 UTC m=+6814.126144786" watchObservedRunningTime="2025-12-05 08:48:33.601854689 +0000 UTC m=+6814.130761950" Dec 05 08:48:39 crc kubenswrapper[4997]: I1205 08:48:39.460102 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7d9mb" Dec 05 08:48:39 crc kubenswrapper[4997]: I1205 08:48:39.460696 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7d9mb" Dec 05 08:48:39 crc kubenswrapper[4997]: I1205 08:48:39.507200 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7d9mb" Dec 05 08:48:39 crc kubenswrapper[4997]: I1205 08:48:39.675206 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7d9mb" Dec 05 08:48:39 crc kubenswrapper[4997]: I1205 08:48:39.736382 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7d9mb"] Dec 05 08:48:41 crc kubenswrapper[4997]: I1205 08:48:41.638340 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7d9mb" podUID="ef7d64a6-69e2-4bf1-bf71-a19aca9af20e" containerName="registry-server" containerID="cri-o://62f2d24b22487df6f97797460824c6058f8ad738c902248f1a5a99d421fa36fa" gracePeriod=2 Dec 05 08:48:42 crc kubenswrapper[4997]: I1205 08:48:42.646851 4997 generic.go:334] "Generic (PLEG): container finished" podID="ef7d64a6-69e2-4bf1-bf71-a19aca9af20e" containerID="62f2d24b22487df6f97797460824c6058f8ad738c902248f1a5a99d421fa36fa" exitCode=0 Dec 05 08:48:42 crc kubenswrapper[4997]: I1205 08:48:42.646934 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7d9mb" event={"ID":"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e","Type":"ContainerDied","Data":"62f2d24b22487df6f97797460824c6058f8ad738c902248f1a5a99d421fa36fa"} Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.264503 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7d9mb" Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.381263 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfddp\" (UniqueName: \"kubernetes.io/projected/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-kube-api-access-xfddp\") pod \"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e\" (UID: \"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e\") " Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.381355 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-utilities\") pod \"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e\" (UID: \"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e\") " Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.381431 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-catalog-content\") pod \"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e\" (UID: \"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e\") " Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.382759 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-utilities" (OuterVolumeSpecName: "utilities") pod "ef7d64a6-69e2-4bf1-bf71-a19aca9af20e" (UID: "ef7d64a6-69e2-4bf1-bf71-a19aca9af20e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.387025 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-kube-api-access-xfddp" (OuterVolumeSpecName: "kube-api-access-xfddp") pod "ef7d64a6-69e2-4bf1-bf71-a19aca9af20e" (UID: "ef7d64a6-69e2-4bf1-bf71-a19aca9af20e"). InnerVolumeSpecName "kube-api-access-xfddp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.448604 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ef7d64a6-69e2-4bf1-bf71-a19aca9af20e" (UID: "ef7d64a6-69e2-4bf1-bf71-a19aca9af20e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.482716 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.482754 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfddp\" (UniqueName: \"kubernetes.io/projected/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-kube-api-access-xfddp\") on node \"crc\" DevicePath \"\"" Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.482768 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.656330 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7d9mb" event={"ID":"ef7d64a6-69e2-4bf1-bf71-a19aca9af20e","Type":"ContainerDied","Data":"37108dc36ff4f16a405d9bf29fbf85ed908e1ea14cb6a8ebe6c47976e34c399f"} Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.656378 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7d9mb" Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.656383 4997 scope.go:117] "RemoveContainer" containerID="62f2d24b22487df6f97797460824c6058f8ad738c902248f1a5a99d421fa36fa" Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.677438 4997 scope.go:117] "RemoveContainer" containerID="02e00b48287235601c0407246daa81b3ad738cd32a1527c3eb7bf3130c0e5497" Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.689956 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7d9mb"] Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.701091 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7d9mb"] Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.717266 4997 scope.go:117] "RemoveContainer" containerID="ae009c91a57e4b5a5bbca8fda7eaed3090c3dbec9aa8c5ee3103c86159f34a45" Dec 05 08:48:43 crc kubenswrapper[4997]: I1205 08:48:43.759434 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef7d64a6-69e2-4bf1-bf71-a19aca9af20e" path="/var/lib/kubelet/pods/ef7d64a6-69e2-4bf1-bf71-a19aca9af20e/volumes" Dec 05 08:48:47 crc kubenswrapper[4997]: I1205 08:48:47.749439 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:48:47 crc kubenswrapper[4997]: E1205 08:48:47.751567 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:49:01 crc kubenswrapper[4997]: I1205 08:49:01.748937 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:49:01 crc kubenswrapper[4997]: E1205 08:49:01.749953 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.022289 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Dec 05 08:49:09 crc kubenswrapper[4997]: E1205 08:49:09.023428 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef7d64a6-69e2-4bf1-bf71-a19aca9af20e" containerName="extract-utilities" Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.023499 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef7d64a6-69e2-4bf1-bf71-a19aca9af20e" containerName="extract-utilities" Dec 05 08:49:09 crc kubenswrapper[4997]: E1205 08:49:09.023525 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef7d64a6-69e2-4bf1-bf71-a19aca9af20e" containerName="registry-server" Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.023537 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef7d64a6-69e2-4bf1-bf71-a19aca9af20e" containerName="registry-server" Dec 05 08:49:09 crc kubenswrapper[4997]: E1205 08:49:09.023577 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef7d64a6-69e2-4bf1-bf71-a19aca9af20e" containerName="extract-content" Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.023593 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef7d64a6-69e2-4bf1-bf71-a19aca9af20e" containerName="extract-content" Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.024077 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef7d64a6-69e2-4bf1-bf71-a19aca9af20e" containerName="registry-server" Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.025086 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.028493 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-qsjdm" Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.062778 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.078124 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5b900c62-9e5f-4417-9360-c4cb3361b7a6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5b900c62-9e5f-4417-9360-c4cb3361b7a6\") pod \"mariadb-copy-data\" (UID: \"ce49c282-2c15-4809-9fea-1df3e8750974\") " pod="openstack/mariadb-copy-data" Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.078244 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnxmb\" (UniqueName: \"kubernetes.io/projected/ce49c282-2c15-4809-9fea-1df3e8750974-kube-api-access-wnxmb\") pod \"mariadb-copy-data\" (UID: \"ce49c282-2c15-4809-9fea-1df3e8750974\") " pod="openstack/mariadb-copy-data" Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.179750 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5b900c62-9e5f-4417-9360-c4cb3361b7a6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5b900c62-9e5f-4417-9360-c4cb3361b7a6\") pod \"mariadb-copy-data\" (UID: \"ce49c282-2c15-4809-9fea-1df3e8750974\") " pod="openstack/mariadb-copy-data" Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.179847 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnxmb\" (UniqueName: \"kubernetes.io/projected/ce49c282-2c15-4809-9fea-1df3e8750974-kube-api-access-wnxmb\") pod \"mariadb-copy-data\" (UID: \"ce49c282-2c15-4809-9fea-1df3e8750974\") " pod="openstack/mariadb-copy-data" Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.182251 4997 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.182288 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5b900c62-9e5f-4417-9360-c4cb3361b7a6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5b900c62-9e5f-4417-9360-c4cb3361b7a6\") pod \"mariadb-copy-data\" (UID: \"ce49c282-2c15-4809-9fea-1df3e8750974\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6ea71fe3aff13e0948ecdbdc836ccc9c08f54b2f583bfd119921a19b0acda25d/globalmount\"" pod="openstack/mariadb-copy-data" Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.201448 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnxmb\" (UniqueName: \"kubernetes.io/projected/ce49c282-2c15-4809-9fea-1df3e8750974-kube-api-access-wnxmb\") pod \"mariadb-copy-data\" (UID: \"ce49c282-2c15-4809-9fea-1df3e8750974\") " pod="openstack/mariadb-copy-data" Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.209178 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5b900c62-9e5f-4417-9360-c4cb3361b7a6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5b900c62-9e5f-4417-9360-c4cb3361b7a6\") pod \"mariadb-copy-data\" (UID: \"ce49c282-2c15-4809-9fea-1df3e8750974\") " pod="openstack/mariadb-copy-data" Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.360068 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.657710 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.887082 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"ce49c282-2c15-4809-9fea-1df3e8750974","Type":"ContainerStarted","Data":"5b48de53aeeb2c625a8c80083e172a9f3b0c470cf58770302dc3610f17279bd5"} Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.887451 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"ce49c282-2c15-4809-9fea-1df3e8750974","Type":"ContainerStarted","Data":"b1f0bc0182fbf5f6d24f64a74e445296f60390f6c735a8e331d68fdf372748e8"} Dec 05 08:49:09 crc kubenswrapper[4997]: I1205 08:49:09.905050 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=2.905021571 podStartE2EDuration="2.905021571s" podCreationTimestamp="2025-12-05 08:49:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:49:09.898873947 +0000 UTC m=+6850.427781208" watchObservedRunningTime="2025-12-05 08:49:09.905021571 +0000 UTC m=+6850.433928842" Dec 05 08:49:13 crc kubenswrapper[4997]: I1205 08:49:13.106180 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Dec 05 08:49:13 crc kubenswrapper[4997]: I1205 08:49:13.107545 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 05 08:49:13 crc kubenswrapper[4997]: I1205 08:49:13.126047 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 05 08:49:13 crc kubenswrapper[4997]: I1205 08:49:13.148178 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl6rv\" (UniqueName: \"kubernetes.io/projected/cff6e9e5-5cd5-4017-b572-30c60a9671dc-kube-api-access-xl6rv\") pod \"mariadb-client\" (UID: \"cff6e9e5-5cd5-4017-b572-30c60a9671dc\") " pod="openstack/mariadb-client" Dec 05 08:49:13 crc kubenswrapper[4997]: I1205 08:49:13.249685 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl6rv\" (UniqueName: \"kubernetes.io/projected/cff6e9e5-5cd5-4017-b572-30c60a9671dc-kube-api-access-xl6rv\") pod \"mariadb-client\" (UID: \"cff6e9e5-5cd5-4017-b572-30c60a9671dc\") " pod="openstack/mariadb-client" Dec 05 08:49:13 crc kubenswrapper[4997]: I1205 08:49:13.268737 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl6rv\" (UniqueName: \"kubernetes.io/projected/cff6e9e5-5cd5-4017-b572-30c60a9671dc-kube-api-access-xl6rv\") pod \"mariadb-client\" (UID: \"cff6e9e5-5cd5-4017-b572-30c60a9671dc\") " pod="openstack/mariadb-client" Dec 05 08:49:13 crc kubenswrapper[4997]: I1205 08:49:13.450273 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 05 08:49:13 crc kubenswrapper[4997]: I1205 08:49:13.864827 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 05 08:49:13 crc kubenswrapper[4997]: W1205 08:49:13.867745 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcff6e9e5_5cd5_4017_b572_30c60a9671dc.slice/crio-42163471b98c125b9e1f1e2c486e23ef90709bf9277ba548f2473b2d050dacb3 WatchSource:0}: Error finding container 42163471b98c125b9e1f1e2c486e23ef90709bf9277ba548f2473b2d050dacb3: Status 404 returned error can't find the container with id 42163471b98c125b9e1f1e2c486e23ef90709bf9277ba548f2473b2d050dacb3 Dec 05 08:49:13 crc kubenswrapper[4997]: I1205 08:49:13.922415 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"cff6e9e5-5cd5-4017-b572-30c60a9671dc","Type":"ContainerStarted","Data":"42163471b98c125b9e1f1e2c486e23ef90709bf9277ba548f2473b2d050dacb3"} Dec 05 08:49:14 crc kubenswrapper[4997]: I1205 08:49:14.932227 4997 generic.go:334] "Generic (PLEG): container finished" podID="cff6e9e5-5cd5-4017-b572-30c60a9671dc" containerID="83bd574bee13eea7f1a9ff7aa3d514e06ed543014dee9ba67759d04ac052ee59" exitCode=0 Dec 05 08:49:14 crc kubenswrapper[4997]: I1205 08:49:14.932276 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"cff6e9e5-5cd5-4017-b572-30c60a9671dc","Type":"ContainerDied","Data":"83bd574bee13eea7f1a9ff7aa3d514e06ed543014dee9ba67759d04ac052ee59"} Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.268667 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.299538 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_cff6e9e5-5cd5-4017-b572-30c60a9671dc/mariadb-client/0.log" Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.305728 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xl6rv\" (UniqueName: \"kubernetes.io/projected/cff6e9e5-5cd5-4017-b572-30c60a9671dc-kube-api-access-xl6rv\") pod \"cff6e9e5-5cd5-4017-b572-30c60a9671dc\" (UID: \"cff6e9e5-5cd5-4017-b572-30c60a9671dc\") " Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.311159 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cff6e9e5-5cd5-4017-b572-30c60a9671dc-kube-api-access-xl6rv" (OuterVolumeSpecName: "kube-api-access-xl6rv") pod "cff6e9e5-5cd5-4017-b572-30c60a9671dc" (UID: "cff6e9e5-5cd5-4017-b572-30c60a9671dc"). InnerVolumeSpecName "kube-api-access-xl6rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.327343 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.334437 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.408132 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xl6rv\" (UniqueName: \"kubernetes.io/projected/cff6e9e5-5cd5-4017-b572-30c60a9671dc-kube-api-access-xl6rv\") on node \"crc\" DevicePath \"\"" Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.461394 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Dec 05 08:49:16 crc kubenswrapper[4997]: E1205 08:49:16.461902 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cff6e9e5-5cd5-4017-b572-30c60a9671dc" containerName="mariadb-client" Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.461923 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="cff6e9e5-5cd5-4017-b572-30c60a9671dc" containerName="mariadb-client" Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.462108 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="cff6e9e5-5cd5-4017-b572-30c60a9671dc" containerName="mariadb-client" Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.462778 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.468085 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.509466 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dftkz\" (UniqueName: \"kubernetes.io/projected/2fb16149-5405-4b52-899b-62beb83ca45a-kube-api-access-dftkz\") pod \"mariadb-client\" (UID: \"2fb16149-5405-4b52-899b-62beb83ca45a\") " pod="openstack/mariadb-client" Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.610673 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dftkz\" (UniqueName: \"kubernetes.io/projected/2fb16149-5405-4b52-899b-62beb83ca45a-kube-api-access-dftkz\") pod \"mariadb-client\" (UID: \"2fb16149-5405-4b52-899b-62beb83ca45a\") " pod="openstack/mariadb-client" Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.635791 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dftkz\" (UniqueName: \"kubernetes.io/projected/2fb16149-5405-4b52-899b-62beb83ca45a-kube-api-access-dftkz\") pod \"mariadb-client\" (UID: \"2fb16149-5405-4b52-899b-62beb83ca45a\") " pod="openstack/mariadb-client" Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.748872 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:49:16 crc kubenswrapper[4997]: E1205 08:49:16.749253 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.784788 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.954086 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42163471b98c125b9e1f1e2c486e23ef90709bf9277ba548f2473b2d050dacb3" Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.954149 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 05 08:49:16 crc kubenswrapper[4997]: I1205 08:49:16.972385 4997 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="cff6e9e5-5cd5-4017-b572-30c60a9671dc" podUID="2fb16149-5405-4b52-899b-62beb83ca45a" Dec 05 08:49:17 crc kubenswrapper[4997]: I1205 08:49:17.007089 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 05 08:49:17 crc kubenswrapper[4997]: W1205 08:49:17.010508 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fb16149_5405_4b52_899b_62beb83ca45a.slice/crio-5f7dbcd82f775a9e1e8a95d17659b6a415d5e113b0a010457140c047217c5062 WatchSource:0}: Error finding container 5f7dbcd82f775a9e1e8a95d17659b6a415d5e113b0a010457140c047217c5062: Status 404 returned error can't find the container with id 5f7dbcd82f775a9e1e8a95d17659b6a415d5e113b0a010457140c047217c5062 Dec 05 08:49:17 crc kubenswrapper[4997]: I1205 08:49:17.761743 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cff6e9e5-5cd5-4017-b572-30c60a9671dc" path="/var/lib/kubelet/pods/cff6e9e5-5cd5-4017-b572-30c60a9671dc/volumes" Dec 05 08:49:17 crc kubenswrapper[4997]: I1205 08:49:17.962772 4997 generic.go:334] "Generic (PLEG): container finished" podID="2fb16149-5405-4b52-899b-62beb83ca45a" containerID="3420f6d362c5b72a2d0e29f026755078694c4418dd38cb7b1b56db3cb45ddb4c" exitCode=0 Dec 05 08:49:17 crc kubenswrapper[4997]: I1205 08:49:17.962823 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"2fb16149-5405-4b52-899b-62beb83ca45a","Type":"ContainerDied","Data":"3420f6d362c5b72a2d0e29f026755078694c4418dd38cb7b1b56db3cb45ddb4c"} Dec 05 08:49:17 crc kubenswrapper[4997]: I1205 08:49:17.962850 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"2fb16149-5405-4b52-899b-62beb83ca45a","Type":"ContainerStarted","Data":"5f7dbcd82f775a9e1e8a95d17659b6a415d5e113b0a010457140c047217c5062"} Dec 05 08:49:19 crc kubenswrapper[4997]: I1205 08:49:19.305220 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 05 08:49:19 crc kubenswrapper[4997]: I1205 08:49:19.323154 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_2fb16149-5405-4b52-899b-62beb83ca45a/mariadb-client/0.log" Dec 05 08:49:19 crc kubenswrapper[4997]: I1205 08:49:19.347381 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Dec 05 08:49:19 crc kubenswrapper[4997]: I1205 08:49:19.354930 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dftkz\" (UniqueName: \"kubernetes.io/projected/2fb16149-5405-4b52-899b-62beb83ca45a-kube-api-access-dftkz\") pod \"2fb16149-5405-4b52-899b-62beb83ca45a\" (UID: \"2fb16149-5405-4b52-899b-62beb83ca45a\") " Dec 05 08:49:19 crc kubenswrapper[4997]: I1205 08:49:19.354952 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Dec 05 08:49:19 crc kubenswrapper[4997]: I1205 08:49:19.361300 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fb16149-5405-4b52-899b-62beb83ca45a-kube-api-access-dftkz" (OuterVolumeSpecName: "kube-api-access-dftkz") pod "2fb16149-5405-4b52-899b-62beb83ca45a" (UID: "2fb16149-5405-4b52-899b-62beb83ca45a"). InnerVolumeSpecName "kube-api-access-dftkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:49:19 crc kubenswrapper[4997]: I1205 08:49:19.456410 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dftkz\" (UniqueName: \"kubernetes.io/projected/2fb16149-5405-4b52-899b-62beb83ca45a-kube-api-access-dftkz\") on node \"crc\" DevicePath \"\"" Dec 05 08:49:19 crc kubenswrapper[4997]: I1205 08:49:19.759607 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fb16149-5405-4b52-899b-62beb83ca45a" path="/var/lib/kubelet/pods/2fb16149-5405-4b52-899b-62beb83ca45a/volumes" Dec 05 08:49:19 crc kubenswrapper[4997]: I1205 08:49:19.981241 4997 scope.go:117] "RemoveContainer" containerID="3420f6d362c5b72a2d0e29f026755078694c4418dd38cb7b1b56db3cb45ddb4c" Dec 05 08:49:19 crc kubenswrapper[4997]: I1205 08:49:19.981370 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 05 08:49:30 crc kubenswrapper[4997]: I1205 08:49:30.749420 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:49:30 crc kubenswrapper[4997]: E1205 08:49:30.750182 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:49:42 crc kubenswrapper[4997]: I1205 08:49:42.749550 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:49:42 crc kubenswrapper[4997]: E1205 08:49:42.750286 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:49:54 crc kubenswrapper[4997]: I1205 08:49:54.749249 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:49:54 crc kubenswrapper[4997]: E1205 08:49:54.750423 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.002871 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 08:49:57 crc kubenswrapper[4997]: E1205 08:49:57.003482 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fb16149-5405-4b52-899b-62beb83ca45a" containerName="mariadb-client" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.003538 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fb16149-5405-4b52-899b-62beb83ca45a" containerName="mariadb-client" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.003722 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fb16149-5405-4b52-899b-62beb83ca45a" containerName="mariadb-client" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.004588 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.006750 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.006841 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-fnmd4" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.008814 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.029841 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.040552 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.041976 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.058712 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.060000 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.065056 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.065942 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43590576-9b6a-436e-9022-d9af8d390cda-config\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.066005 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bf6dd0ec-b448-408c-8b0f-7165be31147f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.066054 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-37bebe28-2117-45cc-b358-3657f117316b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-37bebe28-2117-45cc-b358-3657f117316b\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.066094 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9d6j\" (UniqueName: \"kubernetes.io/projected/bf6dd0ec-b448-408c-8b0f-7165be31147f-kube-api-access-c9d6j\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.066112 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf6dd0ec-b448-408c-8b0f-7165be31147f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.066131 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a05a1996-a23f-4003-83f3-516666df3c39\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a05a1996-a23f-4003-83f3-516666df3c39\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.066151 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43590576-9b6a-436e-9022-d9af8d390cda-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.066201 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf6dd0ec-b448-408c-8b0f-7165be31147f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.066237 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/43590576-9b6a-436e-9022-d9af8d390cda-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.066270 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9br5l\" (UniqueName: \"kubernetes.io/projected/43590576-9b6a-436e-9022-d9af8d390cda-kube-api-access-9br5l\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.066284 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf6dd0ec-b448-408c-8b0f-7165be31147f-config\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.066301 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43590576-9b6a-436e-9022-d9af8d390cda-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.080986 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.168111 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43590576-9b6a-436e-9022-d9af8d390cda-config\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.168167 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bf6dd0ec-b448-408c-8b0f-7165be31147f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.168277 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-37bebe28-2117-45cc-b358-3657f117316b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-37bebe28-2117-45cc-b358-3657f117316b\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.168316 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9d6j\" (UniqueName: \"kubernetes.io/projected/bf6dd0ec-b448-408c-8b0f-7165be31147f-kube-api-access-c9d6j\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.168342 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf6dd0ec-b448-408c-8b0f-7165be31147f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.168372 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a05a1996-a23f-4003-83f3-516666df3c39\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a05a1996-a23f-4003-83f3-516666df3c39\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.168391 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43590576-9b6a-436e-9022-d9af8d390cda-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.168422 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf6dd0ec-b448-408c-8b0f-7165be31147f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.168446 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/43590576-9b6a-436e-9022-d9af8d390cda-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.168471 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9br5l\" (UniqueName: \"kubernetes.io/projected/43590576-9b6a-436e-9022-d9af8d390cda-kube-api-access-9br5l\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.168488 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf6dd0ec-b448-408c-8b0f-7165be31147f-config\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.168504 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43590576-9b6a-436e-9022-d9af8d390cda-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.168863 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bf6dd0ec-b448-408c-8b0f-7165be31147f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.169326 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/43590576-9b6a-436e-9022-d9af8d390cda-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.169378 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43590576-9b6a-436e-9022-d9af8d390cda-config\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.170193 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43590576-9b6a-436e-9022-d9af8d390cda-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.171243 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf6dd0ec-b448-408c-8b0f-7165be31147f-config\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.171726 4997 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.171761 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-37bebe28-2117-45cc-b358-3657f117316b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-37bebe28-2117-45cc-b358-3657f117316b\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/91ab0bdedbee57cfe7e8521161201fdfc7a3d1f19d0311ec4ea14fcf800d154e/globalmount\"" pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.172078 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf6dd0ec-b448-408c-8b0f-7165be31147f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.172600 4997 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.172648 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a05a1996-a23f-4003-83f3-516666df3c39\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a05a1996-a23f-4003-83f3-516666df3c39\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d6c9a51bf0a195735a15a50cc973fdc80e100830bd048b997c8807087ecf3a3c/globalmount\"" pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.178019 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43590576-9b6a-436e-9022-d9af8d390cda-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.184495 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf6dd0ec-b448-408c-8b0f-7165be31147f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.187730 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9br5l\" (UniqueName: \"kubernetes.io/projected/43590576-9b6a-436e-9022-d9af8d390cda-kube-api-access-9br5l\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.197487 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9d6j\" (UniqueName: \"kubernetes.io/projected/bf6dd0ec-b448-408c-8b0f-7165be31147f-kube-api-access-c9d6j\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.202774 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-37bebe28-2117-45cc-b358-3657f117316b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-37bebe28-2117-45cc-b358-3657f117316b\") pod \"ovsdbserver-nb-0\" (UID: \"bf6dd0ec-b448-408c-8b0f-7165be31147f\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.204385 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a05a1996-a23f-4003-83f3-516666df3c39\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a05a1996-a23f-4003-83f3-516666df3c39\") pod \"ovsdbserver-nb-1\" (UID: \"43590576-9b6a-436e-9022-d9af8d390cda\") " pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.222222 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.223521 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.225077 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.225704 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.226580 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-vwx2t" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.237221 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.248312 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.249721 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.265691 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.267148 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.270096 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c57e847-90d0-4769-9eb8-8c90080aee26-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.270185 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9c57e847-90d0-4769-9eb8-8c90080aee26-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.270223 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c57e847-90d0-4769-9eb8-8c90080aee26-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.270275 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lkv2\" (UniqueName: \"kubernetes.io/projected/9c57e847-90d0-4769-9eb8-8c90080aee26-kube-api-access-4lkv2\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.270311 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c57e847-90d0-4769-9eb8-8c90080aee26-config\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.270353 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5cd1ad53-4e17-43a3-b683-1a6044c0ae03\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5cd1ad53-4e17-43a3-b683-1a6044c0ae03\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.277862 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.309222 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.330785 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.359002 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.371963 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c57e847-90d0-4769-9eb8-8c90080aee26-config\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372040 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d406fbf0-f501-46d8-ba4d-dd2dbd646010\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d406fbf0-f501-46d8-ba4d-dd2dbd646010\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372072 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/def832ab-6d2d-474e-ae28-f1e8e868f07e-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372104 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5cd1ad53-4e17-43a3-b683-1a6044c0ae03\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5cd1ad53-4e17-43a3-b683-1a6044c0ae03\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372142 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/def832ab-6d2d-474e-ae28-f1e8e868f07e-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372177 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bd459cc-346a-4432-ae3d-4df900035da3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372214 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0660b4d6-650f-47a4-8f78-b6a4fba01954-config\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372237 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def832ab-6d2d-474e-ae28-f1e8e868f07e-config\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372260 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c57e847-90d0-4769-9eb8-8c90080aee26-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372301 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7bd459cc-346a-4432-ae3d-4df900035da3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372328 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96n7v\" (UniqueName: \"kubernetes.io/projected/7bd459cc-346a-4432-ae3d-4df900035da3-kube-api-access-96n7v\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372342 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0660b4d6-650f-47a4-8f78-b6a4fba01954-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372360 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/def832ab-6d2d-474e-ae28-f1e8e868f07e-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372387 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9c57e847-90d0-4769-9eb8-8c90080aee26-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372410 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0660b4d6-650f-47a4-8f78-b6a4fba01954-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372424 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7bd459cc-346a-4432-ae3d-4df900035da3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372444 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c57e847-90d0-4769-9eb8-8c90080aee26-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372461 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mww5w\" (UniqueName: \"kubernetes.io/projected/def832ab-6d2d-474e-ae28-f1e8e868f07e-kube-api-access-mww5w\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372486 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4jxc\" (UniqueName: \"kubernetes.io/projected/0660b4d6-650f-47a4-8f78-b6a4fba01954-kube-api-access-s4jxc\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372503 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b3f27c31-5cb5-434a-bc1e-1f641e724ab4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b3f27c31-5cb5-434a-bc1e-1f641e724ab4\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372523 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0660b4d6-650f-47a4-8f78-b6a4fba01954-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372544 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b113e294-eb8d-46a4-bbf3-0fe63b0b62f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b113e294-eb8d-46a4-bbf3-0fe63b0b62f0\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372561 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lkv2\" (UniqueName: \"kubernetes.io/projected/9c57e847-90d0-4769-9eb8-8c90080aee26-kube-api-access-4lkv2\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.372580 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bd459cc-346a-4432-ae3d-4df900035da3-config\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.373413 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c57e847-90d0-4769-9eb8-8c90080aee26-config\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.375927 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9c57e847-90d0-4769-9eb8-8c90080aee26-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.376055 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c57e847-90d0-4769-9eb8-8c90080aee26-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.377771 4997 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.377797 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5cd1ad53-4e17-43a3-b683-1a6044c0ae03\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5cd1ad53-4e17-43a3-b683-1a6044c0ae03\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9c40b98e924d3d29f4b29d47b29b0e1634c7c6f9cb247a1995af2a36589fcb4e/globalmount\"" pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.381534 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c57e847-90d0-4769-9eb8-8c90080aee26-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.393093 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lkv2\" (UniqueName: \"kubernetes.io/projected/9c57e847-90d0-4769-9eb8-8c90080aee26-kube-api-access-4lkv2\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.409955 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5cd1ad53-4e17-43a3-b683-1a6044c0ae03\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5cd1ad53-4e17-43a3-b683-1a6044c0ae03\") pod \"ovsdbserver-nb-2\" (UID: \"9c57e847-90d0-4769-9eb8-8c90080aee26\") " pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.474381 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bd459cc-346a-4432-ae3d-4df900035da3-config\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.474854 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d406fbf0-f501-46d8-ba4d-dd2dbd646010\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d406fbf0-f501-46d8-ba4d-dd2dbd646010\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.474883 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/def832ab-6d2d-474e-ae28-f1e8e868f07e-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.474913 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/def832ab-6d2d-474e-ae28-f1e8e868f07e-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.474937 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bd459cc-346a-4432-ae3d-4df900035da3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.474963 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0660b4d6-650f-47a4-8f78-b6a4fba01954-config\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.474979 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def832ab-6d2d-474e-ae28-f1e8e868f07e-config\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.475008 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7bd459cc-346a-4432-ae3d-4df900035da3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.475038 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96n7v\" (UniqueName: \"kubernetes.io/projected/7bd459cc-346a-4432-ae3d-4df900035da3-kube-api-access-96n7v\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.475054 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0660b4d6-650f-47a4-8f78-b6a4fba01954-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.475083 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/def832ab-6d2d-474e-ae28-f1e8e868f07e-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.475131 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0660b4d6-650f-47a4-8f78-b6a4fba01954-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.475152 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7bd459cc-346a-4432-ae3d-4df900035da3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.475176 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mww5w\" (UniqueName: \"kubernetes.io/projected/def832ab-6d2d-474e-ae28-f1e8e868f07e-kube-api-access-mww5w\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.475209 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4jxc\" (UniqueName: \"kubernetes.io/projected/0660b4d6-650f-47a4-8f78-b6a4fba01954-kube-api-access-s4jxc\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.475227 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b3f27c31-5cb5-434a-bc1e-1f641e724ab4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b3f27c31-5cb5-434a-bc1e-1f641e724ab4\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.475252 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0660b4d6-650f-47a4-8f78-b6a4fba01954-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.475267 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b113e294-eb8d-46a4-bbf3-0fe63b0b62f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b113e294-eb8d-46a4-bbf3-0fe63b0b62f0\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.475393 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bd459cc-346a-4432-ae3d-4df900035da3-config\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.476116 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/def832ab-6d2d-474e-ae28-f1e8e868f07e-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.477135 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7bd459cc-346a-4432-ae3d-4df900035da3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.478223 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0660b4d6-650f-47a4-8f78-b6a4fba01954-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.478471 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7bd459cc-346a-4432-ae3d-4df900035da3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.478478 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0660b4d6-650f-47a4-8f78-b6a4fba01954-config\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.478827 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0660b4d6-650f-47a4-8f78-b6a4fba01954-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.478853 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0660b4d6-650f-47a4-8f78-b6a4fba01954-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.481103 4997 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.481133 4997 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.481181 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b113e294-eb8d-46a4-bbf3-0fe63b0b62f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b113e294-eb8d-46a4-bbf3-0fe63b0b62f0\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2e7ae1e192bcaf3b486a78981c5e8f086be65af0e0d71528210d42f0454defe0/globalmount\"" pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.481143 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d406fbf0-f501-46d8-ba4d-dd2dbd646010\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d406fbf0-f501-46d8-ba4d-dd2dbd646010\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f72be5ddf0358ed5ba89a1c7d99a5f8d905c727860fbb81fe2d95ab97cbda1f3/globalmount\"" pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.482851 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bd459cc-346a-4432-ae3d-4df900035da3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.484325 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/def832ab-6d2d-474e-ae28-f1e8e868f07e-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.484519 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def832ab-6d2d-474e-ae28-f1e8e868f07e-config\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.487498 4997 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.487528 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b3f27c31-5cb5-434a-bc1e-1f641e724ab4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b3f27c31-5cb5-434a-bc1e-1f641e724ab4\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5fa19c7d58ce2498c7d1f5cfb65e063435ae3da3a286120cf9ca6e8008833b2a/globalmount\"" pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.493635 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/def832ab-6d2d-474e-ae28-f1e8e868f07e-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.495346 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mww5w\" (UniqueName: \"kubernetes.io/projected/def832ab-6d2d-474e-ae28-f1e8e868f07e-kube-api-access-mww5w\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.498018 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96n7v\" (UniqueName: \"kubernetes.io/projected/7bd459cc-346a-4432-ae3d-4df900035da3-kube-api-access-96n7v\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.498158 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4jxc\" (UniqueName: \"kubernetes.io/projected/0660b4d6-650f-47a4-8f78-b6a4fba01954-kube-api-access-s4jxc\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.527150 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b113e294-eb8d-46a4-bbf3-0fe63b0b62f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b113e294-eb8d-46a4-bbf3-0fe63b0b62f0\") pod \"ovsdbserver-sb-1\" (UID: \"def832ab-6d2d-474e-ae28-f1e8e868f07e\") " pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.535603 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d406fbf0-f501-46d8-ba4d-dd2dbd646010\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d406fbf0-f501-46d8-ba4d-dd2dbd646010\") pod \"ovsdbserver-sb-2\" (UID: \"0660b4d6-650f-47a4-8f78-b6a4fba01954\") " pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.537664 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b3f27c31-5cb5-434a-bc1e-1f641e724ab4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b3f27c31-5cb5-434a-bc1e-1f641e724ab4\") pod \"ovsdbserver-sb-0\" (UID: \"7bd459cc-346a-4432-ae3d-4df900035da3\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.569739 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.588856 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.607092 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.679240 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.877905 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 08:49:57 crc kubenswrapper[4997]: I1205 08:49:57.968402 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 05 08:49:57 crc kubenswrapper[4997]: W1205 08:49:57.978252 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43590576_9b6a_436e_9022_d9af8d390cda.slice/crio-ac49e4afdc61057e915017f6ce7e4bb9e656f268a3402e02bffba853e002fd53 WatchSource:0}: Error finding container ac49e4afdc61057e915017f6ce7e4bb9e656f268a3402e02bffba853e002fd53: Status 404 returned error can't find the container with id ac49e4afdc61057e915017f6ce7e4bb9e656f268a3402e02bffba853e002fd53 Dec 05 08:49:58 crc kubenswrapper[4997]: I1205 08:49:58.084218 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 05 08:49:58 crc kubenswrapper[4997]: W1205 08:49:58.091377 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddef832ab_6d2d_474e_ae28_f1e8e868f07e.slice/crio-18f351d9f3337cd782e40d6dfbc6ac24a09598834ecff147188966ac6571cf46 WatchSource:0}: Error finding container 18f351d9f3337cd782e40d6dfbc6ac24a09598834ecff147188966ac6571cf46: Status 404 returned error can't find the container with id 18f351d9f3337cd782e40d6dfbc6ac24a09598834ecff147188966ac6571cf46 Dec 05 08:49:58 crc kubenswrapper[4997]: I1205 08:49:58.159829 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 05 08:49:58 crc kubenswrapper[4997]: W1205 08:49:58.163250 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0660b4d6_650f_47a4_8f78_b6a4fba01954.slice/crio-0fa57ceb8e3132d8a5d030987f716b1f6511255dbd5e259d1bebc0c859799b55 WatchSource:0}: Error finding container 0fa57ceb8e3132d8a5d030987f716b1f6511255dbd5e259d1bebc0c859799b55: Status 404 returned error can't find the container with id 0fa57ceb8e3132d8a5d030987f716b1f6511255dbd5e259d1bebc0c859799b55 Dec 05 08:49:58 crc kubenswrapper[4997]: I1205 08:49:58.307737 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7bd459cc-346a-4432-ae3d-4df900035da3","Type":"ContainerStarted","Data":"b19226b2796758bd87f4f76010663bb3670ee4bb60faec60b65de41a257cae3a"} Dec 05 08:49:58 crc kubenswrapper[4997]: I1205 08:49:58.308824 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"0660b4d6-650f-47a4-8f78-b6a4fba01954","Type":"ContainerStarted","Data":"0fa57ceb8e3132d8a5d030987f716b1f6511255dbd5e259d1bebc0c859799b55"} Dec 05 08:49:58 crc kubenswrapper[4997]: I1205 08:49:58.309974 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"def832ab-6d2d-474e-ae28-f1e8e868f07e","Type":"ContainerStarted","Data":"18f351d9f3337cd782e40d6dfbc6ac24a09598834ecff147188966ac6571cf46"} Dec 05 08:49:58 crc kubenswrapper[4997]: I1205 08:49:58.311216 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"43590576-9b6a-436e-9022-d9af8d390cda","Type":"ContainerStarted","Data":"ac49e4afdc61057e915017f6ce7e4bb9e656f268a3402e02bffba853e002fd53"} Dec 05 08:49:58 crc kubenswrapper[4997]: W1205 08:49:58.435008 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf6dd0ec_b448_408c_8b0f_7165be31147f.slice/crio-feb6ec5e3a07b02e0e3bfd47372cc106fb45f65943edcfe1ccd7f538be5f1e87 WatchSource:0}: Error finding container feb6ec5e3a07b02e0e3bfd47372cc106fb45f65943edcfe1ccd7f538be5f1e87: Status 404 returned error can't find the container with id feb6ec5e3a07b02e0e3bfd47372cc106fb45f65943edcfe1ccd7f538be5f1e87 Dec 05 08:49:58 crc kubenswrapper[4997]: I1205 08:49:58.436988 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 08:49:59 crc kubenswrapper[4997]: I1205 08:49:59.155065 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 05 08:49:59 crc kubenswrapper[4997]: W1205 08:49:59.162158 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c57e847_90d0_4769_9eb8_8c90080aee26.slice/crio-860ea82944b3a08d2b7e48c6f5c402e6cc643bb68fc52a55b0af174351d0f81b WatchSource:0}: Error finding container 860ea82944b3a08d2b7e48c6f5c402e6cc643bb68fc52a55b0af174351d0f81b: Status 404 returned error can't find the container with id 860ea82944b3a08d2b7e48c6f5c402e6cc643bb68fc52a55b0af174351d0f81b Dec 05 08:49:59 crc kubenswrapper[4997]: I1205 08:49:59.321288 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"9c57e847-90d0-4769-9eb8-8c90080aee26","Type":"ContainerStarted","Data":"860ea82944b3a08d2b7e48c6f5c402e6cc643bb68fc52a55b0af174351d0f81b"} Dec 05 08:49:59 crc kubenswrapper[4997]: I1205 08:49:59.322751 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bf6dd0ec-b448-408c-8b0f-7165be31147f","Type":"ContainerStarted","Data":"feb6ec5e3a07b02e0e3bfd47372cc106fb45f65943edcfe1ccd7f538be5f1e87"} Dec 05 08:50:08 crc kubenswrapper[4997]: I1205 08:50:08.749533 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:50:08 crc kubenswrapper[4997]: E1205 08:50:08.750252 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:50:19 crc kubenswrapper[4997]: I1205 08:50:19.754172 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:50:19 crc kubenswrapper[4997]: E1205 08:50:19.754923 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:50:20 crc kubenswrapper[4997]: I1205 08:50:20.491493 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bf6dd0ec-b448-408c-8b0f-7165be31147f","Type":"ContainerStarted","Data":"a24b15bfaf6d57370469d530ee60fb3674dfdb78ccd4f74adb9f2e91adac896e"} Dec 05 08:50:20 crc kubenswrapper[4997]: I1205 08:50:20.491941 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bf6dd0ec-b448-408c-8b0f-7165be31147f","Type":"ContainerStarted","Data":"e2cfb1db29620c68580ff9dfc7c677f5cc820073a14a7636661fc41812384c47"} Dec 05 08:50:20 crc kubenswrapper[4997]: I1205 08:50:20.496294 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7bd459cc-346a-4432-ae3d-4df900035da3","Type":"ContainerStarted","Data":"8b0873f3a47f043b1d98bd4ab3111aab998147acdc77da4be7564da9b294e6b5"} Dec 05 08:50:20 crc kubenswrapper[4997]: I1205 08:50:20.496329 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7bd459cc-346a-4432-ae3d-4df900035da3","Type":"ContainerStarted","Data":"4d15647b0358f507f0b2eb998aba8326426cf4a18111a074cdd032a9cd038c62"} Dec 05 08:50:20 crc kubenswrapper[4997]: I1205 08:50:20.498820 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"0660b4d6-650f-47a4-8f78-b6a4fba01954","Type":"ContainerStarted","Data":"b5107e69c2a10f38e71414b73a50782f205b655ef2aad0cdbb90fe4cafe10a05"} Dec 05 08:50:20 crc kubenswrapper[4997]: I1205 08:50:20.498850 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"0660b4d6-650f-47a4-8f78-b6a4fba01954","Type":"ContainerStarted","Data":"7cd34cd32d100571ceae4a44e4b8722eba273af729cc1a1b9be212cb49df2e35"} Dec 05 08:50:20 crc kubenswrapper[4997]: I1205 08:50:20.501319 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"9c57e847-90d0-4769-9eb8-8c90080aee26","Type":"ContainerStarted","Data":"84ab8323729b9d9a598e8a51464197b8ff90ae11ea1683267fc56932c97a3758"} Dec 05 08:50:20 crc kubenswrapper[4997]: I1205 08:50:20.501352 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"9c57e847-90d0-4769-9eb8-8c90080aee26","Type":"ContainerStarted","Data":"68e852fd509993baefc16657129b2c10dabafa9b2b54a870a8f4fdf8c795922b"} Dec 05 08:50:20 crc kubenswrapper[4997]: I1205 08:50:20.502537 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"def832ab-6d2d-474e-ae28-f1e8e868f07e","Type":"ContainerStarted","Data":"9024b9ef8c56c2bcfb9e17c47a3319f7d6744d5655c08b19e7c09f974608a3a8"} Dec 05 08:50:20 crc kubenswrapper[4997]: I1205 08:50:20.502557 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"def832ab-6d2d-474e-ae28-f1e8e868f07e","Type":"ContainerStarted","Data":"7548776250b7b1248e6a9a5de8a425773f1d5c4e8d72cc0c3471127c04db4259"} Dec 05 08:50:20 crc kubenswrapper[4997]: I1205 08:50:20.519457 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.321984864 podStartE2EDuration="25.519433366s" podCreationTimestamp="2025-12-05 08:49:55 +0000 UTC" firstStartedPulling="2025-12-05 08:49:58.437835307 +0000 UTC m=+6898.966742568" lastFinishedPulling="2025-12-05 08:50:19.635283809 +0000 UTC m=+6920.164191070" observedRunningTime="2025-12-05 08:50:20.512009896 +0000 UTC m=+6921.040917177" watchObservedRunningTime="2025-12-05 08:50:20.519433366 +0000 UTC m=+6921.048340647" Dec 05 08:50:20 crc kubenswrapper[4997]: I1205 08:50:20.532637 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=5.052550201 podStartE2EDuration="25.53261114s" podCreationTimestamp="2025-12-05 08:49:55 +0000 UTC" firstStartedPulling="2025-12-05 08:49:59.164670134 +0000 UTC m=+6899.693577395" lastFinishedPulling="2025-12-05 08:50:19.644731073 +0000 UTC m=+6920.173638334" observedRunningTime="2025-12-05 08:50:20.529016804 +0000 UTC m=+6921.057924075" watchObservedRunningTime="2025-12-05 08:50:20.53261114 +0000 UTC m=+6921.061518401" Dec 05 08:50:20 crc kubenswrapper[4997]: I1205 08:50:20.558153 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=2.807519513 podStartE2EDuration="24.558130986s" podCreationTimestamp="2025-12-05 08:49:56 +0000 UTC" firstStartedPulling="2025-12-05 08:49:57.883987578 +0000 UTC m=+6898.412894839" lastFinishedPulling="2025-12-05 08:50:19.634599051 +0000 UTC m=+6920.163506312" observedRunningTime="2025-12-05 08:50:20.545274911 +0000 UTC m=+6921.074182172" watchObservedRunningTime="2025-12-05 08:50:20.558130986 +0000 UTC m=+6921.087038247" Dec 05 08:50:20 crc kubenswrapper[4997]: I1205 08:50:20.565545 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=3.024003825 podStartE2EDuration="24.565525746s" podCreationTimestamp="2025-12-05 08:49:56 +0000 UTC" firstStartedPulling="2025-12-05 08:49:58.093158822 +0000 UTC m=+6898.622066083" lastFinishedPulling="2025-12-05 08:50:19.634680743 +0000 UTC m=+6920.163588004" observedRunningTime="2025-12-05 08:50:20.563781618 +0000 UTC m=+6921.092688889" watchObservedRunningTime="2025-12-05 08:50:20.565525746 +0000 UTC m=+6921.094432997" Dec 05 08:50:20 crc kubenswrapper[4997]: I1205 08:50:20.586920 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=3.119625095 podStartE2EDuration="24.58689646s" podCreationTimestamp="2025-12-05 08:49:56 +0000 UTC" firstStartedPulling="2025-12-05 08:49:58.165961059 +0000 UTC m=+6898.694868320" lastFinishedPulling="2025-12-05 08:50:19.633232414 +0000 UTC m=+6920.162139685" observedRunningTime="2025-12-05 08:50:20.581553236 +0000 UTC m=+6921.110460517" watchObservedRunningTime="2025-12-05 08:50:20.58689646 +0000 UTC m=+6921.115803731" Dec 05 08:50:21 crc kubenswrapper[4997]: I1205 08:50:21.331553 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 05 08:50:21 crc kubenswrapper[4997]: I1205 08:50:21.515829 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"43590576-9b6a-436e-9022-d9af8d390cda","Type":"ContainerStarted","Data":"08cf167c6488704d745abcebbee3b7f0734ffbc78c4fb2bf01f2916ee919e8a7"} Dec 05 08:50:21 crc kubenswrapper[4997]: I1205 08:50:21.515933 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"43590576-9b6a-436e-9022-d9af8d390cda","Type":"ContainerStarted","Data":"f653ae2403bbd9496332c3d793925ca5e67fd1fd8063e442fb9c306a2850245a"} Dec 05 08:50:21 crc kubenswrapper[4997]: I1205 08:50:21.538464 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=3.922923855 podStartE2EDuration="26.538433658s" podCreationTimestamp="2025-12-05 08:49:55 +0000 UTC" firstStartedPulling="2025-12-05 08:49:57.980759949 +0000 UTC m=+6898.509667210" lastFinishedPulling="2025-12-05 08:50:20.596269752 +0000 UTC m=+6921.125177013" observedRunningTime="2025-12-05 08:50:21.53440202 +0000 UTC m=+6922.063309291" watchObservedRunningTime="2025-12-05 08:50:21.538433658 +0000 UTC m=+6922.067340919" Dec 05 08:50:21 crc kubenswrapper[4997]: I1205 08:50:21.571166 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 05 08:50:21 crc kubenswrapper[4997]: I1205 08:50:21.589975 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Dec 05 08:50:21 crc kubenswrapper[4997]: I1205 08:50:21.607890 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Dec 05 08:50:21 crc kubenswrapper[4997]: I1205 08:50:21.679981 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Dec 05 08:50:22 crc kubenswrapper[4997]: I1205 08:50:22.331791 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 05 08:50:22 crc kubenswrapper[4997]: I1205 08:50:22.359854 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Dec 05 08:50:22 crc kubenswrapper[4997]: I1205 08:50:22.571544 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 05 08:50:22 crc kubenswrapper[4997]: I1205 08:50:22.589784 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Dec 05 08:50:22 crc kubenswrapper[4997]: I1205 08:50:22.608295 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Dec 05 08:50:22 crc kubenswrapper[4997]: I1205 08:50:22.679990 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.359291 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.376422 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.401128 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.600425 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.610068 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.640515 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.658341 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.676548 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.750701 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.757283 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.808973 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.853798 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-658cfcb57-kzftr"] Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.856007 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658cfcb57-kzftr" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.859730 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.866828 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-658cfcb57-kzftr"] Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.867166 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.983512 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-ovsdbserver-nb\") pod \"dnsmasq-dns-658cfcb57-kzftr\" (UID: \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\") " pod="openstack/dnsmasq-dns-658cfcb57-kzftr" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.983761 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd7fk\" (UniqueName: \"kubernetes.io/projected/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-kube-api-access-dd7fk\") pod \"dnsmasq-dns-658cfcb57-kzftr\" (UID: \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\") " pod="openstack/dnsmasq-dns-658cfcb57-kzftr" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.983928 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-config\") pod \"dnsmasq-dns-658cfcb57-kzftr\" (UID: \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\") " pod="openstack/dnsmasq-dns-658cfcb57-kzftr" Dec 05 08:50:24 crc kubenswrapper[4997]: I1205 08:50:24.984080 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-dns-svc\") pod \"dnsmasq-dns-658cfcb57-kzftr\" (UID: \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\") " pod="openstack/dnsmasq-dns-658cfcb57-kzftr" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.043582 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-658cfcb57-kzftr"] Dec 05 08:50:25 crc kubenswrapper[4997]: E1205 08:50:25.044175 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-dd7fk ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-658cfcb57-kzftr" podUID="1dfd96ab-668d-4428-a1b9-781eb89a8eb1" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.067888 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-574bcff97-jb2rd"] Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.074974 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.081265 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.085483 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-config\") pod \"dnsmasq-dns-658cfcb57-kzftr\" (UID: \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\") " pod="openstack/dnsmasq-dns-658cfcb57-kzftr" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.085545 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-dns-svc\") pod \"dnsmasq-dns-658cfcb57-kzftr\" (UID: \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\") " pod="openstack/dnsmasq-dns-658cfcb57-kzftr" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.085607 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-574bcff97-jb2rd"] Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.085689 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-ovsdbserver-nb\") pod \"dnsmasq-dns-658cfcb57-kzftr\" (UID: \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\") " pod="openstack/dnsmasq-dns-658cfcb57-kzftr" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.085729 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd7fk\" (UniqueName: \"kubernetes.io/projected/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-kube-api-access-dd7fk\") pod \"dnsmasq-dns-658cfcb57-kzftr\" (UID: \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\") " pod="openstack/dnsmasq-dns-658cfcb57-kzftr" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.086607 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-config\") pod \"dnsmasq-dns-658cfcb57-kzftr\" (UID: \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\") " pod="openstack/dnsmasq-dns-658cfcb57-kzftr" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.086951 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-dns-svc\") pod \"dnsmasq-dns-658cfcb57-kzftr\" (UID: \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\") " pod="openstack/dnsmasq-dns-658cfcb57-kzftr" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.087165 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-ovsdbserver-nb\") pod \"dnsmasq-dns-658cfcb57-kzftr\" (UID: \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\") " pod="openstack/dnsmasq-dns-658cfcb57-kzftr" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.129771 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd7fk\" (UniqueName: \"kubernetes.io/projected/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-kube-api-access-dd7fk\") pod \"dnsmasq-dns-658cfcb57-kzftr\" (UID: \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\") " pod="openstack/dnsmasq-dns-658cfcb57-kzftr" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.188815 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-ovsdbserver-sb\") pod \"dnsmasq-dns-574bcff97-jb2rd\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.189171 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-config\") pod \"dnsmasq-dns-574bcff97-jb2rd\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.189208 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-ovsdbserver-nb\") pod \"dnsmasq-dns-574bcff97-jb2rd\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.189516 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-dns-svc\") pod \"dnsmasq-dns-574bcff97-jb2rd\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.189885 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpwd2\" (UniqueName: \"kubernetes.io/projected/c2241378-742b-4a1f-b21a-01213d3745e8-kube-api-access-qpwd2\") pod \"dnsmasq-dns-574bcff97-jb2rd\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.291487 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpwd2\" (UniqueName: \"kubernetes.io/projected/c2241378-742b-4a1f-b21a-01213d3745e8-kube-api-access-qpwd2\") pod \"dnsmasq-dns-574bcff97-jb2rd\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.291546 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-ovsdbserver-sb\") pod \"dnsmasq-dns-574bcff97-jb2rd\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.291574 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-config\") pod \"dnsmasq-dns-574bcff97-jb2rd\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.291604 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-ovsdbserver-nb\") pod \"dnsmasq-dns-574bcff97-jb2rd\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.291670 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-dns-svc\") pod \"dnsmasq-dns-574bcff97-jb2rd\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.292524 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-dns-svc\") pod \"dnsmasq-dns-574bcff97-jb2rd\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.292838 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-ovsdbserver-sb\") pod \"dnsmasq-dns-574bcff97-jb2rd\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.293154 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-config\") pod \"dnsmasq-dns-574bcff97-jb2rd\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.293685 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-ovsdbserver-nb\") pod \"dnsmasq-dns-574bcff97-jb2rd\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.337948 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpwd2\" (UniqueName: \"kubernetes.io/projected/c2241378-742b-4a1f-b21a-01213d3745e8-kube-api-access-qpwd2\") pod \"dnsmasq-dns-574bcff97-jb2rd\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.398359 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.566227 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658cfcb57-kzftr" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.584279 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658cfcb57-kzftr" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.613682 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.696714 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dd7fk\" (UniqueName: \"kubernetes.io/projected/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-kube-api-access-dd7fk\") pod \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\" (UID: \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\") " Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.696796 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-config\") pod \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\" (UID: \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\") " Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.696864 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-dns-svc\") pod \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\" (UID: \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\") " Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.696890 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-ovsdbserver-nb\") pod \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\" (UID: \"1dfd96ab-668d-4428-a1b9-781eb89a8eb1\") " Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.697357 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1dfd96ab-668d-4428-a1b9-781eb89a8eb1" (UID: "1dfd96ab-668d-4428-a1b9-781eb89a8eb1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.697661 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1dfd96ab-668d-4428-a1b9-781eb89a8eb1" (UID: "1dfd96ab-668d-4428-a1b9-781eb89a8eb1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.697791 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.697816 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.698645 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-config" (OuterVolumeSpecName: "config") pod "1dfd96ab-668d-4428-a1b9-781eb89a8eb1" (UID: "1dfd96ab-668d-4428-a1b9-781eb89a8eb1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.701044 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-kube-api-access-dd7fk" (OuterVolumeSpecName: "kube-api-access-dd7fk") pod "1dfd96ab-668d-4428-a1b9-781eb89a8eb1" (UID: "1dfd96ab-668d-4428-a1b9-781eb89a8eb1"). InnerVolumeSpecName "kube-api-access-dd7fk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.800918 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dd7fk\" (UniqueName: \"kubernetes.io/projected/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-kube-api-access-dd7fk\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.800956 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dfd96ab-668d-4428-a1b9-781eb89a8eb1-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:25 crc kubenswrapper[4997]: I1205 08:50:25.816663 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-574bcff97-jb2rd"] Dec 05 08:50:26 crc kubenswrapper[4997]: I1205 08:50:26.576059 4997 generic.go:334] "Generic (PLEG): container finished" podID="c2241378-742b-4a1f-b21a-01213d3745e8" containerID="3495ae35600724700c55e556d248037fcb36635e0759c13ed5d0041ad88980cb" exitCode=0 Dec 05 08:50:26 crc kubenswrapper[4997]: I1205 08:50:26.576106 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574bcff97-jb2rd" event={"ID":"c2241378-742b-4a1f-b21a-01213d3745e8","Type":"ContainerDied","Data":"3495ae35600724700c55e556d248037fcb36635e0759c13ed5d0041ad88980cb"} Dec 05 08:50:26 crc kubenswrapper[4997]: I1205 08:50:26.576430 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574bcff97-jb2rd" event={"ID":"c2241378-742b-4a1f-b21a-01213d3745e8","Type":"ContainerStarted","Data":"369c1fa922661a6aa1f3e7aeaff28e4d0c6d107e22b786a7d47e1cf21687fa25"} Dec 05 08:50:26 crc kubenswrapper[4997]: I1205 08:50:26.576489 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658cfcb57-kzftr" Dec 05 08:50:26 crc kubenswrapper[4997]: I1205 08:50:26.717101 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-658cfcb57-kzftr"] Dec 05 08:50:26 crc kubenswrapper[4997]: I1205 08:50:26.724056 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-658cfcb57-kzftr"] Dec 05 08:50:27 crc kubenswrapper[4997]: I1205 08:50:27.586601 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574bcff97-jb2rd" event={"ID":"c2241378-742b-4a1f-b21a-01213d3745e8","Type":"ContainerStarted","Data":"10881e6afa1e07aa4b4a1ce0e66000192dc87d8f59b7c1e12df58e18b4c282a7"} Dec 05 08:50:27 crc kubenswrapper[4997]: I1205 08:50:27.759287 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dfd96ab-668d-4428-a1b9-781eb89a8eb1" path="/var/lib/kubelet/pods/1dfd96ab-668d-4428-a1b9-781eb89a8eb1/volumes" Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.402719 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.404318 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.407611 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.419764 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.545915 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/650f03f7-065f-42d0-8dc4-9092bc28a2d4-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"650f03f7-065f-42d0-8dc4-9092bc28a2d4\") " pod="openstack/ovn-copy-data" Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.546192 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k77h\" (UniqueName: \"kubernetes.io/projected/650f03f7-065f-42d0-8dc4-9092bc28a2d4-kube-api-access-9k77h\") pod \"ovn-copy-data\" (UID: \"650f03f7-065f-42d0-8dc4-9092bc28a2d4\") " pod="openstack/ovn-copy-data" Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.546245 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e75b94d1-ea00-4a16-832d-09cf519ec342\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b94d1-ea00-4a16-832d-09cf519ec342\") pod \"ovn-copy-data\" (UID: \"650f03f7-065f-42d0-8dc4-9092bc28a2d4\") " pod="openstack/ovn-copy-data" Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.593104 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.618646 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-574bcff97-jb2rd" podStartSLOduration=3.6186071269999998 podStartE2EDuration="3.618607127s" podCreationTimestamp="2025-12-05 08:50:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:50:28.613728577 +0000 UTC m=+6929.142635848" watchObservedRunningTime="2025-12-05 08:50:28.618607127 +0000 UTC m=+6929.147514388" Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.647604 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/650f03f7-065f-42d0-8dc4-9092bc28a2d4-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"650f03f7-065f-42d0-8dc4-9092bc28a2d4\") " pod="openstack/ovn-copy-data" Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.647708 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k77h\" (UniqueName: \"kubernetes.io/projected/650f03f7-065f-42d0-8dc4-9092bc28a2d4-kube-api-access-9k77h\") pod \"ovn-copy-data\" (UID: \"650f03f7-065f-42d0-8dc4-9092bc28a2d4\") " pod="openstack/ovn-copy-data" Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.647781 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e75b94d1-ea00-4a16-832d-09cf519ec342\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b94d1-ea00-4a16-832d-09cf519ec342\") pod \"ovn-copy-data\" (UID: \"650f03f7-065f-42d0-8dc4-9092bc28a2d4\") " pod="openstack/ovn-copy-data" Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.651056 4997 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.651321 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e75b94d1-ea00-4a16-832d-09cf519ec342\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b94d1-ea00-4a16-832d-09cf519ec342\") pod \"ovn-copy-data\" (UID: \"650f03f7-065f-42d0-8dc4-9092bc28a2d4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/325b39fa8682e311a111c1e3f53e610f5672ae674a0ea3156f356eab93cd2f66/globalmount\"" pod="openstack/ovn-copy-data" Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.655122 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/650f03f7-065f-42d0-8dc4-9092bc28a2d4-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"650f03f7-065f-42d0-8dc4-9092bc28a2d4\") " pod="openstack/ovn-copy-data" Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.667098 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k77h\" (UniqueName: \"kubernetes.io/projected/650f03f7-065f-42d0-8dc4-9092bc28a2d4-kube-api-access-9k77h\") pod \"ovn-copy-data\" (UID: \"650f03f7-065f-42d0-8dc4-9092bc28a2d4\") " pod="openstack/ovn-copy-data" Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.679879 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e75b94d1-ea00-4a16-832d-09cf519ec342\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b94d1-ea00-4a16-832d-09cf519ec342\") pod \"ovn-copy-data\" (UID: \"650f03f7-065f-42d0-8dc4-9092bc28a2d4\") " pod="openstack/ovn-copy-data" Dec 05 08:50:28 crc kubenswrapper[4997]: I1205 08:50:28.752448 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 05 08:50:29 crc kubenswrapper[4997]: I1205 08:50:29.029208 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Dec 05 08:50:29 crc kubenswrapper[4997]: W1205 08:50:29.030944 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod650f03f7_065f_42d0_8dc4_9092bc28a2d4.slice/crio-22d57cff64dd19e8c21b3bf2cf4f1e80d3b172b7ae29908ea4264aed24e9d91d WatchSource:0}: Error finding container 22d57cff64dd19e8c21b3bf2cf4f1e80d3b172b7ae29908ea4264aed24e9d91d: Status 404 returned error can't find the container with id 22d57cff64dd19e8c21b3bf2cf4f1e80d3b172b7ae29908ea4264aed24e9d91d Dec 05 08:50:29 crc kubenswrapper[4997]: I1205 08:50:29.602261 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"650f03f7-065f-42d0-8dc4-9092bc28a2d4","Type":"ContainerStarted","Data":"22d57cff64dd19e8c21b3bf2cf4f1e80d3b172b7ae29908ea4264aed24e9d91d"} Dec 05 08:50:30 crc kubenswrapper[4997]: I1205 08:50:30.613724 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"650f03f7-065f-42d0-8dc4-9092bc28a2d4","Type":"ContainerStarted","Data":"2889232d7b711b570b629cbe31d13490494287816ec612e7302c2aaa68fab403"} Dec 05 08:50:30 crc kubenswrapper[4997]: I1205 08:50:30.635861 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=2.482589561 podStartE2EDuration="3.635836942s" podCreationTimestamp="2025-12-05 08:50:27 +0000 UTC" firstStartedPulling="2025-12-05 08:50:29.03313895 +0000 UTC m=+6929.562046211" lastFinishedPulling="2025-12-05 08:50:30.186386331 +0000 UTC m=+6930.715293592" observedRunningTime="2025-12-05 08:50:30.627913239 +0000 UTC m=+6931.156820510" watchObservedRunningTime="2025-12-05 08:50:30.635836942 +0000 UTC m=+6931.164744203" Dec 05 08:50:31 crc kubenswrapper[4997]: I1205 08:50:31.749189 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:50:31 crc kubenswrapper[4997]: E1205 08:50:31.749763 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:50:35 crc kubenswrapper[4997]: I1205 08:50:35.399820 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:50:35 crc kubenswrapper[4997]: I1205 08:50:35.452263 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-778d75ccf7-nv7g4"] Dec 05 08:50:35 crc kubenswrapper[4997]: I1205 08:50:35.452604 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" podUID="9acdab75-6b5a-4252-a746-03f7d3b5493e" containerName="dnsmasq-dns" containerID="cri-o://4a19c7414bb6f4f547d19d92cae5e409eff005da06fbeed721ba71d19c0f112a" gracePeriod=10 Dec 05 08:50:35 crc kubenswrapper[4997]: I1205 08:50:35.651299 4997 generic.go:334] "Generic (PLEG): container finished" podID="9acdab75-6b5a-4252-a746-03f7d3b5493e" containerID="4a19c7414bb6f4f547d19d92cae5e409eff005da06fbeed721ba71d19c0f112a" exitCode=0 Dec 05 08:50:35 crc kubenswrapper[4997]: I1205 08:50:35.651355 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" event={"ID":"9acdab75-6b5a-4252-a746-03f7d3b5493e","Type":"ContainerDied","Data":"4a19c7414bb6f4f547d19d92cae5e409eff005da06fbeed721ba71d19c0f112a"} Dec 05 08:50:36 crc kubenswrapper[4997]: I1205 08:50:36.024488 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" Dec 05 08:50:36 crc kubenswrapper[4997]: I1205 08:50:36.168828 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9acdab75-6b5a-4252-a746-03f7d3b5493e-config\") pod \"9acdab75-6b5a-4252-a746-03f7d3b5493e\" (UID: \"9acdab75-6b5a-4252-a746-03f7d3b5493e\") " Dec 05 08:50:36 crc kubenswrapper[4997]: I1205 08:50:36.168993 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rf95\" (UniqueName: \"kubernetes.io/projected/9acdab75-6b5a-4252-a746-03f7d3b5493e-kube-api-access-7rf95\") pod \"9acdab75-6b5a-4252-a746-03f7d3b5493e\" (UID: \"9acdab75-6b5a-4252-a746-03f7d3b5493e\") " Dec 05 08:50:36 crc kubenswrapper[4997]: I1205 08:50:36.169020 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9acdab75-6b5a-4252-a746-03f7d3b5493e-dns-svc\") pod \"9acdab75-6b5a-4252-a746-03f7d3b5493e\" (UID: \"9acdab75-6b5a-4252-a746-03f7d3b5493e\") " Dec 05 08:50:36 crc kubenswrapper[4997]: I1205 08:50:36.178824 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9acdab75-6b5a-4252-a746-03f7d3b5493e-kube-api-access-7rf95" (OuterVolumeSpecName: "kube-api-access-7rf95") pod "9acdab75-6b5a-4252-a746-03f7d3b5493e" (UID: "9acdab75-6b5a-4252-a746-03f7d3b5493e"). InnerVolumeSpecName "kube-api-access-7rf95". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:50:36 crc kubenswrapper[4997]: I1205 08:50:36.235566 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9acdab75-6b5a-4252-a746-03f7d3b5493e-config" (OuterVolumeSpecName: "config") pod "9acdab75-6b5a-4252-a746-03f7d3b5493e" (UID: "9acdab75-6b5a-4252-a746-03f7d3b5493e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:50:36 crc kubenswrapper[4997]: I1205 08:50:36.249367 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9acdab75-6b5a-4252-a746-03f7d3b5493e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9acdab75-6b5a-4252-a746-03f7d3b5493e" (UID: "9acdab75-6b5a-4252-a746-03f7d3b5493e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:50:36 crc kubenswrapper[4997]: I1205 08:50:36.270961 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rf95\" (UniqueName: \"kubernetes.io/projected/9acdab75-6b5a-4252-a746-03f7d3b5493e-kube-api-access-7rf95\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:36 crc kubenswrapper[4997]: I1205 08:50:36.271013 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9acdab75-6b5a-4252-a746-03f7d3b5493e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:36 crc kubenswrapper[4997]: I1205 08:50:36.271025 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9acdab75-6b5a-4252-a746-03f7d3b5493e-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:36 crc kubenswrapper[4997]: I1205 08:50:36.660672 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" event={"ID":"9acdab75-6b5a-4252-a746-03f7d3b5493e","Type":"ContainerDied","Data":"372bb81a7dcfd924f98e3d6332da8c67626d26660faec3b7882899235e820b7b"} Dec 05 08:50:36 crc kubenswrapper[4997]: I1205 08:50:36.661005 4997 scope.go:117] "RemoveContainer" containerID="4a19c7414bb6f4f547d19d92cae5e409eff005da06fbeed721ba71d19c0f112a" Dec 05 08:50:36 crc kubenswrapper[4997]: I1205 08:50:36.660714 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778d75ccf7-nv7g4" Dec 05 08:50:36 crc kubenswrapper[4997]: I1205 08:50:36.692005 4997 scope.go:117] "RemoveContainer" containerID="a7fc8e5261025fc95708fd530d4697762c97e37594e8098b3548d3b1244d6816" Dec 05 08:50:36 crc kubenswrapper[4997]: I1205 08:50:36.695716 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-778d75ccf7-nv7g4"] Dec 05 08:50:36 crc kubenswrapper[4997]: I1205 08:50:36.705808 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-778d75ccf7-nv7g4"] Dec 05 08:50:37 crc kubenswrapper[4997]: I1205 08:50:37.760429 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9acdab75-6b5a-4252-a746-03f7d3b5493e" path="/var/lib/kubelet/pods/9acdab75-6b5a-4252-a746-03f7d3b5493e/volumes" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.042368 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 05 08:50:38 crc kubenswrapper[4997]: E1205 08:50:38.042761 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9acdab75-6b5a-4252-a746-03f7d3b5493e" containerName="init" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.042781 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="9acdab75-6b5a-4252-a746-03f7d3b5493e" containerName="init" Dec 05 08:50:38 crc kubenswrapper[4997]: E1205 08:50:38.042794 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9acdab75-6b5a-4252-a746-03f7d3b5493e" containerName="dnsmasq-dns" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.042800 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="9acdab75-6b5a-4252-a746-03f7d3b5493e" containerName="dnsmasq-dns" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.042978 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="9acdab75-6b5a-4252-a746-03f7d3b5493e" containerName="dnsmasq-dns" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.043890 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 08:50:38 crc kubenswrapper[4997]: W1205 08:50:38.053825 4997 reflector.go:561] object-"openstack"/"ovnnorthd-scripts": failed to list *v1.ConfigMap: configmaps "ovnnorthd-scripts" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Dec 05 08:50:38 crc kubenswrapper[4997]: E1205 08:50:38.054270 4997 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"ovnnorthd-scripts\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"ovnnorthd-scripts\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.061292 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.061347 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-mt9gm" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.074781 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.102665 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a10d62a7-cf49-42bd-8d9b-72cba21becd8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a10d62a7-cf49-42bd-8d9b-72cba21becd8\") " pod="openstack/ovn-northd-0" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.102829 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2q8g\" (UniqueName: \"kubernetes.io/projected/a10d62a7-cf49-42bd-8d9b-72cba21becd8-kube-api-access-x2q8g\") pod \"ovn-northd-0\" (UID: \"a10d62a7-cf49-42bd-8d9b-72cba21becd8\") " pod="openstack/ovn-northd-0" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.102904 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a10d62a7-cf49-42bd-8d9b-72cba21becd8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a10d62a7-cf49-42bd-8d9b-72cba21becd8\") " pod="openstack/ovn-northd-0" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.102942 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a10d62a7-cf49-42bd-8d9b-72cba21becd8-scripts\") pod \"ovn-northd-0\" (UID: \"a10d62a7-cf49-42bd-8d9b-72cba21becd8\") " pod="openstack/ovn-northd-0" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.102965 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a10d62a7-cf49-42bd-8d9b-72cba21becd8-config\") pod \"ovn-northd-0\" (UID: \"a10d62a7-cf49-42bd-8d9b-72cba21becd8\") " pod="openstack/ovn-northd-0" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.204864 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a10d62a7-cf49-42bd-8d9b-72cba21becd8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a10d62a7-cf49-42bd-8d9b-72cba21becd8\") " pod="openstack/ovn-northd-0" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.204968 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2q8g\" (UniqueName: \"kubernetes.io/projected/a10d62a7-cf49-42bd-8d9b-72cba21becd8-kube-api-access-x2q8g\") pod \"ovn-northd-0\" (UID: \"a10d62a7-cf49-42bd-8d9b-72cba21becd8\") " pod="openstack/ovn-northd-0" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.205025 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a10d62a7-cf49-42bd-8d9b-72cba21becd8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a10d62a7-cf49-42bd-8d9b-72cba21becd8\") " pod="openstack/ovn-northd-0" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.205045 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a10d62a7-cf49-42bd-8d9b-72cba21becd8-scripts\") pod \"ovn-northd-0\" (UID: \"a10d62a7-cf49-42bd-8d9b-72cba21becd8\") " pod="openstack/ovn-northd-0" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.205065 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a10d62a7-cf49-42bd-8d9b-72cba21becd8-config\") pod \"ovn-northd-0\" (UID: \"a10d62a7-cf49-42bd-8d9b-72cba21becd8\") " pod="openstack/ovn-northd-0" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.205591 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a10d62a7-cf49-42bd-8d9b-72cba21becd8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a10d62a7-cf49-42bd-8d9b-72cba21becd8\") " pod="openstack/ovn-northd-0" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.205993 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a10d62a7-cf49-42bd-8d9b-72cba21becd8-config\") pod \"ovn-northd-0\" (UID: \"a10d62a7-cf49-42bd-8d9b-72cba21becd8\") " pod="openstack/ovn-northd-0" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.213472 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a10d62a7-cf49-42bd-8d9b-72cba21becd8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a10d62a7-cf49-42bd-8d9b-72cba21becd8\") " pod="openstack/ovn-northd-0" Dec 05 08:50:38 crc kubenswrapper[4997]: I1205 08:50:38.228042 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2q8g\" (UniqueName: \"kubernetes.io/projected/a10d62a7-cf49-42bd-8d9b-72cba21becd8-kube-api-access-x2q8g\") pod \"ovn-northd-0\" (UID: \"a10d62a7-cf49-42bd-8d9b-72cba21becd8\") " pod="openstack/ovn-northd-0" Dec 05 08:50:39 crc kubenswrapper[4997]: I1205 08:50:39.080109 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 05 08:50:39 crc kubenswrapper[4997]: I1205 08:50:39.086282 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a10d62a7-cf49-42bd-8d9b-72cba21becd8-scripts\") pod \"ovn-northd-0\" (UID: \"a10d62a7-cf49-42bd-8d9b-72cba21becd8\") " pod="openstack/ovn-northd-0" Dec 05 08:50:39 crc kubenswrapper[4997]: I1205 08:50:39.270098 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 08:50:39 crc kubenswrapper[4997]: I1205 08:50:39.719738 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 08:50:39 crc kubenswrapper[4997]: W1205 08:50:39.726176 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda10d62a7_cf49_42bd_8d9b_72cba21becd8.slice/crio-ddb200a81781312392cd8fcf00b59801386bb09406164705ec9c2f117f0cfd05 WatchSource:0}: Error finding container ddb200a81781312392cd8fcf00b59801386bb09406164705ec9c2f117f0cfd05: Status 404 returned error can't find the container with id ddb200a81781312392cd8fcf00b59801386bb09406164705ec9c2f117f0cfd05 Dec 05 08:50:40 crc kubenswrapper[4997]: I1205 08:50:40.720902 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a10d62a7-cf49-42bd-8d9b-72cba21becd8","Type":"ContainerStarted","Data":"ddb200a81781312392cd8fcf00b59801386bb09406164705ec9c2f117f0cfd05"} Dec 05 08:50:41 crc kubenswrapper[4997]: I1205 08:50:41.729399 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a10d62a7-cf49-42bd-8d9b-72cba21becd8","Type":"ContainerStarted","Data":"93d6be91fc3d20a84abbe7e684b92ce990cf8787d6393c8f347a0e41cc666e7c"} Dec 05 08:50:41 crc kubenswrapper[4997]: I1205 08:50:41.729731 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a10d62a7-cf49-42bd-8d9b-72cba21becd8","Type":"ContainerStarted","Data":"920dffbbc3d9beedeeb89d575716251a400460fccb2602e4d850f403a16a46fb"} Dec 05 08:50:41 crc kubenswrapper[4997]: I1205 08:50:41.730574 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 05 08:50:41 crc kubenswrapper[4997]: I1205 08:50:41.752975 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.7226848439999998 podStartE2EDuration="3.752955308s" podCreationTimestamp="2025-12-05 08:50:38 +0000 UTC" firstStartedPulling="2025-12-05 08:50:39.728177561 +0000 UTC m=+6940.257084822" lastFinishedPulling="2025-12-05 08:50:40.758448025 +0000 UTC m=+6941.287355286" observedRunningTime="2025-12-05 08:50:41.746601076 +0000 UTC m=+6942.275508357" watchObservedRunningTime="2025-12-05 08:50:41.752955308 +0000 UTC m=+6942.281862569" Dec 05 08:50:42 crc kubenswrapper[4997]: I1205 08:50:42.749082 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:50:42 crc kubenswrapper[4997]: E1205 08:50:42.749771 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.408690 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-pmf4c"] Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.410489 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-pmf4c" Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.418274 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-pmf4c"] Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.488864 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-e2cd-account-create-update-c54l4"] Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.490053 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e2cd-account-create-update-c54l4" Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.496551 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e2cd-account-create-update-c54l4"] Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.514701 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.539406 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54fc3d55-a800-4f20-ad4b-eb48952d3660-operator-scripts\") pod \"keystone-db-create-pmf4c\" (UID: \"54fc3d55-a800-4f20-ad4b-eb48952d3660\") " pod="openstack/keystone-db-create-pmf4c" Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.539559 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxkzg\" (UniqueName: \"kubernetes.io/projected/54fc3d55-a800-4f20-ad4b-eb48952d3660-kube-api-access-gxkzg\") pod \"keystone-db-create-pmf4c\" (UID: \"54fc3d55-a800-4f20-ad4b-eb48952d3660\") " pod="openstack/keystone-db-create-pmf4c" Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.640864 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54fc3d55-a800-4f20-ad4b-eb48952d3660-operator-scripts\") pod \"keystone-db-create-pmf4c\" (UID: \"54fc3d55-a800-4f20-ad4b-eb48952d3660\") " pod="openstack/keystone-db-create-pmf4c" Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.640938 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxkzg\" (UniqueName: \"kubernetes.io/projected/54fc3d55-a800-4f20-ad4b-eb48952d3660-kube-api-access-gxkzg\") pod \"keystone-db-create-pmf4c\" (UID: \"54fc3d55-a800-4f20-ad4b-eb48952d3660\") " pod="openstack/keystone-db-create-pmf4c" Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.640982 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f94c087-4666-45bd-92b5-8d47209a0bc6-operator-scripts\") pod \"keystone-e2cd-account-create-update-c54l4\" (UID: \"8f94c087-4666-45bd-92b5-8d47209a0bc6\") " pod="openstack/keystone-e2cd-account-create-update-c54l4" Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.641026 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvxp7\" (UniqueName: \"kubernetes.io/projected/8f94c087-4666-45bd-92b5-8d47209a0bc6-kube-api-access-kvxp7\") pod \"keystone-e2cd-account-create-update-c54l4\" (UID: \"8f94c087-4666-45bd-92b5-8d47209a0bc6\") " pod="openstack/keystone-e2cd-account-create-update-c54l4" Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.641706 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54fc3d55-a800-4f20-ad4b-eb48952d3660-operator-scripts\") pod \"keystone-db-create-pmf4c\" (UID: \"54fc3d55-a800-4f20-ad4b-eb48952d3660\") " pod="openstack/keystone-db-create-pmf4c" Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.659118 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxkzg\" (UniqueName: \"kubernetes.io/projected/54fc3d55-a800-4f20-ad4b-eb48952d3660-kube-api-access-gxkzg\") pod \"keystone-db-create-pmf4c\" (UID: \"54fc3d55-a800-4f20-ad4b-eb48952d3660\") " pod="openstack/keystone-db-create-pmf4c" Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.734875 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-pmf4c" Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.742762 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvxp7\" (UniqueName: \"kubernetes.io/projected/8f94c087-4666-45bd-92b5-8d47209a0bc6-kube-api-access-kvxp7\") pod \"keystone-e2cd-account-create-update-c54l4\" (UID: \"8f94c087-4666-45bd-92b5-8d47209a0bc6\") " pod="openstack/keystone-e2cd-account-create-update-c54l4" Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.742918 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f94c087-4666-45bd-92b5-8d47209a0bc6-operator-scripts\") pod \"keystone-e2cd-account-create-update-c54l4\" (UID: \"8f94c087-4666-45bd-92b5-8d47209a0bc6\") " pod="openstack/keystone-e2cd-account-create-update-c54l4" Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.743687 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f94c087-4666-45bd-92b5-8d47209a0bc6-operator-scripts\") pod \"keystone-e2cd-account-create-update-c54l4\" (UID: \"8f94c087-4666-45bd-92b5-8d47209a0bc6\") " pod="openstack/keystone-e2cd-account-create-update-c54l4" Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.767425 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvxp7\" (UniqueName: \"kubernetes.io/projected/8f94c087-4666-45bd-92b5-8d47209a0bc6-kube-api-access-kvxp7\") pod \"keystone-e2cd-account-create-update-c54l4\" (UID: \"8f94c087-4666-45bd-92b5-8d47209a0bc6\") " pod="openstack/keystone-e2cd-account-create-update-c54l4" Dec 05 08:50:45 crc kubenswrapper[4997]: I1205 08:50:45.804522 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e2cd-account-create-update-c54l4" Dec 05 08:50:46 crc kubenswrapper[4997]: I1205 08:50:46.170755 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-pmf4c"] Dec 05 08:50:46 crc kubenswrapper[4997]: I1205 08:50:46.347070 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e2cd-account-create-update-c54l4"] Dec 05 08:50:46 crc kubenswrapper[4997]: W1205 08:50:46.349783 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f94c087_4666_45bd_92b5_8d47209a0bc6.slice/crio-44cd78d9ee2fcd38b70158a4e513118c2ac94583fd6e4990876dbe0ad3bdae1f WatchSource:0}: Error finding container 44cd78d9ee2fcd38b70158a4e513118c2ac94583fd6e4990876dbe0ad3bdae1f: Status 404 returned error can't find the container with id 44cd78d9ee2fcd38b70158a4e513118c2ac94583fd6e4990876dbe0ad3bdae1f Dec 05 08:50:46 crc kubenswrapper[4997]: I1205 08:50:46.773119 4997 generic.go:334] "Generic (PLEG): container finished" podID="54fc3d55-a800-4f20-ad4b-eb48952d3660" containerID="afe5d97d369a654bc9dbb3bbb1778a60d7e857b4f56127530b272e3a87343b85" exitCode=0 Dec 05 08:50:46 crc kubenswrapper[4997]: I1205 08:50:46.773166 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-pmf4c" event={"ID":"54fc3d55-a800-4f20-ad4b-eb48952d3660","Type":"ContainerDied","Data":"afe5d97d369a654bc9dbb3bbb1778a60d7e857b4f56127530b272e3a87343b85"} Dec 05 08:50:46 crc kubenswrapper[4997]: I1205 08:50:46.773347 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-pmf4c" event={"ID":"54fc3d55-a800-4f20-ad4b-eb48952d3660","Type":"ContainerStarted","Data":"281aa531c99b8085281442fce97cfb12505445d5f6a3242e896e1613aa312024"} Dec 05 08:50:46 crc kubenswrapper[4997]: I1205 08:50:46.775179 4997 generic.go:334] "Generic (PLEG): container finished" podID="8f94c087-4666-45bd-92b5-8d47209a0bc6" containerID="ccccf89fa650efef7e7ae39fcfbbf18cb328eab8e34c99bb541980db90e9a2e8" exitCode=0 Dec 05 08:50:46 crc kubenswrapper[4997]: I1205 08:50:46.775223 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e2cd-account-create-update-c54l4" event={"ID":"8f94c087-4666-45bd-92b5-8d47209a0bc6","Type":"ContainerDied","Data":"ccccf89fa650efef7e7ae39fcfbbf18cb328eab8e34c99bb541980db90e9a2e8"} Dec 05 08:50:46 crc kubenswrapper[4997]: I1205 08:50:46.775285 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e2cd-account-create-update-c54l4" event={"ID":"8f94c087-4666-45bd-92b5-8d47209a0bc6","Type":"ContainerStarted","Data":"44cd78d9ee2fcd38b70158a4e513118c2ac94583fd6e4990876dbe0ad3bdae1f"} Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.227852 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e2cd-account-create-update-c54l4" Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.234979 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-pmf4c" Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.285761 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f94c087-4666-45bd-92b5-8d47209a0bc6-operator-scripts\") pod \"8f94c087-4666-45bd-92b5-8d47209a0bc6\" (UID: \"8f94c087-4666-45bd-92b5-8d47209a0bc6\") " Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.285845 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54fc3d55-a800-4f20-ad4b-eb48952d3660-operator-scripts\") pod \"54fc3d55-a800-4f20-ad4b-eb48952d3660\" (UID: \"54fc3d55-a800-4f20-ad4b-eb48952d3660\") " Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.285959 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvxp7\" (UniqueName: \"kubernetes.io/projected/8f94c087-4666-45bd-92b5-8d47209a0bc6-kube-api-access-kvxp7\") pod \"8f94c087-4666-45bd-92b5-8d47209a0bc6\" (UID: \"8f94c087-4666-45bd-92b5-8d47209a0bc6\") " Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.286815 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f94c087-4666-45bd-92b5-8d47209a0bc6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8f94c087-4666-45bd-92b5-8d47209a0bc6" (UID: "8f94c087-4666-45bd-92b5-8d47209a0bc6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.286847 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxkzg\" (UniqueName: \"kubernetes.io/projected/54fc3d55-a800-4f20-ad4b-eb48952d3660-kube-api-access-gxkzg\") pod \"54fc3d55-a800-4f20-ad4b-eb48952d3660\" (UID: \"54fc3d55-a800-4f20-ad4b-eb48952d3660\") " Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.287131 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54fc3d55-a800-4f20-ad4b-eb48952d3660-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "54fc3d55-a800-4f20-ad4b-eb48952d3660" (UID: "54fc3d55-a800-4f20-ad4b-eb48952d3660"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.287345 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f94c087-4666-45bd-92b5-8d47209a0bc6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.287366 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54fc3d55-a800-4f20-ad4b-eb48952d3660-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.291255 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54fc3d55-a800-4f20-ad4b-eb48952d3660-kube-api-access-gxkzg" (OuterVolumeSpecName: "kube-api-access-gxkzg") pod "54fc3d55-a800-4f20-ad4b-eb48952d3660" (UID: "54fc3d55-a800-4f20-ad4b-eb48952d3660"). InnerVolumeSpecName "kube-api-access-gxkzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.296931 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f94c087-4666-45bd-92b5-8d47209a0bc6-kube-api-access-kvxp7" (OuterVolumeSpecName: "kube-api-access-kvxp7") pod "8f94c087-4666-45bd-92b5-8d47209a0bc6" (UID: "8f94c087-4666-45bd-92b5-8d47209a0bc6"). InnerVolumeSpecName "kube-api-access-kvxp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.388802 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxkzg\" (UniqueName: \"kubernetes.io/projected/54fc3d55-a800-4f20-ad4b-eb48952d3660-kube-api-access-gxkzg\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.389042 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvxp7\" (UniqueName: \"kubernetes.io/projected/8f94c087-4666-45bd-92b5-8d47209a0bc6-kube-api-access-kvxp7\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.791304 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-pmf4c" event={"ID":"54fc3d55-a800-4f20-ad4b-eb48952d3660","Type":"ContainerDied","Data":"281aa531c99b8085281442fce97cfb12505445d5f6a3242e896e1613aa312024"} Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.791548 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="281aa531c99b8085281442fce97cfb12505445d5f6a3242e896e1613aa312024" Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.791365 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-pmf4c" Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.792581 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e2cd-account-create-update-c54l4" event={"ID":"8f94c087-4666-45bd-92b5-8d47209a0bc6","Type":"ContainerDied","Data":"44cd78d9ee2fcd38b70158a4e513118c2ac94583fd6e4990876dbe0ad3bdae1f"} Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.792603 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44cd78d9ee2fcd38b70158a4e513118c2ac94583fd6e4990876dbe0ad3bdae1f" Dec 05 08:50:48 crc kubenswrapper[4997]: I1205 08:50:48.792683 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e2cd-account-create-update-c54l4" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.008826 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-6vpxj"] Dec 05 08:50:51 crc kubenswrapper[4997]: E1205 08:50:51.009166 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54fc3d55-a800-4f20-ad4b-eb48952d3660" containerName="mariadb-database-create" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.009179 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="54fc3d55-a800-4f20-ad4b-eb48952d3660" containerName="mariadb-database-create" Dec 05 08:50:51 crc kubenswrapper[4997]: E1205 08:50:51.009196 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f94c087-4666-45bd-92b5-8d47209a0bc6" containerName="mariadb-account-create-update" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.009203 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f94c087-4666-45bd-92b5-8d47209a0bc6" containerName="mariadb-account-create-update" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.009370 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f94c087-4666-45bd-92b5-8d47209a0bc6" containerName="mariadb-account-create-update" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.009385 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="54fc3d55-a800-4f20-ad4b-eb48952d3660" containerName="mariadb-database-create" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.009962 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6vpxj" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.012202 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.012315 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-g9bms" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.012708 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.014164 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.025880 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-6vpxj"] Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.130883 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b68a777-809c-4794-bb7f-6b336a78130c-config-data\") pod \"keystone-db-sync-6vpxj\" (UID: \"0b68a777-809c-4794-bb7f-6b336a78130c\") " pod="openstack/keystone-db-sync-6vpxj" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.130983 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bj2w\" (UniqueName: \"kubernetes.io/projected/0b68a777-809c-4794-bb7f-6b336a78130c-kube-api-access-2bj2w\") pod \"keystone-db-sync-6vpxj\" (UID: \"0b68a777-809c-4794-bb7f-6b336a78130c\") " pod="openstack/keystone-db-sync-6vpxj" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.131026 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b68a777-809c-4794-bb7f-6b336a78130c-combined-ca-bundle\") pod \"keystone-db-sync-6vpxj\" (UID: \"0b68a777-809c-4794-bb7f-6b336a78130c\") " pod="openstack/keystone-db-sync-6vpxj" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.232688 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b68a777-809c-4794-bb7f-6b336a78130c-config-data\") pod \"keystone-db-sync-6vpxj\" (UID: \"0b68a777-809c-4794-bb7f-6b336a78130c\") " pod="openstack/keystone-db-sync-6vpxj" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.232797 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bj2w\" (UniqueName: \"kubernetes.io/projected/0b68a777-809c-4794-bb7f-6b336a78130c-kube-api-access-2bj2w\") pod \"keystone-db-sync-6vpxj\" (UID: \"0b68a777-809c-4794-bb7f-6b336a78130c\") " pod="openstack/keystone-db-sync-6vpxj" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.232833 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b68a777-809c-4794-bb7f-6b336a78130c-combined-ca-bundle\") pod \"keystone-db-sync-6vpxj\" (UID: \"0b68a777-809c-4794-bb7f-6b336a78130c\") " pod="openstack/keystone-db-sync-6vpxj" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.237945 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b68a777-809c-4794-bb7f-6b336a78130c-config-data\") pod \"keystone-db-sync-6vpxj\" (UID: \"0b68a777-809c-4794-bb7f-6b336a78130c\") " pod="openstack/keystone-db-sync-6vpxj" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.239130 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b68a777-809c-4794-bb7f-6b336a78130c-combined-ca-bundle\") pod \"keystone-db-sync-6vpxj\" (UID: \"0b68a777-809c-4794-bb7f-6b336a78130c\") " pod="openstack/keystone-db-sync-6vpxj" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.251666 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bj2w\" (UniqueName: \"kubernetes.io/projected/0b68a777-809c-4794-bb7f-6b336a78130c-kube-api-access-2bj2w\") pod \"keystone-db-sync-6vpxj\" (UID: \"0b68a777-809c-4794-bb7f-6b336a78130c\") " pod="openstack/keystone-db-sync-6vpxj" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.329092 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6vpxj" Dec 05 08:50:51 crc kubenswrapper[4997]: I1205 08:50:51.805339 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-6vpxj"] Dec 05 08:50:51 crc kubenswrapper[4997]: W1205 08:50:51.820932 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b68a777_809c_4794_bb7f_6b336a78130c.slice/crio-3a5c3561e0117793bcaa4838bf5bf3f778ca176cc7df85afc58e5fcd1dcadfbd WatchSource:0}: Error finding container 3a5c3561e0117793bcaa4838bf5bf3f778ca176cc7df85afc58e5fcd1dcadfbd: Status 404 returned error can't find the container with id 3a5c3561e0117793bcaa4838bf5bf3f778ca176cc7df85afc58e5fcd1dcadfbd Dec 05 08:50:52 crc kubenswrapper[4997]: I1205 08:50:52.820539 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6vpxj" event={"ID":"0b68a777-809c-4794-bb7f-6b336a78130c","Type":"ContainerStarted","Data":"3a5c3561e0117793bcaa4838bf5bf3f778ca176cc7df85afc58e5fcd1dcadfbd"} Dec 05 08:50:54 crc kubenswrapper[4997]: I1205 08:50:54.334423 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 05 08:50:55 crc kubenswrapper[4997]: I1205 08:50:55.749748 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:50:55 crc kubenswrapper[4997]: E1205 08:50:55.750029 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:50:57 crc kubenswrapper[4997]: I1205 08:50:57.865229 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6vpxj" event={"ID":"0b68a777-809c-4794-bb7f-6b336a78130c","Type":"ContainerStarted","Data":"3ed6e708fc661912301a30eb43674bccd45365ffa4096fdfa3fb1cf18c6ee779"} Dec 05 08:50:57 crc kubenswrapper[4997]: I1205 08:50:57.882844 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-6vpxj" podStartSLOduration=3.127391441 podStartE2EDuration="7.88282376s" podCreationTimestamp="2025-12-05 08:50:50 +0000 UTC" firstStartedPulling="2025-12-05 08:50:51.831125676 +0000 UTC m=+6952.360032937" lastFinishedPulling="2025-12-05 08:50:56.586557985 +0000 UTC m=+6957.115465256" observedRunningTime="2025-12-05 08:50:57.878490823 +0000 UTC m=+6958.407398104" watchObservedRunningTime="2025-12-05 08:50:57.88282376 +0000 UTC m=+6958.411731021" Dec 05 08:50:58 crc kubenswrapper[4997]: I1205 08:50:58.874740 4997 generic.go:334] "Generic (PLEG): container finished" podID="0b68a777-809c-4794-bb7f-6b336a78130c" containerID="3ed6e708fc661912301a30eb43674bccd45365ffa4096fdfa3fb1cf18c6ee779" exitCode=0 Dec 05 08:50:58 crc kubenswrapper[4997]: I1205 08:50:58.874853 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6vpxj" event={"ID":"0b68a777-809c-4794-bb7f-6b336a78130c","Type":"ContainerDied","Data":"3ed6e708fc661912301a30eb43674bccd45365ffa4096fdfa3fb1cf18c6ee779"} Dec 05 08:51:00 crc kubenswrapper[4997]: I1205 08:51:00.233070 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6vpxj" Dec 05 08:51:00 crc kubenswrapper[4997]: I1205 08:51:00.303602 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b68a777-809c-4794-bb7f-6b336a78130c-combined-ca-bundle\") pod \"0b68a777-809c-4794-bb7f-6b336a78130c\" (UID: \"0b68a777-809c-4794-bb7f-6b336a78130c\") " Dec 05 08:51:00 crc kubenswrapper[4997]: I1205 08:51:00.303919 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b68a777-809c-4794-bb7f-6b336a78130c-config-data\") pod \"0b68a777-809c-4794-bb7f-6b336a78130c\" (UID: \"0b68a777-809c-4794-bb7f-6b336a78130c\") " Dec 05 08:51:00 crc kubenswrapper[4997]: I1205 08:51:00.304058 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bj2w\" (UniqueName: \"kubernetes.io/projected/0b68a777-809c-4794-bb7f-6b336a78130c-kube-api-access-2bj2w\") pod \"0b68a777-809c-4794-bb7f-6b336a78130c\" (UID: \"0b68a777-809c-4794-bb7f-6b336a78130c\") " Dec 05 08:51:00 crc kubenswrapper[4997]: I1205 08:51:00.313862 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b68a777-809c-4794-bb7f-6b336a78130c-kube-api-access-2bj2w" (OuterVolumeSpecName: "kube-api-access-2bj2w") pod "0b68a777-809c-4794-bb7f-6b336a78130c" (UID: "0b68a777-809c-4794-bb7f-6b336a78130c"). InnerVolumeSpecName "kube-api-access-2bj2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:51:00 crc kubenswrapper[4997]: I1205 08:51:00.325651 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b68a777-809c-4794-bb7f-6b336a78130c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b68a777-809c-4794-bb7f-6b336a78130c" (UID: "0b68a777-809c-4794-bb7f-6b336a78130c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:00 crc kubenswrapper[4997]: I1205 08:51:00.344122 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b68a777-809c-4794-bb7f-6b336a78130c-config-data" (OuterVolumeSpecName: "config-data") pod "0b68a777-809c-4794-bb7f-6b336a78130c" (UID: "0b68a777-809c-4794-bb7f-6b336a78130c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:00 crc kubenswrapper[4997]: I1205 08:51:00.406760 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b68a777-809c-4794-bb7f-6b336a78130c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:00 crc kubenswrapper[4997]: I1205 08:51:00.406799 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b68a777-809c-4794-bb7f-6b336a78130c-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:00 crc kubenswrapper[4997]: I1205 08:51:00.406811 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bj2w\" (UniqueName: \"kubernetes.io/projected/0b68a777-809c-4794-bb7f-6b336a78130c-kube-api-access-2bj2w\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:00 crc kubenswrapper[4997]: I1205 08:51:00.888823 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6vpxj" event={"ID":"0b68a777-809c-4794-bb7f-6b336a78130c","Type":"ContainerDied","Data":"3a5c3561e0117793bcaa4838bf5bf3f778ca176cc7df85afc58e5fcd1dcadfbd"} Dec 05 08:51:00 crc kubenswrapper[4997]: I1205 08:51:00.888869 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a5c3561e0117793bcaa4838bf5bf3f778ca176cc7df85afc58e5fcd1dcadfbd" Dec 05 08:51:00 crc kubenswrapper[4997]: I1205 08:51:00.888875 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6vpxj" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.199381 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-9glv9"] Dec 05 08:51:01 crc kubenswrapper[4997]: E1205 08:51:01.200271 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b68a777-809c-4794-bb7f-6b336a78130c" containerName="keystone-db-sync" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.200283 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b68a777-809c-4794-bb7f-6b336a78130c" containerName="keystone-db-sync" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.200455 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b68a777-809c-4794-bb7f-6b336a78130c" containerName="keystone-db-sync" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.201115 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.212303 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.212509 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.212635 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.213204 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-g9bms" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.214630 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6468849cd7-rbk8q"] Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.217474 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.221689 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.226085 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9glv9"] Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.282496 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6468849cd7-rbk8q"] Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.321771 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-ovsdbserver-nb\") pod \"dnsmasq-dns-6468849cd7-rbk8q\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.321871 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-config\") pod \"dnsmasq-dns-6468849cd7-rbk8q\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.321895 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-ovsdbserver-sb\") pod \"dnsmasq-dns-6468849cd7-rbk8q\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.321937 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-combined-ca-bundle\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.322112 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-fernet-keys\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.322234 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-dns-svc\") pod \"dnsmasq-dns-6468849cd7-rbk8q\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.322293 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-credential-keys\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.322364 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8h7q\" (UniqueName: \"kubernetes.io/projected/a17a3529-5b5c-4029-bb2a-f5883d800c4d-kube-api-access-p8h7q\") pod \"dnsmasq-dns-6468849cd7-rbk8q\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.322402 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42rvn\" (UniqueName: \"kubernetes.io/projected/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-kube-api-access-42rvn\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.322421 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-scripts\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.322436 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-config-data\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.423897 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8h7q\" (UniqueName: \"kubernetes.io/projected/a17a3529-5b5c-4029-bb2a-f5883d800c4d-kube-api-access-p8h7q\") pod \"dnsmasq-dns-6468849cd7-rbk8q\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.423953 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42rvn\" (UniqueName: \"kubernetes.io/projected/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-kube-api-access-42rvn\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.423975 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-scripts\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.423992 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-config-data\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.424008 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-ovsdbserver-nb\") pod \"dnsmasq-dns-6468849cd7-rbk8q\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.424044 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-config\") pod \"dnsmasq-dns-6468849cd7-rbk8q\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.424059 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-ovsdbserver-sb\") pod \"dnsmasq-dns-6468849cd7-rbk8q\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.424098 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-combined-ca-bundle\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.424132 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-fernet-keys\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.424163 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-dns-svc\") pod \"dnsmasq-dns-6468849cd7-rbk8q\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.424184 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-credential-keys\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.425541 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-ovsdbserver-sb\") pod \"dnsmasq-dns-6468849cd7-rbk8q\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.425561 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-dns-svc\") pod \"dnsmasq-dns-6468849cd7-rbk8q\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.425665 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-ovsdbserver-nb\") pod \"dnsmasq-dns-6468849cd7-rbk8q\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.426003 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-config\") pod \"dnsmasq-dns-6468849cd7-rbk8q\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.428671 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-config-data\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.440134 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-scripts\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.442602 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-fernet-keys\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.446585 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-credential-keys\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.458269 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-combined-ca-bundle\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.474272 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42rvn\" (UniqueName: \"kubernetes.io/projected/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-kube-api-access-42rvn\") pod \"keystone-bootstrap-9glv9\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.476632 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8h7q\" (UniqueName: \"kubernetes.io/projected/a17a3529-5b5c-4029-bb2a-f5883d800c4d-kube-api-access-p8h7q\") pod \"dnsmasq-dns-6468849cd7-rbk8q\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.521800 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:01 crc kubenswrapper[4997]: I1205 08:51:01.544966 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:02 crc kubenswrapper[4997]: I1205 08:51:02.044524 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9glv9"] Dec 05 08:51:02 crc kubenswrapper[4997]: W1205 08:51:02.049511 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cc6de13_ef1f_455e_9bc9_d1a3bbdb66b5.slice/crio-18be92d7f1f1b9199a8ebfdf4c0305983f7830cfdc207185b21fa16993f3aa90 WatchSource:0}: Error finding container 18be92d7f1f1b9199a8ebfdf4c0305983f7830cfdc207185b21fa16993f3aa90: Status 404 returned error can't find the container with id 18be92d7f1f1b9199a8ebfdf4c0305983f7830cfdc207185b21fa16993f3aa90 Dec 05 08:51:02 crc kubenswrapper[4997]: W1205 08:51:02.207296 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda17a3529_5b5c_4029_bb2a_f5883d800c4d.slice/crio-07112dd7e7b4204671cfc0280d87dfb0cf40b44d0c78dfd9bae828364615d91b WatchSource:0}: Error finding container 07112dd7e7b4204671cfc0280d87dfb0cf40b44d0c78dfd9bae828364615d91b: Status 404 returned error can't find the container with id 07112dd7e7b4204671cfc0280d87dfb0cf40b44d0c78dfd9bae828364615d91b Dec 05 08:51:02 crc kubenswrapper[4997]: I1205 08:51:02.209234 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6468849cd7-rbk8q"] Dec 05 08:51:02 crc kubenswrapper[4997]: I1205 08:51:02.912296 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" event={"ID":"a17a3529-5b5c-4029-bb2a-f5883d800c4d","Type":"ContainerStarted","Data":"07112dd7e7b4204671cfc0280d87dfb0cf40b44d0c78dfd9bae828364615d91b"} Dec 05 08:51:02 crc kubenswrapper[4997]: I1205 08:51:02.913833 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9glv9" event={"ID":"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5","Type":"ContainerStarted","Data":"18be92d7f1f1b9199a8ebfdf4c0305983f7830cfdc207185b21fa16993f3aa90"} Dec 05 08:51:03 crc kubenswrapper[4997]: I1205 08:51:03.922326 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9glv9" event={"ID":"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5","Type":"ContainerStarted","Data":"26e04d5978d8454a5e28dbfd786bc8d7c723835268518aac180c43f3aa839922"} Dec 05 08:51:03 crc kubenswrapper[4997]: I1205 08:51:03.923922 4997 generic.go:334] "Generic (PLEG): container finished" podID="a17a3529-5b5c-4029-bb2a-f5883d800c4d" containerID="9b8a0e39b7c8a7a70943fe99e20575281db387be1814f84d64f6a8c13eb4c93a" exitCode=0 Dec 05 08:51:03 crc kubenswrapper[4997]: I1205 08:51:03.923972 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" event={"ID":"a17a3529-5b5c-4029-bb2a-f5883d800c4d","Type":"ContainerDied","Data":"9b8a0e39b7c8a7a70943fe99e20575281db387be1814f84d64f6a8c13eb4c93a"} Dec 05 08:51:03 crc kubenswrapper[4997]: I1205 08:51:03.954488 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-9glv9" podStartSLOduration=2.9544714499999998 podStartE2EDuration="2.95447145s" podCreationTimestamp="2025-12-05 08:51:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:51:03.94776923 +0000 UTC m=+6964.476676491" watchObservedRunningTime="2025-12-05 08:51:03.95447145 +0000 UTC m=+6964.483378711" Dec 05 08:51:04 crc kubenswrapper[4997]: I1205 08:51:04.940158 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" event={"ID":"a17a3529-5b5c-4029-bb2a-f5883d800c4d","Type":"ContainerStarted","Data":"ea64bf5ff5e848152bb443a88b44980e48b57427a15e6bb36d885c2bb453ca13"} Dec 05 08:51:04 crc kubenswrapper[4997]: I1205 08:51:04.973111 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" podStartSLOduration=3.973088561 podStartE2EDuration="3.973088561s" podCreationTimestamp="2025-12-05 08:51:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:51:04.964036498 +0000 UTC m=+6965.492943779" watchObservedRunningTime="2025-12-05 08:51:04.973088561 +0000 UTC m=+6965.501995842" Dec 05 08:51:05 crc kubenswrapper[4997]: I1205 08:51:05.946272 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:07 crc kubenswrapper[4997]: I1205 08:51:07.963050 4997 generic.go:334] "Generic (PLEG): container finished" podID="5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5" containerID="26e04d5978d8454a5e28dbfd786bc8d7c723835268518aac180c43f3aa839922" exitCode=0 Dec 05 08:51:07 crc kubenswrapper[4997]: I1205 08:51:07.963153 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9glv9" event={"ID":"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5","Type":"ContainerDied","Data":"26e04d5978d8454a5e28dbfd786bc8d7c723835268518aac180c43f3aa839922"} Dec 05 08:51:08 crc kubenswrapper[4997]: I1205 08:51:08.749280 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:51:08 crc kubenswrapper[4997]: E1205 08:51:08.749572 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.290378 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.400286 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42rvn\" (UniqueName: \"kubernetes.io/projected/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-kube-api-access-42rvn\") pod \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.400351 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-scripts\") pod \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.400388 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-combined-ca-bundle\") pod \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.400420 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-config-data\") pod \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.400507 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-fernet-keys\") pod \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.400605 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-credential-keys\") pod \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\" (UID: \"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5\") " Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.410856 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5" (UID: "5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.410877 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-scripts" (OuterVolumeSpecName: "scripts") pod "5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5" (UID: "5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.410911 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-kube-api-access-42rvn" (OuterVolumeSpecName: "kube-api-access-42rvn") pod "5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5" (UID: "5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5"). InnerVolumeSpecName "kube-api-access-42rvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.410938 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5" (UID: "5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.427959 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-config-data" (OuterVolumeSpecName: "config-data") pod "5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5" (UID: "5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.429077 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5" (UID: "5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.502681 4997 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.502725 4997 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.502736 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42rvn\" (UniqueName: \"kubernetes.io/projected/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-kube-api-access-42rvn\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.502745 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.502754 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.502761 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.977337 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9glv9" event={"ID":"5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5","Type":"ContainerDied","Data":"18be92d7f1f1b9199a8ebfdf4c0305983f7830cfdc207185b21fa16993f3aa90"} Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.977375 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18be92d7f1f1b9199a8ebfdf4c0305983f7830cfdc207185b21fa16993f3aa90" Dec 05 08:51:09 crc kubenswrapper[4997]: I1205 08:51:09.977456 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9glv9" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.075041 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-9glv9"] Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.084425 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-9glv9"] Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.168100 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-sxn7d"] Dec 05 08:51:10 crc kubenswrapper[4997]: E1205 08:51:10.168726 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5" containerName="keystone-bootstrap" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.168751 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5" containerName="keystone-bootstrap" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.168953 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5" containerName="keystone-bootstrap" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.169766 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.171653 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.172377 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.172471 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-g9bms" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.172498 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.172498 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.190602 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sxn7d"] Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.219543 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-combined-ca-bundle\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.219603 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-scripts\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.219686 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-credential-keys\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.219731 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-config-data\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.219760 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-fernet-keys\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.219800 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qg7b\" (UniqueName: \"kubernetes.io/projected/8d1240db-51c8-4e89-9165-d6a27b973908-kube-api-access-6qg7b\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.321958 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-config-data\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.322026 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-fernet-keys\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.322073 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qg7b\" (UniqueName: \"kubernetes.io/projected/8d1240db-51c8-4e89-9165-d6a27b973908-kube-api-access-6qg7b\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.322121 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-combined-ca-bundle\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.322148 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-scripts\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.322199 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-credential-keys\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.332334 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-scripts\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.335400 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-fernet-keys\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.341754 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-config-data\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.342673 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-credential-keys\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.346225 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-combined-ca-bundle\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.354373 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qg7b\" (UniqueName: \"kubernetes.io/projected/8d1240db-51c8-4e89-9165-d6a27b973908-kube-api-access-6qg7b\") pod \"keystone-bootstrap-sxn7d\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.497156 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.722510 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sxn7d"] Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.987110 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sxn7d" event={"ID":"8d1240db-51c8-4e89-9165-d6a27b973908","Type":"ContainerStarted","Data":"54fc76a663c1f19d294beeaff96a9faae058033865ef0a7de32aa3f7206c65e7"} Dec 05 08:51:10 crc kubenswrapper[4997]: I1205 08:51:10.987434 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sxn7d" event={"ID":"8d1240db-51c8-4e89-9165-d6a27b973908","Type":"ContainerStarted","Data":"99e9651f6b41d0e09d4a4515a5b0c12a5b11fd7f0edfc36fea0710974e68bb19"} Dec 05 08:51:11 crc kubenswrapper[4997]: I1205 08:51:11.004130 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-sxn7d" podStartSLOduration=1.004098099 podStartE2EDuration="1.004098099s" podCreationTimestamp="2025-12-05 08:51:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:51:11.003136163 +0000 UTC m=+6971.532043444" watchObservedRunningTime="2025-12-05 08:51:11.004098099 +0000 UTC m=+6971.533005360" Dec 05 08:51:11 crc kubenswrapper[4997]: I1205 08:51:11.547566 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:51:11 crc kubenswrapper[4997]: I1205 08:51:11.622402 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-574bcff97-jb2rd"] Dec 05 08:51:11 crc kubenswrapper[4997]: I1205 08:51:11.622722 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-574bcff97-jb2rd" podUID="c2241378-742b-4a1f-b21a-01213d3745e8" containerName="dnsmasq-dns" containerID="cri-o://10881e6afa1e07aa4b4a1ce0e66000192dc87d8f59b7c1e12df58e18b4c282a7" gracePeriod=10 Dec 05 08:51:11 crc kubenswrapper[4997]: I1205 08:51:11.785863 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5" path="/var/lib/kubelet/pods/5cc6de13-ef1f-455e-9bc9-d1a3bbdb66b5/volumes" Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.012013 4997 generic.go:334] "Generic (PLEG): container finished" podID="c2241378-742b-4a1f-b21a-01213d3745e8" containerID="10881e6afa1e07aa4b4a1ce0e66000192dc87d8f59b7c1e12df58e18b4c282a7" exitCode=0 Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.012743 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574bcff97-jb2rd" event={"ID":"c2241378-742b-4a1f-b21a-01213d3745e8","Type":"ContainerDied","Data":"10881e6afa1e07aa4b4a1ce0e66000192dc87d8f59b7c1e12df58e18b4c282a7"} Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.152147 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.258419 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-config\") pod \"c2241378-742b-4a1f-b21a-01213d3745e8\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.258578 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-dns-svc\") pod \"c2241378-742b-4a1f-b21a-01213d3745e8\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.258656 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-ovsdbserver-nb\") pod \"c2241378-742b-4a1f-b21a-01213d3745e8\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.258687 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-ovsdbserver-sb\") pod \"c2241378-742b-4a1f-b21a-01213d3745e8\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.258721 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpwd2\" (UniqueName: \"kubernetes.io/projected/c2241378-742b-4a1f-b21a-01213d3745e8-kube-api-access-qpwd2\") pod \"c2241378-742b-4a1f-b21a-01213d3745e8\" (UID: \"c2241378-742b-4a1f-b21a-01213d3745e8\") " Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.265885 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2241378-742b-4a1f-b21a-01213d3745e8-kube-api-access-qpwd2" (OuterVolumeSpecName: "kube-api-access-qpwd2") pod "c2241378-742b-4a1f-b21a-01213d3745e8" (UID: "c2241378-742b-4a1f-b21a-01213d3745e8"). InnerVolumeSpecName "kube-api-access-qpwd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.299919 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c2241378-742b-4a1f-b21a-01213d3745e8" (UID: "c2241378-742b-4a1f-b21a-01213d3745e8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.300207 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-config" (OuterVolumeSpecName: "config") pod "c2241378-742b-4a1f-b21a-01213d3745e8" (UID: "c2241378-742b-4a1f-b21a-01213d3745e8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.302756 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c2241378-742b-4a1f-b21a-01213d3745e8" (UID: "c2241378-742b-4a1f-b21a-01213d3745e8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.303356 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c2241378-742b-4a1f-b21a-01213d3745e8" (UID: "c2241378-742b-4a1f-b21a-01213d3745e8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.362868 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.362912 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.362924 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.362935 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2241378-742b-4a1f-b21a-01213d3745e8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:12 crc kubenswrapper[4997]: I1205 08:51:12.362952 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpwd2\" (UniqueName: \"kubernetes.io/projected/c2241378-742b-4a1f-b21a-01213d3745e8-kube-api-access-qpwd2\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:13 crc kubenswrapper[4997]: I1205 08:51:13.024446 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574bcff97-jb2rd" event={"ID":"c2241378-742b-4a1f-b21a-01213d3745e8","Type":"ContainerDied","Data":"369c1fa922661a6aa1f3e7aeaff28e4d0c6d107e22b786a7d47e1cf21687fa25"} Dec 05 08:51:13 crc kubenswrapper[4997]: I1205 08:51:13.024760 4997 scope.go:117] "RemoveContainer" containerID="10881e6afa1e07aa4b4a1ce0e66000192dc87d8f59b7c1e12df58e18b4c282a7" Dec 05 08:51:13 crc kubenswrapper[4997]: I1205 08:51:13.024657 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574bcff97-jb2rd" Dec 05 08:51:13 crc kubenswrapper[4997]: I1205 08:51:13.049462 4997 scope.go:117] "RemoveContainer" containerID="3495ae35600724700c55e556d248037fcb36635e0759c13ed5d0041ad88980cb" Dec 05 08:51:13 crc kubenswrapper[4997]: I1205 08:51:13.075745 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-574bcff97-jb2rd"] Dec 05 08:51:13 crc kubenswrapper[4997]: I1205 08:51:13.085174 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-574bcff97-jb2rd"] Dec 05 08:51:13 crc kubenswrapper[4997]: I1205 08:51:13.759397 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2241378-742b-4a1f-b21a-01213d3745e8" path="/var/lib/kubelet/pods/c2241378-742b-4a1f-b21a-01213d3745e8/volumes" Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.037405 4997 generic.go:334] "Generic (PLEG): container finished" podID="8d1240db-51c8-4e89-9165-d6a27b973908" containerID="54fc76a663c1f19d294beeaff96a9faae058033865ef0a7de32aa3f7206c65e7" exitCode=0 Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.037474 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sxn7d" event={"ID":"8d1240db-51c8-4e89-9165-d6a27b973908","Type":"ContainerDied","Data":"54fc76a663c1f19d294beeaff96a9faae058033865ef0a7de32aa3f7206c65e7"} Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.038975 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t9jkl"] Dec 05 08:51:14 crc kubenswrapper[4997]: E1205 08:51:14.039529 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2241378-742b-4a1f-b21a-01213d3745e8" containerName="dnsmasq-dns" Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.039591 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2241378-742b-4a1f-b21a-01213d3745e8" containerName="dnsmasq-dns" Dec 05 08:51:14 crc kubenswrapper[4997]: E1205 08:51:14.039647 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2241378-742b-4a1f-b21a-01213d3745e8" containerName="init" Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.039656 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2241378-742b-4a1f-b21a-01213d3745e8" containerName="init" Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.041004 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2241378-742b-4a1f-b21a-01213d3745e8" containerName="dnsmasq-dns" Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.042528 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t9jkl" Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.059523 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t9jkl"] Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.190093 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/210251f1-d0c4-483e-ac78-20aa69438b25-catalog-content\") pod \"redhat-operators-t9jkl\" (UID: \"210251f1-d0c4-483e-ac78-20aa69438b25\") " pod="openshift-marketplace/redhat-operators-t9jkl" Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.190146 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bn5k\" (UniqueName: \"kubernetes.io/projected/210251f1-d0c4-483e-ac78-20aa69438b25-kube-api-access-6bn5k\") pod \"redhat-operators-t9jkl\" (UID: \"210251f1-d0c4-483e-ac78-20aa69438b25\") " pod="openshift-marketplace/redhat-operators-t9jkl" Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.190241 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/210251f1-d0c4-483e-ac78-20aa69438b25-utilities\") pod \"redhat-operators-t9jkl\" (UID: \"210251f1-d0c4-483e-ac78-20aa69438b25\") " pod="openshift-marketplace/redhat-operators-t9jkl" Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.291543 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bn5k\" (UniqueName: \"kubernetes.io/projected/210251f1-d0c4-483e-ac78-20aa69438b25-kube-api-access-6bn5k\") pod \"redhat-operators-t9jkl\" (UID: \"210251f1-d0c4-483e-ac78-20aa69438b25\") " pod="openshift-marketplace/redhat-operators-t9jkl" Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.291679 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/210251f1-d0c4-483e-ac78-20aa69438b25-utilities\") pod \"redhat-operators-t9jkl\" (UID: \"210251f1-d0c4-483e-ac78-20aa69438b25\") " pod="openshift-marketplace/redhat-operators-t9jkl" Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.291767 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/210251f1-d0c4-483e-ac78-20aa69438b25-catalog-content\") pod \"redhat-operators-t9jkl\" (UID: \"210251f1-d0c4-483e-ac78-20aa69438b25\") " pod="openshift-marketplace/redhat-operators-t9jkl" Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.292398 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/210251f1-d0c4-483e-ac78-20aa69438b25-catalog-content\") pod \"redhat-operators-t9jkl\" (UID: \"210251f1-d0c4-483e-ac78-20aa69438b25\") " pod="openshift-marketplace/redhat-operators-t9jkl" Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.295797 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/210251f1-d0c4-483e-ac78-20aa69438b25-utilities\") pod \"redhat-operators-t9jkl\" (UID: \"210251f1-d0c4-483e-ac78-20aa69438b25\") " pod="openshift-marketplace/redhat-operators-t9jkl" Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.318675 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bn5k\" (UniqueName: \"kubernetes.io/projected/210251f1-d0c4-483e-ac78-20aa69438b25-kube-api-access-6bn5k\") pod \"redhat-operators-t9jkl\" (UID: \"210251f1-d0c4-483e-ac78-20aa69438b25\") " pod="openshift-marketplace/redhat-operators-t9jkl" Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.369663 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t9jkl" Dec 05 08:51:14 crc kubenswrapper[4997]: I1205 08:51:14.929056 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t9jkl"] Dec 05 08:51:14 crc kubenswrapper[4997]: W1205 08:51:14.930799 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod210251f1_d0c4_483e_ac78_20aa69438b25.slice/crio-0595270acd252c47e8299f2919382ea91ac8b51152e8f06e4d6c84ed8b5959f2 WatchSource:0}: Error finding container 0595270acd252c47e8299f2919382ea91ac8b51152e8f06e4d6c84ed8b5959f2: Status 404 returned error can't find the container with id 0595270acd252c47e8299f2919382ea91ac8b51152e8f06e4d6c84ed8b5959f2 Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.044865 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t9jkl" event={"ID":"210251f1-d0c4-483e-ac78-20aa69438b25","Type":"ContainerStarted","Data":"0595270acd252c47e8299f2919382ea91ac8b51152e8f06e4d6c84ed8b5959f2"} Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.424911 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.614236 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-config-data\") pod \"8d1240db-51c8-4e89-9165-d6a27b973908\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.614299 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qg7b\" (UniqueName: \"kubernetes.io/projected/8d1240db-51c8-4e89-9165-d6a27b973908-kube-api-access-6qg7b\") pod \"8d1240db-51c8-4e89-9165-d6a27b973908\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.614388 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-combined-ca-bundle\") pod \"8d1240db-51c8-4e89-9165-d6a27b973908\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.614422 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-credential-keys\") pod \"8d1240db-51c8-4e89-9165-d6a27b973908\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.614457 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-fernet-keys\") pod \"8d1240db-51c8-4e89-9165-d6a27b973908\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.614531 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-scripts\") pod \"8d1240db-51c8-4e89-9165-d6a27b973908\" (UID: \"8d1240db-51c8-4e89-9165-d6a27b973908\") " Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.623508 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d1240db-51c8-4e89-9165-d6a27b973908-kube-api-access-6qg7b" (OuterVolumeSpecName: "kube-api-access-6qg7b") pod "8d1240db-51c8-4e89-9165-d6a27b973908" (UID: "8d1240db-51c8-4e89-9165-d6a27b973908"). InnerVolumeSpecName "kube-api-access-6qg7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.624840 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-scripts" (OuterVolumeSpecName: "scripts") pod "8d1240db-51c8-4e89-9165-d6a27b973908" (UID: "8d1240db-51c8-4e89-9165-d6a27b973908"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.627733 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8d1240db-51c8-4e89-9165-d6a27b973908" (UID: "8d1240db-51c8-4e89-9165-d6a27b973908"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.642172 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8d1240db-51c8-4e89-9165-d6a27b973908" (UID: "8d1240db-51c8-4e89-9165-d6a27b973908"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.662814 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d1240db-51c8-4e89-9165-d6a27b973908" (UID: "8d1240db-51c8-4e89-9165-d6a27b973908"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.665590 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-config-data" (OuterVolumeSpecName: "config-data") pod "8d1240db-51c8-4e89-9165-d6a27b973908" (UID: "8d1240db-51c8-4e89-9165-d6a27b973908"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.715913 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.715950 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.715961 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qg7b\" (UniqueName: \"kubernetes.io/projected/8d1240db-51c8-4e89-9165-d6a27b973908-kube-api-access-6qg7b\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.715971 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.715980 4997 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:15 crc kubenswrapper[4997]: I1205 08:51:15.715987 4997 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d1240db-51c8-4e89-9165-d6a27b973908-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:15 crc kubenswrapper[4997]: E1205 08:51:15.876489 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d1240db_51c8_4e89_9165_d6a27b973908.slice\": RecentStats: unable to find data in memory cache]" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.053777 4997 generic.go:334] "Generic (PLEG): container finished" podID="210251f1-d0c4-483e-ac78-20aa69438b25" containerID="d62a40908a16fc0fbf94a28c808c698998b4f9abc84ccdbd9a61a14449256d81" exitCode=0 Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.054405 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t9jkl" event={"ID":"210251f1-d0c4-483e-ac78-20aa69438b25","Type":"ContainerDied","Data":"d62a40908a16fc0fbf94a28c808c698998b4f9abc84ccdbd9a61a14449256d81"} Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.062246 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sxn7d" event={"ID":"8d1240db-51c8-4e89-9165-d6a27b973908","Type":"ContainerDied","Data":"99e9651f6b41d0e09d4a4515a5b0c12a5b11fd7f0edfc36fea0710974e68bb19"} Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.062293 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99e9651f6b41d0e09d4a4515a5b0c12a5b11fd7f0edfc36fea0710974e68bb19" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.062293 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sxn7d" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.156580 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-56858ddfd7-wzhwt"] Dec 05 08:51:16 crc kubenswrapper[4997]: E1205 08:51:16.156982 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1240db-51c8-4e89-9165-d6a27b973908" containerName="keystone-bootstrap" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.157000 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1240db-51c8-4e89-9165-d6a27b973908" containerName="keystone-bootstrap" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.157189 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d1240db-51c8-4e89-9165-d6a27b973908" containerName="keystone-bootstrap" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.157851 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.159881 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.160126 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.160308 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.160511 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-g9bms" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.181263 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-56858ddfd7-wzhwt"] Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.325410 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsglh\" (UniqueName: \"kubernetes.io/projected/56e0b4d4-eadc-456f-b9b1-30db806b94cc-kube-api-access-wsglh\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.325459 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56e0b4d4-eadc-456f-b9b1-30db806b94cc-fernet-keys\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.325559 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56e0b4d4-eadc-456f-b9b1-30db806b94cc-scripts\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.325587 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56e0b4d4-eadc-456f-b9b1-30db806b94cc-credential-keys\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.325631 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e0b4d4-eadc-456f-b9b1-30db806b94cc-combined-ca-bundle\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.325696 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e0b4d4-eadc-456f-b9b1-30db806b94cc-config-data\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.427807 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56e0b4d4-eadc-456f-b9b1-30db806b94cc-scripts\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.427884 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56e0b4d4-eadc-456f-b9b1-30db806b94cc-credential-keys\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.427910 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e0b4d4-eadc-456f-b9b1-30db806b94cc-combined-ca-bundle\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.427969 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e0b4d4-eadc-456f-b9b1-30db806b94cc-config-data\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.428051 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsglh\" (UniqueName: \"kubernetes.io/projected/56e0b4d4-eadc-456f-b9b1-30db806b94cc-kube-api-access-wsglh\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.428068 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56e0b4d4-eadc-456f-b9b1-30db806b94cc-fernet-keys\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.434186 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56e0b4d4-eadc-456f-b9b1-30db806b94cc-credential-keys\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.434253 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e0b4d4-eadc-456f-b9b1-30db806b94cc-combined-ca-bundle\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.434567 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e0b4d4-eadc-456f-b9b1-30db806b94cc-config-data\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.436543 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56e0b4d4-eadc-456f-b9b1-30db806b94cc-fernet-keys\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.441187 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56e0b4d4-eadc-456f-b9b1-30db806b94cc-scripts\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.455413 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsglh\" (UniqueName: \"kubernetes.io/projected/56e0b4d4-eadc-456f-b9b1-30db806b94cc-kube-api-access-wsglh\") pod \"keystone-56858ddfd7-wzhwt\" (UID: \"56e0b4d4-eadc-456f-b9b1-30db806b94cc\") " pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.487814 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:16 crc kubenswrapper[4997]: I1205 08:51:16.923243 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-56858ddfd7-wzhwt"] Dec 05 08:51:16 crc kubenswrapper[4997]: W1205 08:51:16.927118 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56e0b4d4_eadc_456f_b9b1_30db806b94cc.slice/crio-d34e1d52e12fa031ec7fe6fcbbc2d2ddc5ef3902369abf7eec5c80348340c3b0 WatchSource:0}: Error finding container d34e1d52e12fa031ec7fe6fcbbc2d2ddc5ef3902369abf7eec5c80348340c3b0: Status 404 returned error can't find the container with id d34e1d52e12fa031ec7fe6fcbbc2d2ddc5ef3902369abf7eec5c80348340c3b0 Dec 05 08:51:17 crc kubenswrapper[4997]: I1205 08:51:17.070673 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-56858ddfd7-wzhwt" event={"ID":"56e0b4d4-eadc-456f-b9b1-30db806b94cc","Type":"ContainerStarted","Data":"d34e1d52e12fa031ec7fe6fcbbc2d2ddc5ef3902369abf7eec5c80348340c3b0"} Dec 05 08:51:17 crc kubenswrapper[4997]: I1205 08:51:17.073366 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t9jkl" event={"ID":"210251f1-d0c4-483e-ac78-20aa69438b25","Type":"ContainerStarted","Data":"ce2971e520e8310f5970d5b871bc894cd0a53542834b32df06945ac95911316d"} Dec 05 08:51:18 crc kubenswrapper[4997]: I1205 08:51:18.082581 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-56858ddfd7-wzhwt" event={"ID":"56e0b4d4-eadc-456f-b9b1-30db806b94cc","Type":"ContainerStarted","Data":"0454503480ce4c92bb22f1266364cad5622b81911143d4b09bcc73a822a99678"} Dec 05 08:51:18 crc kubenswrapper[4997]: I1205 08:51:18.108983 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-56858ddfd7-wzhwt" podStartSLOduration=2.108956282 podStartE2EDuration="2.108956282s" podCreationTimestamp="2025-12-05 08:51:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:51:18.099926268 +0000 UTC m=+6978.628833569" watchObservedRunningTime="2025-12-05 08:51:18.108956282 +0000 UTC m=+6978.637863563" Dec 05 08:51:19 crc kubenswrapper[4997]: I1205 08:51:19.091214 4997 generic.go:334] "Generic (PLEG): container finished" podID="210251f1-d0c4-483e-ac78-20aa69438b25" containerID="ce2971e520e8310f5970d5b871bc894cd0a53542834b32df06945ac95911316d" exitCode=0 Dec 05 08:51:19 crc kubenswrapper[4997]: I1205 08:51:19.091265 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t9jkl" event={"ID":"210251f1-d0c4-483e-ac78-20aa69438b25","Type":"ContainerDied","Data":"ce2971e520e8310f5970d5b871bc894cd0a53542834b32df06945ac95911316d"} Dec 05 08:51:19 crc kubenswrapper[4997]: I1205 08:51:19.091443 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:20 crc kubenswrapper[4997]: I1205 08:51:20.100055 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t9jkl" event={"ID":"210251f1-d0c4-483e-ac78-20aa69438b25","Type":"ContainerStarted","Data":"fb2ed47b661b5737edb974bb3594a98911f91488afca86bd7315def3c49499d0"} Dec 05 08:51:20 crc kubenswrapper[4997]: I1205 08:51:20.117605 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t9jkl" podStartSLOduration=2.51695247 podStartE2EDuration="6.117586936s" podCreationTimestamp="2025-12-05 08:51:14 +0000 UTC" firstStartedPulling="2025-12-05 08:51:16.057519209 +0000 UTC m=+6976.586426470" lastFinishedPulling="2025-12-05 08:51:19.658153675 +0000 UTC m=+6980.187060936" observedRunningTime="2025-12-05 08:51:20.117184154 +0000 UTC m=+6980.646091425" watchObservedRunningTime="2025-12-05 08:51:20.117586936 +0000 UTC m=+6980.646494217" Dec 05 08:51:23 crc kubenswrapper[4997]: I1205 08:51:23.749162 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:51:23 crc kubenswrapper[4997]: E1205 08:51:23.749718 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:51:24 crc kubenswrapper[4997]: I1205 08:51:24.370744 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t9jkl" Dec 05 08:51:24 crc kubenswrapper[4997]: I1205 08:51:24.370942 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t9jkl" Dec 05 08:51:25 crc kubenswrapper[4997]: I1205 08:51:25.412288 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-t9jkl" podUID="210251f1-d0c4-483e-ac78-20aa69438b25" containerName="registry-server" probeResult="failure" output=< Dec 05 08:51:25 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 08:51:25 crc kubenswrapper[4997]: > Dec 05 08:51:34 crc kubenswrapper[4997]: I1205 08:51:34.417264 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t9jkl" Dec 05 08:51:34 crc kubenswrapper[4997]: I1205 08:51:34.458733 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t9jkl" Dec 05 08:51:36 crc kubenswrapper[4997]: I1205 08:51:36.748803 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:51:36 crc kubenswrapper[4997]: E1205 08:51:36.749316 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:51:37 crc kubenswrapper[4997]: I1205 08:51:37.682605 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t9jkl"] Dec 05 08:51:37 crc kubenswrapper[4997]: I1205 08:51:37.682898 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t9jkl" podUID="210251f1-d0c4-483e-ac78-20aa69438b25" containerName="registry-server" containerID="cri-o://fb2ed47b661b5737edb974bb3594a98911f91488afca86bd7315def3c49499d0" gracePeriod=2 Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.196432 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t9jkl" Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.259855 4997 generic.go:334] "Generic (PLEG): container finished" podID="210251f1-d0c4-483e-ac78-20aa69438b25" containerID="fb2ed47b661b5737edb974bb3594a98911f91488afca86bd7315def3c49499d0" exitCode=0 Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.259926 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t9jkl" event={"ID":"210251f1-d0c4-483e-ac78-20aa69438b25","Type":"ContainerDied","Data":"fb2ed47b661b5737edb974bb3594a98911f91488afca86bd7315def3c49499d0"} Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.259948 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t9jkl" Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.259973 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t9jkl" event={"ID":"210251f1-d0c4-483e-ac78-20aa69438b25","Type":"ContainerDied","Data":"0595270acd252c47e8299f2919382ea91ac8b51152e8f06e4d6c84ed8b5959f2"} Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.259999 4997 scope.go:117] "RemoveContainer" containerID="fb2ed47b661b5737edb974bb3594a98911f91488afca86bd7315def3c49499d0" Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.278675 4997 scope.go:117] "RemoveContainer" containerID="ce2971e520e8310f5970d5b871bc894cd0a53542834b32df06945ac95911316d" Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.302504 4997 scope.go:117] "RemoveContainer" containerID="d62a40908a16fc0fbf94a28c808c698998b4f9abc84ccdbd9a61a14449256d81" Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.326297 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/210251f1-d0c4-483e-ac78-20aa69438b25-catalog-content\") pod \"210251f1-d0c4-483e-ac78-20aa69438b25\" (UID: \"210251f1-d0c4-483e-ac78-20aa69438b25\") " Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.326429 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/210251f1-d0c4-483e-ac78-20aa69438b25-utilities\") pod \"210251f1-d0c4-483e-ac78-20aa69438b25\" (UID: \"210251f1-d0c4-483e-ac78-20aa69438b25\") " Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.326559 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bn5k\" (UniqueName: \"kubernetes.io/projected/210251f1-d0c4-483e-ac78-20aa69438b25-kube-api-access-6bn5k\") pod \"210251f1-d0c4-483e-ac78-20aa69438b25\" (UID: \"210251f1-d0c4-483e-ac78-20aa69438b25\") " Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.335352 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210251f1-d0c4-483e-ac78-20aa69438b25-kube-api-access-6bn5k" (OuterVolumeSpecName: "kube-api-access-6bn5k") pod "210251f1-d0c4-483e-ac78-20aa69438b25" (UID: "210251f1-d0c4-483e-ac78-20aa69438b25"). InnerVolumeSpecName "kube-api-access-6bn5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.338446 4997 scope.go:117] "RemoveContainer" containerID="fb2ed47b661b5737edb974bb3594a98911f91488afca86bd7315def3c49499d0" Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.339429 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/210251f1-d0c4-483e-ac78-20aa69438b25-utilities" (OuterVolumeSpecName: "utilities") pod "210251f1-d0c4-483e-ac78-20aa69438b25" (UID: "210251f1-d0c4-483e-ac78-20aa69438b25"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:51:39 crc kubenswrapper[4997]: E1205 08:51:39.344790 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb2ed47b661b5737edb974bb3594a98911f91488afca86bd7315def3c49499d0\": container with ID starting with fb2ed47b661b5737edb974bb3594a98911f91488afca86bd7315def3c49499d0 not found: ID does not exist" containerID="fb2ed47b661b5737edb974bb3594a98911f91488afca86bd7315def3c49499d0" Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.344845 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb2ed47b661b5737edb974bb3594a98911f91488afca86bd7315def3c49499d0"} err="failed to get container status \"fb2ed47b661b5737edb974bb3594a98911f91488afca86bd7315def3c49499d0\": rpc error: code = NotFound desc = could not find container \"fb2ed47b661b5737edb974bb3594a98911f91488afca86bd7315def3c49499d0\": container with ID starting with fb2ed47b661b5737edb974bb3594a98911f91488afca86bd7315def3c49499d0 not found: ID does not exist" Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.344878 4997 scope.go:117] "RemoveContainer" containerID="ce2971e520e8310f5970d5b871bc894cd0a53542834b32df06945ac95911316d" Dec 05 08:51:39 crc kubenswrapper[4997]: E1205 08:51:39.345325 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce2971e520e8310f5970d5b871bc894cd0a53542834b32df06945ac95911316d\": container with ID starting with ce2971e520e8310f5970d5b871bc894cd0a53542834b32df06945ac95911316d not found: ID does not exist" containerID="ce2971e520e8310f5970d5b871bc894cd0a53542834b32df06945ac95911316d" Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.345386 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce2971e520e8310f5970d5b871bc894cd0a53542834b32df06945ac95911316d"} err="failed to get container status \"ce2971e520e8310f5970d5b871bc894cd0a53542834b32df06945ac95911316d\": rpc error: code = NotFound desc = could not find container \"ce2971e520e8310f5970d5b871bc894cd0a53542834b32df06945ac95911316d\": container with ID starting with ce2971e520e8310f5970d5b871bc894cd0a53542834b32df06945ac95911316d not found: ID does not exist" Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.345409 4997 scope.go:117] "RemoveContainer" containerID="d62a40908a16fc0fbf94a28c808c698998b4f9abc84ccdbd9a61a14449256d81" Dec 05 08:51:39 crc kubenswrapper[4997]: E1205 08:51:39.345833 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d62a40908a16fc0fbf94a28c808c698998b4f9abc84ccdbd9a61a14449256d81\": container with ID starting with d62a40908a16fc0fbf94a28c808c698998b4f9abc84ccdbd9a61a14449256d81 not found: ID does not exist" containerID="d62a40908a16fc0fbf94a28c808c698998b4f9abc84ccdbd9a61a14449256d81" Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.345878 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62a40908a16fc0fbf94a28c808c698998b4f9abc84ccdbd9a61a14449256d81"} err="failed to get container status \"d62a40908a16fc0fbf94a28c808c698998b4f9abc84ccdbd9a61a14449256d81\": rpc error: code = NotFound desc = could not find container \"d62a40908a16fc0fbf94a28c808c698998b4f9abc84ccdbd9a61a14449256d81\": container with ID starting with d62a40908a16fc0fbf94a28c808c698998b4f9abc84ccdbd9a61a14449256d81 not found: ID does not exist" Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.428758 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bn5k\" (UniqueName: \"kubernetes.io/projected/210251f1-d0c4-483e-ac78-20aa69438b25-kube-api-access-6bn5k\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.428802 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/210251f1-d0c4-483e-ac78-20aa69438b25-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.439775 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/210251f1-d0c4-483e-ac78-20aa69438b25-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "210251f1-d0c4-483e-ac78-20aa69438b25" (UID: "210251f1-d0c4-483e-ac78-20aa69438b25"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.530567 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/210251f1-d0c4-483e-ac78-20aa69438b25-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.594023 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t9jkl"] Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.602108 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t9jkl"] Dec 05 08:51:39 crc kubenswrapper[4997]: I1205 08:51:39.758250 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210251f1-d0c4-483e-ac78-20aa69438b25" path="/var/lib/kubelet/pods/210251f1-d0c4-483e-ac78-20aa69438b25/volumes" Dec 05 08:51:47 crc kubenswrapper[4997]: I1205 08:51:47.953443 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-56858ddfd7-wzhwt" Dec 05 08:51:49 crc kubenswrapper[4997]: I1205 08:51:49.753792 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:51:49 crc kubenswrapper[4997]: E1205 08:51:49.754554 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.415370 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 05 08:51:50 crc kubenswrapper[4997]: E1205 08:51:50.415900 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="210251f1-d0c4-483e-ac78-20aa69438b25" containerName="registry-server" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.415920 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="210251f1-d0c4-483e-ac78-20aa69438b25" containerName="registry-server" Dec 05 08:51:50 crc kubenswrapper[4997]: E1205 08:51:50.415933 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="210251f1-d0c4-483e-ac78-20aa69438b25" containerName="extract-utilities" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.415941 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="210251f1-d0c4-483e-ac78-20aa69438b25" containerName="extract-utilities" Dec 05 08:51:50 crc kubenswrapper[4997]: E1205 08:51:50.415978 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="210251f1-d0c4-483e-ac78-20aa69438b25" containerName="extract-content" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.415988 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="210251f1-d0c4-483e-ac78-20aa69438b25" containerName="extract-content" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.416263 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="210251f1-d0c4-483e-ac78-20aa69438b25" containerName="registry-server" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.416976 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.419108 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.419155 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.423890 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.434911 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-h4w6n" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.465261 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d8facf43-b820-4753-a695-34d7e7b89da9-openstack-config\") pod \"openstackclient\" (UID: \"d8facf43-b820-4753-a695-34d7e7b89da9\") " pod="openstack/openstackclient" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.465351 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8facf43-b820-4753-a695-34d7e7b89da9-openstack-config-secret\") pod \"openstackclient\" (UID: \"d8facf43-b820-4753-a695-34d7e7b89da9\") " pod="openstack/openstackclient" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.465439 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drn6r\" (UniqueName: \"kubernetes.io/projected/d8facf43-b820-4753-a695-34d7e7b89da9-kube-api-access-drn6r\") pod \"openstackclient\" (UID: \"d8facf43-b820-4753-a695-34d7e7b89da9\") " pod="openstack/openstackclient" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.566772 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d8facf43-b820-4753-a695-34d7e7b89da9-openstack-config\") pod \"openstackclient\" (UID: \"d8facf43-b820-4753-a695-34d7e7b89da9\") " pod="openstack/openstackclient" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.566853 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8facf43-b820-4753-a695-34d7e7b89da9-openstack-config-secret\") pod \"openstackclient\" (UID: \"d8facf43-b820-4753-a695-34d7e7b89da9\") " pod="openstack/openstackclient" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.566883 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drn6r\" (UniqueName: \"kubernetes.io/projected/d8facf43-b820-4753-a695-34d7e7b89da9-kube-api-access-drn6r\") pod \"openstackclient\" (UID: \"d8facf43-b820-4753-a695-34d7e7b89da9\") " pod="openstack/openstackclient" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.567957 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d8facf43-b820-4753-a695-34d7e7b89da9-openstack-config\") pod \"openstackclient\" (UID: \"d8facf43-b820-4753-a695-34d7e7b89da9\") " pod="openstack/openstackclient" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.576529 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8facf43-b820-4753-a695-34d7e7b89da9-openstack-config-secret\") pod \"openstackclient\" (UID: \"d8facf43-b820-4753-a695-34d7e7b89da9\") " pod="openstack/openstackclient" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.591470 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drn6r\" (UniqueName: \"kubernetes.io/projected/d8facf43-b820-4753-a695-34d7e7b89da9-kube-api-access-drn6r\") pod \"openstackclient\" (UID: \"d8facf43-b820-4753-a695-34d7e7b89da9\") " pod="openstack/openstackclient" Dec 05 08:51:50 crc kubenswrapper[4997]: I1205 08:51:50.736508 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 08:51:51 crc kubenswrapper[4997]: I1205 08:51:51.153851 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 08:51:51 crc kubenswrapper[4997]: I1205 08:51:51.363473 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"d8facf43-b820-4753-a695-34d7e7b89da9","Type":"ContainerStarted","Data":"987b78737d3a9aa3734d830d0a0219bd2e59490f284aba106b5e0a3d0d1e2eda"} Dec 05 08:52:01 crc kubenswrapper[4997]: I1205 08:52:01.749000 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:52:02 crc kubenswrapper[4997]: I1205 08:52:02.486636 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"c7247cb7e5929fb7303801e82d11af6c5303b8ce2c4cf98c220b4780c5e2c9f4"} Dec 05 08:52:02 crc kubenswrapper[4997]: I1205 08:52:02.488007 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"d8facf43-b820-4753-a695-34d7e7b89da9","Type":"ContainerStarted","Data":"ed0731958ed9cfb10278b3374d57d6772ac176a47911b4297300d3e8b2cd651d"} Dec 05 08:52:02 crc kubenswrapper[4997]: I1205 08:52:02.524802 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.7966267889999998 podStartE2EDuration="12.524785209s" podCreationTimestamp="2025-12-05 08:51:50 +0000 UTC" firstStartedPulling="2025-12-05 08:51:51.168009282 +0000 UTC m=+7011.696916543" lastFinishedPulling="2025-12-05 08:52:01.896167702 +0000 UTC m=+7022.425074963" observedRunningTime="2025-12-05 08:52:02.522403656 +0000 UTC m=+7023.051310937" watchObservedRunningTime="2025-12-05 08:52:02.524785209 +0000 UTC m=+7023.053692470" Dec 05 08:53:26 crc kubenswrapper[4997]: I1205 08:53:26.686467 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-vxnqk"] Dec 05 08:53:26 crc kubenswrapper[4997]: I1205 08:53:26.688061 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vxnqk" Dec 05 08:53:26 crc kubenswrapper[4997]: I1205 08:53:26.696731 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-vxnqk"] Dec 05 08:53:26 crc kubenswrapper[4997]: I1205 08:53:26.794368 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-c64a-account-create-update-whrd6"] Dec 05 08:53:26 crc kubenswrapper[4997]: I1205 08:53:26.796127 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c64a-account-create-update-whrd6" Dec 05 08:53:26 crc kubenswrapper[4997]: I1205 08:53:26.798733 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 05 08:53:26 crc kubenswrapper[4997]: I1205 08:53:26.809741 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6st6l\" (UniqueName: \"kubernetes.io/projected/0fa78f95-7c44-48ec-84e5-e34e33ccc0c3-kube-api-access-6st6l\") pod \"barbican-db-create-vxnqk\" (UID: \"0fa78f95-7c44-48ec-84e5-e34e33ccc0c3\") " pod="openstack/barbican-db-create-vxnqk" Dec 05 08:53:26 crc kubenswrapper[4997]: I1205 08:53:26.810171 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0fa78f95-7c44-48ec-84e5-e34e33ccc0c3-operator-scripts\") pod \"barbican-db-create-vxnqk\" (UID: \"0fa78f95-7c44-48ec-84e5-e34e33ccc0c3\") " pod="openstack/barbican-db-create-vxnqk" Dec 05 08:53:26 crc kubenswrapper[4997]: I1205 08:53:26.821071 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c64a-account-create-update-whrd6"] Dec 05 08:53:26 crc kubenswrapper[4997]: I1205 08:53:26.912541 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0fa78f95-7c44-48ec-84e5-e34e33ccc0c3-operator-scripts\") pod \"barbican-db-create-vxnqk\" (UID: \"0fa78f95-7c44-48ec-84e5-e34e33ccc0c3\") " pod="openstack/barbican-db-create-vxnqk" Dec 05 08:53:26 crc kubenswrapper[4997]: I1205 08:53:26.912715 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd-operator-scripts\") pod \"barbican-c64a-account-create-update-whrd6\" (UID: \"da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd\") " pod="openstack/barbican-c64a-account-create-update-whrd6" Dec 05 08:53:26 crc kubenswrapper[4997]: I1205 08:53:26.912774 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6st6l\" (UniqueName: \"kubernetes.io/projected/0fa78f95-7c44-48ec-84e5-e34e33ccc0c3-kube-api-access-6st6l\") pod \"barbican-db-create-vxnqk\" (UID: \"0fa78f95-7c44-48ec-84e5-e34e33ccc0c3\") " pod="openstack/barbican-db-create-vxnqk" Dec 05 08:53:26 crc kubenswrapper[4997]: I1205 08:53:26.912824 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxn65\" (UniqueName: \"kubernetes.io/projected/da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd-kube-api-access-wxn65\") pod \"barbican-c64a-account-create-update-whrd6\" (UID: \"da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd\") " pod="openstack/barbican-c64a-account-create-update-whrd6" Dec 05 08:53:26 crc kubenswrapper[4997]: I1205 08:53:26.913330 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0fa78f95-7c44-48ec-84e5-e34e33ccc0c3-operator-scripts\") pod \"barbican-db-create-vxnqk\" (UID: \"0fa78f95-7c44-48ec-84e5-e34e33ccc0c3\") " pod="openstack/barbican-db-create-vxnqk" Dec 05 08:53:26 crc kubenswrapper[4997]: I1205 08:53:26.940220 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6st6l\" (UniqueName: \"kubernetes.io/projected/0fa78f95-7c44-48ec-84e5-e34e33ccc0c3-kube-api-access-6st6l\") pod \"barbican-db-create-vxnqk\" (UID: \"0fa78f95-7c44-48ec-84e5-e34e33ccc0c3\") " pod="openstack/barbican-db-create-vxnqk" Dec 05 08:53:27 crc kubenswrapper[4997]: I1205 08:53:27.010093 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vxnqk" Dec 05 08:53:27 crc kubenswrapper[4997]: I1205 08:53:27.014673 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd-operator-scripts\") pod \"barbican-c64a-account-create-update-whrd6\" (UID: \"da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd\") " pod="openstack/barbican-c64a-account-create-update-whrd6" Dec 05 08:53:27 crc kubenswrapper[4997]: I1205 08:53:27.014725 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxn65\" (UniqueName: \"kubernetes.io/projected/da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd-kube-api-access-wxn65\") pod \"barbican-c64a-account-create-update-whrd6\" (UID: \"da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd\") " pod="openstack/barbican-c64a-account-create-update-whrd6" Dec 05 08:53:27 crc kubenswrapper[4997]: I1205 08:53:27.015379 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd-operator-scripts\") pod \"barbican-c64a-account-create-update-whrd6\" (UID: \"da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd\") " pod="openstack/barbican-c64a-account-create-update-whrd6" Dec 05 08:53:27 crc kubenswrapper[4997]: I1205 08:53:27.030824 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxn65\" (UniqueName: \"kubernetes.io/projected/da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd-kube-api-access-wxn65\") pod \"barbican-c64a-account-create-update-whrd6\" (UID: \"da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd\") " pod="openstack/barbican-c64a-account-create-update-whrd6" Dec 05 08:53:27 crc kubenswrapper[4997]: I1205 08:53:27.126711 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c64a-account-create-update-whrd6" Dec 05 08:53:27 crc kubenswrapper[4997]: I1205 08:53:27.258844 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-vxnqk"] Dec 05 08:53:27 crc kubenswrapper[4997]: I1205 08:53:27.364216 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c64a-account-create-update-whrd6"] Dec 05 08:53:28 crc kubenswrapper[4997]: I1205 08:53:28.252207 4997 generic.go:334] "Generic (PLEG): container finished" podID="0fa78f95-7c44-48ec-84e5-e34e33ccc0c3" containerID="58bd0e4ffef3300c91df54143cbfffb7d5044628d0d6372254962c416b7e6506" exitCode=0 Dec 05 08:53:28 crc kubenswrapper[4997]: I1205 08:53:28.252448 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-vxnqk" event={"ID":"0fa78f95-7c44-48ec-84e5-e34e33ccc0c3","Type":"ContainerDied","Data":"58bd0e4ffef3300c91df54143cbfffb7d5044628d0d6372254962c416b7e6506"} Dec 05 08:53:28 crc kubenswrapper[4997]: I1205 08:53:28.252738 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-vxnqk" event={"ID":"0fa78f95-7c44-48ec-84e5-e34e33ccc0c3","Type":"ContainerStarted","Data":"c9a9307cd48f4372029262542ecd021aad26fba108dfe6a98a798e72d205812e"} Dec 05 08:53:28 crc kubenswrapper[4997]: I1205 08:53:28.255378 4997 generic.go:334] "Generic (PLEG): container finished" podID="da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd" containerID="f76545c8d04a765c1ffec4d2255263ce7b7a54d45a8fca50f1cb6e5bee8214c0" exitCode=0 Dec 05 08:53:28 crc kubenswrapper[4997]: I1205 08:53:28.255439 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c64a-account-create-update-whrd6" event={"ID":"da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd","Type":"ContainerDied","Data":"f76545c8d04a765c1ffec4d2255263ce7b7a54d45a8fca50f1cb6e5bee8214c0"} Dec 05 08:53:28 crc kubenswrapper[4997]: I1205 08:53:28.255472 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c64a-account-create-update-whrd6" event={"ID":"da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd","Type":"ContainerStarted","Data":"579160360f9a18fd7c041074afcd018858891ae1507ae5c1abcd64355b357a3e"} Dec 05 08:53:29 crc kubenswrapper[4997]: I1205 08:53:29.617860 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c64a-account-create-update-whrd6" Dec 05 08:53:29 crc kubenswrapper[4997]: I1205 08:53:29.628812 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vxnqk" Dec 05 08:53:29 crc kubenswrapper[4997]: I1205 08:53:29.765853 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd-operator-scripts\") pod \"da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd\" (UID: \"da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd\") " Dec 05 08:53:29 crc kubenswrapper[4997]: I1205 08:53:29.765911 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6st6l\" (UniqueName: \"kubernetes.io/projected/0fa78f95-7c44-48ec-84e5-e34e33ccc0c3-kube-api-access-6st6l\") pod \"0fa78f95-7c44-48ec-84e5-e34e33ccc0c3\" (UID: \"0fa78f95-7c44-48ec-84e5-e34e33ccc0c3\") " Dec 05 08:53:29 crc kubenswrapper[4997]: I1205 08:53:29.765979 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxn65\" (UniqueName: \"kubernetes.io/projected/da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd-kube-api-access-wxn65\") pod \"da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd\" (UID: \"da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd\") " Dec 05 08:53:29 crc kubenswrapper[4997]: I1205 08:53:29.766010 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0fa78f95-7c44-48ec-84e5-e34e33ccc0c3-operator-scripts\") pod \"0fa78f95-7c44-48ec-84e5-e34e33ccc0c3\" (UID: \"0fa78f95-7c44-48ec-84e5-e34e33ccc0c3\") " Dec 05 08:53:29 crc kubenswrapper[4997]: I1205 08:53:29.766289 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd" (UID: "da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:53:29 crc kubenswrapper[4997]: I1205 08:53:29.766566 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fa78f95-7c44-48ec-84e5-e34e33ccc0c3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0fa78f95-7c44-48ec-84e5-e34e33ccc0c3" (UID: "0fa78f95-7c44-48ec-84e5-e34e33ccc0c3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:53:29 crc kubenswrapper[4997]: I1205 08:53:29.772850 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fa78f95-7c44-48ec-84e5-e34e33ccc0c3-kube-api-access-6st6l" (OuterVolumeSpecName: "kube-api-access-6st6l") pod "0fa78f95-7c44-48ec-84e5-e34e33ccc0c3" (UID: "0fa78f95-7c44-48ec-84e5-e34e33ccc0c3"). InnerVolumeSpecName "kube-api-access-6st6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:53:29 crc kubenswrapper[4997]: I1205 08:53:29.772958 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd-kube-api-access-wxn65" (OuterVolumeSpecName: "kube-api-access-wxn65") pod "da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd" (UID: "da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd"). InnerVolumeSpecName "kube-api-access-wxn65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:53:29 crc kubenswrapper[4997]: I1205 08:53:29.867738 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:53:29 crc kubenswrapper[4997]: I1205 08:53:29.867781 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6st6l\" (UniqueName: \"kubernetes.io/projected/0fa78f95-7c44-48ec-84e5-e34e33ccc0c3-kube-api-access-6st6l\") on node \"crc\" DevicePath \"\"" Dec 05 08:53:29 crc kubenswrapper[4997]: I1205 08:53:29.867796 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxn65\" (UniqueName: \"kubernetes.io/projected/da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd-kube-api-access-wxn65\") on node \"crc\" DevicePath \"\"" Dec 05 08:53:29 crc kubenswrapper[4997]: I1205 08:53:29.867805 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0fa78f95-7c44-48ec-84e5-e34e33ccc0c3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:53:30 crc kubenswrapper[4997]: I1205 08:53:30.278720 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-vxnqk" event={"ID":"0fa78f95-7c44-48ec-84e5-e34e33ccc0c3","Type":"ContainerDied","Data":"c9a9307cd48f4372029262542ecd021aad26fba108dfe6a98a798e72d205812e"} Dec 05 08:53:30 crc kubenswrapper[4997]: I1205 08:53:30.278738 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vxnqk" Dec 05 08:53:30 crc kubenswrapper[4997]: I1205 08:53:30.278759 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9a9307cd48f4372029262542ecd021aad26fba108dfe6a98a798e72d205812e" Dec 05 08:53:30 crc kubenswrapper[4997]: I1205 08:53:30.289309 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c64a-account-create-update-whrd6" event={"ID":"da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd","Type":"ContainerDied","Data":"579160360f9a18fd7c041074afcd018858891ae1507ae5c1abcd64355b357a3e"} Dec 05 08:53:30 crc kubenswrapper[4997]: I1205 08:53:30.289368 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="579160360f9a18fd7c041074afcd018858891ae1507ae5c1abcd64355b357a3e" Dec 05 08:53:30 crc kubenswrapper[4997]: I1205 08:53:30.289475 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c64a-account-create-update-whrd6" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.162357 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-fmhcf"] Dec 05 08:53:32 crc kubenswrapper[4997]: E1205 08:53:32.163127 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd" containerName="mariadb-account-create-update" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.163145 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd" containerName="mariadb-account-create-update" Dec 05 08:53:32 crc kubenswrapper[4997]: E1205 08:53:32.163155 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fa78f95-7c44-48ec-84e5-e34e33ccc0c3" containerName="mariadb-database-create" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.163165 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fa78f95-7c44-48ec-84e5-e34e33ccc0c3" containerName="mariadb-database-create" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.163392 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd" containerName="mariadb-account-create-update" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.163414 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fa78f95-7c44-48ec-84e5-e34e33ccc0c3" containerName="mariadb-database-create" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.164113 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-fmhcf" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.165797 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.166665 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-5nsj2" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.174688 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-fmhcf"] Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.305689 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/797fa92f-8010-4e29-9a0d-9c3045136c25-combined-ca-bundle\") pod \"barbican-db-sync-fmhcf\" (UID: \"797fa92f-8010-4e29-9a0d-9c3045136c25\") " pod="openstack/barbican-db-sync-fmhcf" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.305741 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/797fa92f-8010-4e29-9a0d-9c3045136c25-db-sync-config-data\") pod \"barbican-db-sync-fmhcf\" (UID: \"797fa92f-8010-4e29-9a0d-9c3045136c25\") " pod="openstack/barbican-db-sync-fmhcf" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.305767 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpzsd\" (UniqueName: \"kubernetes.io/projected/797fa92f-8010-4e29-9a0d-9c3045136c25-kube-api-access-kpzsd\") pod \"barbican-db-sync-fmhcf\" (UID: \"797fa92f-8010-4e29-9a0d-9c3045136c25\") " pod="openstack/barbican-db-sync-fmhcf" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.406860 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/797fa92f-8010-4e29-9a0d-9c3045136c25-combined-ca-bundle\") pod \"barbican-db-sync-fmhcf\" (UID: \"797fa92f-8010-4e29-9a0d-9c3045136c25\") " pod="openstack/barbican-db-sync-fmhcf" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.406928 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/797fa92f-8010-4e29-9a0d-9c3045136c25-db-sync-config-data\") pod \"barbican-db-sync-fmhcf\" (UID: \"797fa92f-8010-4e29-9a0d-9c3045136c25\") " pod="openstack/barbican-db-sync-fmhcf" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.406959 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpzsd\" (UniqueName: \"kubernetes.io/projected/797fa92f-8010-4e29-9a0d-9c3045136c25-kube-api-access-kpzsd\") pod \"barbican-db-sync-fmhcf\" (UID: \"797fa92f-8010-4e29-9a0d-9c3045136c25\") " pod="openstack/barbican-db-sync-fmhcf" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.420361 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/797fa92f-8010-4e29-9a0d-9c3045136c25-db-sync-config-data\") pod \"barbican-db-sync-fmhcf\" (UID: \"797fa92f-8010-4e29-9a0d-9c3045136c25\") " pod="openstack/barbican-db-sync-fmhcf" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.420406 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/797fa92f-8010-4e29-9a0d-9c3045136c25-combined-ca-bundle\") pod \"barbican-db-sync-fmhcf\" (UID: \"797fa92f-8010-4e29-9a0d-9c3045136c25\") " pod="openstack/barbican-db-sync-fmhcf" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.427215 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpzsd\" (UniqueName: \"kubernetes.io/projected/797fa92f-8010-4e29-9a0d-9c3045136c25-kube-api-access-kpzsd\") pod \"barbican-db-sync-fmhcf\" (UID: \"797fa92f-8010-4e29-9a0d-9c3045136c25\") " pod="openstack/barbican-db-sync-fmhcf" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.514837 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-fmhcf" Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.958013 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-fmhcf"] Dec 05 08:53:32 crc kubenswrapper[4997]: I1205 08:53:32.967341 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 08:53:33 crc kubenswrapper[4997]: I1205 08:53:33.312582 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-fmhcf" event={"ID":"797fa92f-8010-4e29-9a0d-9c3045136c25","Type":"ContainerStarted","Data":"94e84f5dbb3aa48db05c439958f70066b9815d78da05ced0fa735fc04a5d7d47"} Dec 05 08:53:37 crc kubenswrapper[4997]: I1205 08:53:37.346662 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-fmhcf" event={"ID":"797fa92f-8010-4e29-9a0d-9c3045136c25","Type":"ContainerStarted","Data":"70df3d45cf80aa0a7179ad46c24d0545895732c981a5c6bbcb27a8ee215be892"} Dec 05 08:53:37 crc kubenswrapper[4997]: I1205 08:53:37.367422 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-fmhcf" podStartSLOduration=1.427830146 podStartE2EDuration="5.367402925s" podCreationTimestamp="2025-12-05 08:53:32 +0000 UTC" firstStartedPulling="2025-12-05 08:53:32.967088591 +0000 UTC m=+7113.495995852" lastFinishedPulling="2025-12-05 08:53:36.90666137 +0000 UTC m=+7117.435568631" observedRunningTime="2025-12-05 08:53:37.365982946 +0000 UTC m=+7117.894890217" watchObservedRunningTime="2025-12-05 08:53:37.367402925 +0000 UTC m=+7117.896310206" Dec 05 08:53:42 crc kubenswrapper[4997]: I1205 08:53:42.390773 4997 generic.go:334] "Generic (PLEG): container finished" podID="797fa92f-8010-4e29-9a0d-9c3045136c25" containerID="70df3d45cf80aa0a7179ad46c24d0545895732c981a5c6bbcb27a8ee215be892" exitCode=0 Dec 05 08:53:42 crc kubenswrapper[4997]: I1205 08:53:42.390834 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-fmhcf" event={"ID":"797fa92f-8010-4e29-9a0d-9c3045136c25","Type":"ContainerDied","Data":"70df3d45cf80aa0a7179ad46c24d0545895732c981a5c6bbcb27a8ee215be892"} Dec 05 08:53:43 crc kubenswrapper[4997]: I1205 08:53:43.718986 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-fmhcf" Dec 05 08:53:43 crc kubenswrapper[4997]: I1205 08:53:43.790797 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/797fa92f-8010-4e29-9a0d-9c3045136c25-db-sync-config-data\") pod \"797fa92f-8010-4e29-9a0d-9c3045136c25\" (UID: \"797fa92f-8010-4e29-9a0d-9c3045136c25\") " Dec 05 08:53:43 crc kubenswrapper[4997]: I1205 08:53:43.790858 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpzsd\" (UniqueName: \"kubernetes.io/projected/797fa92f-8010-4e29-9a0d-9c3045136c25-kube-api-access-kpzsd\") pod \"797fa92f-8010-4e29-9a0d-9c3045136c25\" (UID: \"797fa92f-8010-4e29-9a0d-9c3045136c25\") " Dec 05 08:53:43 crc kubenswrapper[4997]: I1205 08:53:43.790991 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/797fa92f-8010-4e29-9a0d-9c3045136c25-combined-ca-bundle\") pod \"797fa92f-8010-4e29-9a0d-9c3045136c25\" (UID: \"797fa92f-8010-4e29-9a0d-9c3045136c25\") " Dec 05 08:53:43 crc kubenswrapper[4997]: I1205 08:53:43.796227 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/797fa92f-8010-4e29-9a0d-9c3045136c25-kube-api-access-kpzsd" (OuterVolumeSpecName: "kube-api-access-kpzsd") pod "797fa92f-8010-4e29-9a0d-9c3045136c25" (UID: "797fa92f-8010-4e29-9a0d-9c3045136c25"). InnerVolumeSpecName "kube-api-access-kpzsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:53:43 crc kubenswrapper[4997]: I1205 08:53:43.796338 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797fa92f-8010-4e29-9a0d-9c3045136c25-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "797fa92f-8010-4e29-9a0d-9c3045136c25" (UID: "797fa92f-8010-4e29-9a0d-9c3045136c25"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:53:43 crc kubenswrapper[4997]: I1205 08:53:43.813269 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797fa92f-8010-4e29-9a0d-9c3045136c25-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "797fa92f-8010-4e29-9a0d-9c3045136c25" (UID: "797fa92f-8010-4e29-9a0d-9c3045136c25"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:53:43 crc kubenswrapper[4997]: I1205 08:53:43.892703 4997 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/797fa92f-8010-4e29-9a0d-9c3045136c25-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:53:43 crc kubenswrapper[4997]: I1205 08:53:43.892758 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpzsd\" (UniqueName: \"kubernetes.io/projected/797fa92f-8010-4e29-9a0d-9c3045136c25-kube-api-access-kpzsd\") on node \"crc\" DevicePath \"\"" Dec 05 08:53:43 crc kubenswrapper[4997]: I1205 08:53:43.892770 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/797fa92f-8010-4e29-9a0d-9c3045136c25-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.409389 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-fmhcf" event={"ID":"797fa92f-8010-4e29-9a0d-9c3045136c25","Type":"ContainerDied","Data":"94e84f5dbb3aa48db05c439958f70066b9815d78da05ced0fa735fc04a5d7d47"} Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.409739 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94e84f5dbb3aa48db05c439958f70066b9815d78da05ced0fa735fc04a5d7d47" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.409507 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-fmhcf" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.618712 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5756b5f4c9-wj8k4"] Dec 05 08:53:44 crc kubenswrapper[4997]: E1205 08:53:44.619116 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="797fa92f-8010-4e29-9a0d-9c3045136c25" containerName="barbican-db-sync" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.619133 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="797fa92f-8010-4e29-9a0d-9c3045136c25" containerName="barbican-db-sync" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.619362 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="797fa92f-8010-4e29-9a0d-9c3045136c25" containerName="barbican-db-sync" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.620462 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5756b5f4c9-wj8k4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.622838 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.623477 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-5nsj2" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.623720 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.646153 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5756b5f4c9-wj8k4"] Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.660261 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-77b797d5d4-vh4h4"] Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.685421 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.704693 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.707335 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22mws\" (UniqueName: \"kubernetes.io/projected/0666a02f-d3ed-4403-b2af-5450e2901d82-kube-api-access-22mws\") pod \"barbican-worker-5756b5f4c9-wj8k4\" (UID: \"0666a02f-d3ed-4403-b2af-5450e2901d82\") " pod="openstack/barbican-worker-5756b5f4c9-wj8k4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.708856 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0666a02f-d3ed-4403-b2af-5450e2901d82-combined-ca-bundle\") pod \"barbican-worker-5756b5f4c9-wj8k4\" (UID: \"0666a02f-d3ed-4403-b2af-5450e2901d82\") " pod="openstack/barbican-worker-5756b5f4c9-wj8k4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.709121 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0666a02f-d3ed-4403-b2af-5450e2901d82-config-data-custom\") pod \"barbican-worker-5756b5f4c9-wj8k4\" (UID: \"0666a02f-d3ed-4403-b2af-5450e2901d82\") " pod="openstack/barbican-worker-5756b5f4c9-wj8k4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.709207 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0666a02f-d3ed-4403-b2af-5450e2901d82-config-data\") pod \"barbican-worker-5756b5f4c9-wj8k4\" (UID: \"0666a02f-d3ed-4403-b2af-5450e2901d82\") " pod="openstack/barbican-worker-5756b5f4c9-wj8k4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.709248 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0666a02f-d3ed-4403-b2af-5450e2901d82-logs\") pod \"barbican-worker-5756b5f4c9-wj8k4\" (UID: \"0666a02f-d3ed-4403-b2af-5450e2901d82\") " pod="openstack/barbican-worker-5756b5f4c9-wj8k4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.721201 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-77b797d5d4-vh4h4"] Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.760550 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d79c65b4f-pfwn8"] Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.762322 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.768529 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d79c65b4f-pfwn8"] Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.810423 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c326c07-9fee-4b86-b7c1-bd9ce32c95f3-config-data\") pod \"barbican-keystone-listener-77b797d5d4-vh4h4\" (UID: \"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3\") " pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.810468 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22mws\" (UniqueName: \"kubernetes.io/projected/0666a02f-d3ed-4403-b2af-5450e2901d82-kube-api-access-22mws\") pod \"barbican-worker-5756b5f4c9-wj8k4\" (UID: \"0666a02f-d3ed-4403-b2af-5450e2901d82\") " pod="openstack/barbican-worker-5756b5f4c9-wj8k4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.810516 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c326c07-9fee-4b86-b7c1-bd9ce32c95f3-config-data-custom\") pod \"barbican-keystone-listener-77b797d5d4-vh4h4\" (UID: \"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3\") " pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.810549 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0666a02f-d3ed-4403-b2af-5450e2901d82-combined-ca-bundle\") pod \"barbican-worker-5756b5f4c9-wj8k4\" (UID: \"0666a02f-d3ed-4403-b2af-5450e2901d82\") " pod="openstack/barbican-worker-5756b5f4c9-wj8k4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.810593 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c326c07-9fee-4b86-b7c1-bd9ce32c95f3-combined-ca-bundle\") pod \"barbican-keystone-listener-77b797d5d4-vh4h4\" (UID: \"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3\") " pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.810655 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0666a02f-d3ed-4403-b2af-5450e2901d82-config-data-custom\") pod \"barbican-worker-5756b5f4c9-wj8k4\" (UID: \"0666a02f-d3ed-4403-b2af-5450e2901d82\") " pod="openstack/barbican-worker-5756b5f4c9-wj8k4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.810674 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c326c07-9fee-4b86-b7c1-bd9ce32c95f3-logs\") pod \"barbican-keystone-listener-77b797d5d4-vh4h4\" (UID: \"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3\") " pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.810707 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57vnt\" (UniqueName: \"kubernetes.io/projected/6c326c07-9fee-4b86-b7c1-bd9ce32c95f3-kube-api-access-57vnt\") pod \"barbican-keystone-listener-77b797d5d4-vh4h4\" (UID: \"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3\") " pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.810725 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0666a02f-d3ed-4403-b2af-5450e2901d82-config-data\") pod \"barbican-worker-5756b5f4c9-wj8k4\" (UID: \"0666a02f-d3ed-4403-b2af-5450e2901d82\") " pod="openstack/barbican-worker-5756b5f4c9-wj8k4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.810744 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0666a02f-d3ed-4403-b2af-5450e2901d82-logs\") pod \"barbican-worker-5756b5f4c9-wj8k4\" (UID: \"0666a02f-d3ed-4403-b2af-5450e2901d82\") " pod="openstack/barbican-worker-5756b5f4c9-wj8k4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.811253 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0666a02f-d3ed-4403-b2af-5450e2901d82-logs\") pod \"barbican-worker-5756b5f4c9-wj8k4\" (UID: \"0666a02f-d3ed-4403-b2af-5450e2901d82\") " pod="openstack/barbican-worker-5756b5f4c9-wj8k4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.812286 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-9f58cdf4d-fz24c"] Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.813851 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.816417 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0666a02f-d3ed-4403-b2af-5450e2901d82-config-data\") pod \"barbican-worker-5756b5f4c9-wj8k4\" (UID: \"0666a02f-d3ed-4403-b2af-5450e2901d82\") " pod="openstack/barbican-worker-5756b5f4c9-wj8k4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.816501 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.816912 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0666a02f-d3ed-4403-b2af-5450e2901d82-config-data-custom\") pod \"barbican-worker-5756b5f4c9-wj8k4\" (UID: \"0666a02f-d3ed-4403-b2af-5450e2901d82\") " pod="openstack/barbican-worker-5756b5f4c9-wj8k4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.835261 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0666a02f-d3ed-4403-b2af-5450e2901d82-combined-ca-bundle\") pod \"barbican-worker-5756b5f4c9-wj8k4\" (UID: \"0666a02f-d3ed-4403-b2af-5450e2901d82\") " pod="openstack/barbican-worker-5756b5f4c9-wj8k4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.837825 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9f58cdf4d-fz24c"] Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.842607 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22mws\" (UniqueName: \"kubernetes.io/projected/0666a02f-d3ed-4403-b2af-5450e2901d82-kube-api-access-22mws\") pod \"barbican-worker-5756b5f4c9-wj8k4\" (UID: \"0666a02f-d3ed-4403-b2af-5450e2901d82\") " pod="openstack/barbican-worker-5756b5f4c9-wj8k4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.912590 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-ovsdbserver-sb\") pod \"dnsmasq-dns-d79c65b4f-pfwn8\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.912671 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnw6f\" (UniqueName: \"kubernetes.io/projected/f762f277-36aa-4ed7-8244-374a0481c789-kube-api-access-wnw6f\") pod \"dnsmasq-dns-d79c65b4f-pfwn8\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.912698 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-config\") pod \"dnsmasq-dns-d79c65b4f-pfwn8\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.912751 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c326c07-9fee-4b86-b7c1-bd9ce32c95f3-combined-ca-bundle\") pod \"barbican-keystone-listener-77b797d5d4-vh4h4\" (UID: \"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3\") " pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.913174 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-dns-svc\") pod \"dnsmasq-dns-d79c65b4f-pfwn8\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.913368 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08bdbc96-88c5-4744-bc48-bc2d232a8d6a-config-data\") pod \"barbican-api-9f58cdf4d-fz24c\" (UID: \"08bdbc96-88c5-4744-bc48-bc2d232a8d6a\") " pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.913461 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08bdbc96-88c5-4744-bc48-bc2d232a8d6a-combined-ca-bundle\") pod \"barbican-api-9f58cdf4d-fz24c\" (UID: \"08bdbc96-88c5-4744-bc48-bc2d232a8d6a\") " pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.913564 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c326c07-9fee-4b86-b7c1-bd9ce32c95f3-logs\") pod \"barbican-keystone-listener-77b797d5d4-vh4h4\" (UID: \"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3\") " pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.913599 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/08bdbc96-88c5-4744-bc48-bc2d232a8d6a-config-data-custom\") pod \"barbican-api-9f58cdf4d-fz24c\" (UID: \"08bdbc96-88c5-4744-bc48-bc2d232a8d6a\") " pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.913641 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-ovsdbserver-nb\") pod \"dnsmasq-dns-d79c65b4f-pfwn8\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.913785 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57vnt\" (UniqueName: \"kubernetes.io/projected/6c326c07-9fee-4b86-b7c1-bd9ce32c95f3-kube-api-access-57vnt\") pod \"barbican-keystone-listener-77b797d5d4-vh4h4\" (UID: \"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3\") " pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.913814 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08bdbc96-88c5-4744-bc48-bc2d232a8d6a-logs\") pod \"barbican-api-9f58cdf4d-fz24c\" (UID: \"08bdbc96-88c5-4744-bc48-bc2d232a8d6a\") " pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.914270 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zj4s\" (UniqueName: \"kubernetes.io/projected/08bdbc96-88c5-4744-bc48-bc2d232a8d6a-kube-api-access-5zj4s\") pod \"barbican-api-9f58cdf4d-fz24c\" (UID: \"08bdbc96-88c5-4744-bc48-bc2d232a8d6a\") " pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.913959 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c326c07-9fee-4b86-b7c1-bd9ce32c95f3-logs\") pod \"barbican-keystone-listener-77b797d5d4-vh4h4\" (UID: \"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3\") " pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.914420 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c326c07-9fee-4b86-b7c1-bd9ce32c95f3-config-data\") pod \"barbican-keystone-listener-77b797d5d4-vh4h4\" (UID: \"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3\") " pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.914479 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c326c07-9fee-4b86-b7c1-bd9ce32c95f3-config-data-custom\") pod \"barbican-keystone-listener-77b797d5d4-vh4h4\" (UID: \"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3\") " pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.917249 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c326c07-9fee-4b86-b7c1-bd9ce32c95f3-combined-ca-bundle\") pod \"barbican-keystone-listener-77b797d5d4-vh4h4\" (UID: \"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3\") " pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.918765 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c326c07-9fee-4b86-b7c1-bd9ce32c95f3-config-data\") pod \"barbican-keystone-listener-77b797d5d4-vh4h4\" (UID: \"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3\") " pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.921584 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c326c07-9fee-4b86-b7c1-bd9ce32c95f3-config-data-custom\") pod \"barbican-keystone-listener-77b797d5d4-vh4h4\" (UID: \"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3\") " pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.933154 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57vnt\" (UniqueName: \"kubernetes.io/projected/6c326c07-9fee-4b86-b7c1-bd9ce32c95f3-kube-api-access-57vnt\") pod \"barbican-keystone-listener-77b797d5d4-vh4h4\" (UID: \"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3\") " pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" Dec 05 08:53:44 crc kubenswrapper[4997]: I1205 08:53:44.946447 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5756b5f4c9-wj8k4" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.016812 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08bdbc96-88c5-4744-bc48-bc2d232a8d6a-logs\") pod \"barbican-api-9f58cdf4d-fz24c\" (UID: \"08bdbc96-88c5-4744-bc48-bc2d232a8d6a\") " pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.016901 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zj4s\" (UniqueName: \"kubernetes.io/projected/08bdbc96-88c5-4744-bc48-bc2d232a8d6a-kube-api-access-5zj4s\") pod \"barbican-api-9f58cdf4d-fz24c\" (UID: \"08bdbc96-88c5-4744-bc48-bc2d232a8d6a\") " pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.016982 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-ovsdbserver-sb\") pod \"dnsmasq-dns-d79c65b4f-pfwn8\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.017013 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnw6f\" (UniqueName: \"kubernetes.io/projected/f762f277-36aa-4ed7-8244-374a0481c789-kube-api-access-wnw6f\") pod \"dnsmasq-dns-d79c65b4f-pfwn8\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.017035 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-config\") pod \"dnsmasq-dns-d79c65b4f-pfwn8\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.017116 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-dns-svc\") pod \"dnsmasq-dns-d79c65b4f-pfwn8\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.017144 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08bdbc96-88c5-4744-bc48-bc2d232a8d6a-config-data\") pod \"barbican-api-9f58cdf4d-fz24c\" (UID: \"08bdbc96-88c5-4744-bc48-bc2d232a8d6a\") " pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.017181 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08bdbc96-88c5-4744-bc48-bc2d232a8d6a-combined-ca-bundle\") pod \"barbican-api-9f58cdf4d-fz24c\" (UID: \"08bdbc96-88c5-4744-bc48-bc2d232a8d6a\") " pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.017201 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08bdbc96-88c5-4744-bc48-bc2d232a8d6a-logs\") pod \"barbican-api-9f58cdf4d-fz24c\" (UID: \"08bdbc96-88c5-4744-bc48-bc2d232a8d6a\") " pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.017222 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/08bdbc96-88c5-4744-bc48-bc2d232a8d6a-config-data-custom\") pod \"barbican-api-9f58cdf4d-fz24c\" (UID: \"08bdbc96-88c5-4744-bc48-bc2d232a8d6a\") " pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.017274 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-ovsdbserver-nb\") pod \"dnsmasq-dns-d79c65b4f-pfwn8\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.017924 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-config\") pod \"dnsmasq-dns-d79c65b4f-pfwn8\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.018172 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-ovsdbserver-nb\") pod \"dnsmasq-dns-d79c65b4f-pfwn8\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.018433 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-ovsdbserver-sb\") pod \"dnsmasq-dns-d79c65b4f-pfwn8\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.018903 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-dns-svc\") pod \"dnsmasq-dns-d79c65b4f-pfwn8\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.023640 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08bdbc96-88c5-4744-bc48-bc2d232a8d6a-config-data\") pod \"barbican-api-9f58cdf4d-fz24c\" (UID: \"08bdbc96-88c5-4744-bc48-bc2d232a8d6a\") " pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.025788 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.026934 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/08bdbc96-88c5-4744-bc48-bc2d232a8d6a-config-data-custom\") pod \"barbican-api-9f58cdf4d-fz24c\" (UID: \"08bdbc96-88c5-4744-bc48-bc2d232a8d6a\") " pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.027468 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08bdbc96-88c5-4744-bc48-bc2d232a8d6a-combined-ca-bundle\") pod \"barbican-api-9f58cdf4d-fz24c\" (UID: \"08bdbc96-88c5-4744-bc48-bc2d232a8d6a\") " pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.039756 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zj4s\" (UniqueName: \"kubernetes.io/projected/08bdbc96-88c5-4744-bc48-bc2d232a8d6a-kube-api-access-5zj4s\") pod \"barbican-api-9f58cdf4d-fz24c\" (UID: \"08bdbc96-88c5-4744-bc48-bc2d232a8d6a\") " pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.042947 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnw6f\" (UniqueName: \"kubernetes.io/projected/f762f277-36aa-4ed7-8244-374a0481c789-kube-api-access-wnw6f\") pod \"dnsmasq-dns-d79c65b4f-pfwn8\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.087250 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.195258 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.466133 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5756b5f4c9-wj8k4"] Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.539469 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-77b797d5d4-vh4h4"] Dec 05 08:53:45 crc kubenswrapper[4997]: W1205 08:53:45.544265 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c326c07_9fee_4b86_b7c1_bd9ce32c95f3.slice/crio-e5b9174b45f71a5269e995143579e43df41742f18156c87ab5756ab37fbd66de WatchSource:0}: Error finding container e5b9174b45f71a5269e995143579e43df41742f18156c87ab5756ab37fbd66de: Status 404 returned error can't find the container with id e5b9174b45f71a5269e995143579e43df41742f18156c87ab5756ab37fbd66de Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.638375 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d79c65b4f-pfwn8"] Dec 05 08:53:45 crc kubenswrapper[4997]: W1205 08:53:45.647742 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf762f277_36aa_4ed7_8244_374a0481c789.slice/crio-1dce4183c06dfb3bb6e4f0a68105ddf682a164f01e1ca0c0db9a28ca8857b766 WatchSource:0}: Error finding container 1dce4183c06dfb3bb6e4f0a68105ddf682a164f01e1ca0c0db9a28ca8857b766: Status 404 returned error can't find the container with id 1dce4183c06dfb3bb6e4f0a68105ddf682a164f01e1ca0c0db9a28ca8857b766 Dec 05 08:53:45 crc kubenswrapper[4997]: I1205 08:53:45.786284 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9f58cdf4d-fz24c"] Dec 05 08:53:45 crc kubenswrapper[4997]: W1205 08:53:45.799928 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08bdbc96_88c5_4744_bc48_bc2d232a8d6a.slice/crio-aea19423ae724c7332be9c3e274137d8dc66257733cf6431bdb18efd8d70f278 WatchSource:0}: Error finding container aea19423ae724c7332be9c3e274137d8dc66257733cf6431bdb18efd8d70f278: Status 404 returned error can't find the container with id aea19423ae724c7332be9c3e274137d8dc66257733cf6431bdb18efd8d70f278 Dec 05 08:53:46 crc kubenswrapper[4997]: I1205 08:53:46.442883 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5756b5f4c9-wj8k4" event={"ID":"0666a02f-d3ed-4403-b2af-5450e2901d82","Type":"ContainerStarted","Data":"d5fe594313352c17d84cc056ffd3b912f3b749f8ed3ccbca242a5a3908e879dc"} Dec 05 08:53:46 crc kubenswrapper[4997]: I1205 08:53:46.444824 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9f58cdf4d-fz24c" event={"ID":"08bdbc96-88c5-4744-bc48-bc2d232a8d6a","Type":"ContainerStarted","Data":"5d65c2623fa57978cd922581b4f10f731763eadd1edfd55627605d4e24eeee58"} Dec 05 08:53:46 crc kubenswrapper[4997]: I1205 08:53:46.444902 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9f58cdf4d-fz24c" event={"ID":"08bdbc96-88c5-4744-bc48-bc2d232a8d6a","Type":"ContainerStarted","Data":"aea19423ae724c7332be9c3e274137d8dc66257733cf6431bdb18efd8d70f278"} Dec 05 08:53:46 crc kubenswrapper[4997]: I1205 08:53:46.447148 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" event={"ID":"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3","Type":"ContainerStarted","Data":"e5b9174b45f71a5269e995143579e43df41742f18156c87ab5756ab37fbd66de"} Dec 05 08:53:46 crc kubenswrapper[4997]: I1205 08:53:46.450102 4997 generic.go:334] "Generic (PLEG): container finished" podID="f762f277-36aa-4ed7-8244-374a0481c789" containerID="88acceecb88fcb5a6a4f66c6b1648d19b3763b00a6d08872ad086d84fd8d3213" exitCode=0 Dec 05 08:53:46 crc kubenswrapper[4997]: I1205 08:53:46.450149 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" event={"ID":"f762f277-36aa-4ed7-8244-374a0481c789","Type":"ContainerDied","Data":"88acceecb88fcb5a6a4f66c6b1648d19b3763b00a6d08872ad086d84fd8d3213"} Dec 05 08:53:46 crc kubenswrapper[4997]: I1205 08:53:46.450176 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" event={"ID":"f762f277-36aa-4ed7-8244-374a0481c789","Type":"ContainerStarted","Data":"1dce4183c06dfb3bb6e4f0a68105ddf682a164f01e1ca0c0db9a28ca8857b766"} Dec 05 08:53:47 crc kubenswrapper[4997]: I1205 08:53:47.485225 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5756b5f4c9-wj8k4" event={"ID":"0666a02f-d3ed-4403-b2af-5450e2901d82","Type":"ContainerStarted","Data":"fd6b69fbdd8749e2007d3222f1baae2935d2289bc1cfc54d4f470bc813815229"} Dec 05 08:53:47 crc kubenswrapper[4997]: I1205 08:53:47.485858 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5756b5f4c9-wj8k4" event={"ID":"0666a02f-d3ed-4403-b2af-5450e2901d82","Type":"ContainerStarted","Data":"9c6e6f5fcd7c99708a1caec2ea0d77111549a9617afc9ed36a640b7cdf51bf75"} Dec 05 08:53:47 crc kubenswrapper[4997]: I1205 08:53:47.500861 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9f58cdf4d-fz24c" event={"ID":"08bdbc96-88c5-4744-bc48-bc2d232a8d6a","Type":"ContainerStarted","Data":"b3a1ec97f74e2bab01162ff17e189503dd217bed176ca6e17657251b41e5a459"} Dec 05 08:53:47 crc kubenswrapper[4997]: I1205 08:53:47.501420 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:47 crc kubenswrapper[4997]: I1205 08:53:47.501468 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:47 crc kubenswrapper[4997]: I1205 08:53:47.504058 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" event={"ID":"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3","Type":"ContainerStarted","Data":"1272f27caf26858a3a865a14c7d224f550077fadafe1718a4cf2abe967119341"} Dec 05 08:53:47 crc kubenswrapper[4997]: I1205 08:53:47.504104 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" event={"ID":"6c326c07-9fee-4b86-b7c1-bd9ce32c95f3","Type":"ContainerStarted","Data":"a804301e05be4fb145566751c8647816e5f3a2e8adb9060f350f5d7d88a76d02"} Dec 05 08:53:47 crc kubenswrapper[4997]: I1205 08:53:47.514556 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5756b5f4c9-wj8k4" podStartSLOduration=2.326821301 podStartE2EDuration="3.514533857s" podCreationTimestamp="2025-12-05 08:53:44 +0000 UTC" firstStartedPulling="2025-12-05 08:53:45.470289626 +0000 UTC m=+7125.999196897" lastFinishedPulling="2025-12-05 08:53:46.658002202 +0000 UTC m=+7127.186909453" observedRunningTime="2025-12-05 08:53:47.510489738 +0000 UTC m=+7128.039396999" watchObservedRunningTime="2025-12-05 08:53:47.514533857 +0000 UTC m=+7128.043441128" Dec 05 08:53:47 crc kubenswrapper[4997]: I1205 08:53:47.525783 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" event={"ID":"f762f277-36aa-4ed7-8244-374a0481c789","Type":"ContainerStarted","Data":"040da46b1bf79012d4d595d1f632625f98ef4b2bacadb1f74b0de36a646516b1"} Dec 05 08:53:47 crc kubenswrapper[4997]: I1205 08:53:47.525967 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:47 crc kubenswrapper[4997]: I1205 08:53:47.566515 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-9f58cdf4d-fz24c" podStartSLOduration=3.566499393 podStartE2EDuration="3.566499393s" podCreationTimestamp="2025-12-05 08:53:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:53:47.563929704 +0000 UTC m=+7128.092836965" watchObservedRunningTime="2025-12-05 08:53:47.566499393 +0000 UTC m=+7128.095406644" Dec 05 08:53:47 crc kubenswrapper[4997]: I1205 08:53:47.587764 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-77b797d5d4-vh4h4" podStartSLOduration=2.476217445 podStartE2EDuration="3.587744974s" podCreationTimestamp="2025-12-05 08:53:44 +0000 UTC" firstStartedPulling="2025-12-05 08:53:45.547194903 +0000 UTC m=+7126.076102164" lastFinishedPulling="2025-12-05 08:53:46.658722412 +0000 UTC m=+7127.187629693" observedRunningTime="2025-12-05 08:53:47.585075023 +0000 UTC m=+7128.113982294" watchObservedRunningTime="2025-12-05 08:53:47.587744974 +0000 UTC m=+7128.116652235" Dec 05 08:53:47 crc kubenswrapper[4997]: I1205 08:53:47.619627 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" podStartSLOduration=3.619593291 podStartE2EDuration="3.619593291s" podCreationTimestamp="2025-12-05 08:53:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:53:47.619008125 +0000 UTC m=+7128.147915406" watchObservedRunningTime="2025-12-05 08:53:47.619593291 +0000 UTC m=+7128.148500552" Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.088863 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.182440 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6468849cd7-rbk8q"] Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.189740 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" podUID="a17a3529-5b5c-4029-bb2a-f5883d800c4d" containerName="dnsmasq-dns" containerID="cri-o://ea64bf5ff5e848152bb443a88b44980e48b57427a15e6bb36d885c2bb453ca13" gracePeriod=10 Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.600818 4997 generic.go:334] "Generic (PLEG): container finished" podID="a17a3529-5b5c-4029-bb2a-f5883d800c4d" containerID="ea64bf5ff5e848152bb443a88b44980e48b57427a15e6bb36d885c2bb453ca13" exitCode=0 Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.600925 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" event={"ID":"a17a3529-5b5c-4029-bb2a-f5883d800c4d","Type":"ContainerDied","Data":"ea64bf5ff5e848152bb443a88b44980e48b57427a15e6bb36d885c2bb453ca13"} Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.705842 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.853031 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-ovsdbserver-sb\") pod \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.853111 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8h7q\" (UniqueName: \"kubernetes.io/projected/a17a3529-5b5c-4029-bb2a-f5883d800c4d-kube-api-access-p8h7q\") pod \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.853231 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-config\") pod \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.853249 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-ovsdbserver-nb\") pod \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.853288 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-dns-svc\") pod \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\" (UID: \"a17a3529-5b5c-4029-bb2a-f5883d800c4d\") " Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.874385 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a17a3529-5b5c-4029-bb2a-f5883d800c4d-kube-api-access-p8h7q" (OuterVolumeSpecName: "kube-api-access-p8h7q") pod "a17a3529-5b5c-4029-bb2a-f5883d800c4d" (UID: "a17a3529-5b5c-4029-bb2a-f5883d800c4d"). InnerVolumeSpecName "kube-api-access-p8h7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.894182 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a17a3529-5b5c-4029-bb2a-f5883d800c4d" (UID: "a17a3529-5b5c-4029-bb2a-f5883d800c4d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.901819 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-config" (OuterVolumeSpecName: "config") pod "a17a3529-5b5c-4029-bb2a-f5883d800c4d" (UID: "a17a3529-5b5c-4029-bb2a-f5883d800c4d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.907545 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a17a3529-5b5c-4029-bb2a-f5883d800c4d" (UID: "a17a3529-5b5c-4029-bb2a-f5883d800c4d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.917369 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a17a3529-5b5c-4029-bb2a-f5883d800c4d" (UID: "a17a3529-5b5c-4029-bb2a-f5883d800c4d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.955130 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.955190 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.955236 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.955245 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a17a3529-5b5c-4029-bb2a-f5883d800c4d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:53:55 crc kubenswrapper[4997]: I1205 08:53:55.955255 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8h7q\" (UniqueName: \"kubernetes.io/projected/a17a3529-5b5c-4029-bb2a-f5883d800c4d-kube-api-access-p8h7q\") on node \"crc\" DevicePath \"\"" Dec 05 08:53:56 crc kubenswrapper[4997]: I1205 08:53:56.610278 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" event={"ID":"a17a3529-5b5c-4029-bb2a-f5883d800c4d","Type":"ContainerDied","Data":"07112dd7e7b4204671cfc0280d87dfb0cf40b44d0c78dfd9bae828364615d91b"} Dec 05 08:53:56 crc kubenswrapper[4997]: I1205 08:53:56.610332 4997 scope.go:117] "RemoveContainer" containerID="ea64bf5ff5e848152bb443a88b44980e48b57427a15e6bb36d885c2bb453ca13" Dec 05 08:53:56 crc kubenswrapper[4997]: I1205 08:53:56.610444 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6468849cd7-rbk8q" Dec 05 08:53:56 crc kubenswrapper[4997]: I1205 08:53:56.642896 4997 scope.go:117] "RemoveContainer" containerID="9b8a0e39b7c8a7a70943fe99e20575281db387be1814f84d64f6a8c13eb4c93a" Dec 05 08:53:56 crc kubenswrapper[4997]: I1205 08:53:56.651253 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6468849cd7-rbk8q"] Dec 05 08:53:56 crc kubenswrapper[4997]: I1205 08:53:56.658864 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6468849cd7-rbk8q"] Dec 05 08:53:56 crc kubenswrapper[4997]: I1205 08:53:56.860176 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:56 crc kubenswrapper[4997]: I1205 08:53:56.913519 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-9f58cdf4d-fz24c" Dec 05 08:53:57 crc kubenswrapper[4997]: I1205 08:53:57.759368 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a17a3529-5b5c-4029-bb2a-f5883d800c4d" path="/var/lib/kubelet/pods/a17a3529-5b5c-4029-bb2a-f5883d800c4d/volumes" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.046707 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-q75gz"] Dec 05 08:54:03 crc kubenswrapper[4997]: E1205 08:54:03.047546 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a17a3529-5b5c-4029-bb2a-f5883d800c4d" containerName="dnsmasq-dns" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.047559 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a17a3529-5b5c-4029-bb2a-f5883d800c4d" containerName="dnsmasq-dns" Dec 05 08:54:03 crc kubenswrapper[4997]: E1205 08:54:03.047572 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a17a3529-5b5c-4029-bb2a-f5883d800c4d" containerName="init" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.047577 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a17a3529-5b5c-4029-bb2a-f5883d800c4d" containerName="init" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.047757 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a17a3529-5b5c-4029-bb2a-f5883d800c4d" containerName="dnsmasq-dns" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.048274 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-q75gz" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.070433 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-q75gz"] Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.150988 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-67c3-account-create-update-2mshm"] Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.152279 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67c3-account-create-update-2mshm" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.154829 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.164607 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-67c3-account-create-update-2mshm"] Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.175814 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538190ea-f66a-408c-9623-ced05985673f-operator-scripts\") pod \"neutron-db-create-q75gz\" (UID: \"538190ea-f66a-408c-9623-ced05985673f\") " pod="openstack/neutron-db-create-q75gz" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.175880 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpbxm\" (UniqueName: \"kubernetes.io/projected/538190ea-f66a-408c-9623-ced05985673f-kube-api-access-qpbxm\") pod \"neutron-db-create-q75gz\" (UID: \"538190ea-f66a-408c-9623-ced05985673f\") " pod="openstack/neutron-db-create-q75gz" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.277743 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86gzq\" (UniqueName: \"kubernetes.io/projected/bde1371d-4c9d-4e91-99a5-dcd2bba4a22e-kube-api-access-86gzq\") pod \"neutron-67c3-account-create-update-2mshm\" (UID: \"bde1371d-4c9d-4e91-99a5-dcd2bba4a22e\") " pod="openstack/neutron-67c3-account-create-update-2mshm" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.278047 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538190ea-f66a-408c-9623-ced05985673f-operator-scripts\") pod \"neutron-db-create-q75gz\" (UID: \"538190ea-f66a-408c-9623-ced05985673f\") " pod="openstack/neutron-db-create-q75gz" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.278099 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpbxm\" (UniqueName: \"kubernetes.io/projected/538190ea-f66a-408c-9623-ced05985673f-kube-api-access-qpbxm\") pod \"neutron-db-create-q75gz\" (UID: \"538190ea-f66a-408c-9623-ced05985673f\") " pod="openstack/neutron-db-create-q75gz" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.278147 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bde1371d-4c9d-4e91-99a5-dcd2bba4a22e-operator-scripts\") pod \"neutron-67c3-account-create-update-2mshm\" (UID: \"bde1371d-4c9d-4e91-99a5-dcd2bba4a22e\") " pod="openstack/neutron-67c3-account-create-update-2mshm" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.279204 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538190ea-f66a-408c-9623-ced05985673f-operator-scripts\") pod \"neutron-db-create-q75gz\" (UID: \"538190ea-f66a-408c-9623-ced05985673f\") " pod="openstack/neutron-db-create-q75gz" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.300826 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpbxm\" (UniqueName: \"kubernetes.io/projected/538190ea-f66a-408c-9623-ced05985673f-kube-api-access-qpbxm\") pod \"neutron-db-create-q75gz\" (UID: \"538190ea-f66a-408c-9623-ced05985673f\") " pod="openstack/neutron-db-create-q75gz" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.364198 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-q75gz" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.379708 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bde1371d-4c9d-4e91-99a5-dcd2bba4a22e-operator-scripts\") pod \"neutron-67c3-account-create-update-2mshm\" (UID: \"bde1371d-4c9d-4e91-99a5-dcd2bba4a22e\") " pod="openstack/neutron-67c3-account-create-update-2mshm" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.379835 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86gzq\" (UniqueName: \"kubernetes.io/projected/bde1371d-4c9d-4e91-99a5-dcd2bba4a22e-kube-api-access-86gzq\") pod \"neutron-67c3-account-create-update-2mshm\" (UID: \"bde1371d-4c9d-4e91-99a5-dcd2bba4a22e\") " pod="openstack/neutron-67c3-account-create-update-2mshm" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.381217 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bde1371d-4c9d-4e91-99a5-dcd2bba4a22e-operator-scripts\") pod \"neutron-67c3-account-create-update-2mshm\" (UID: \"bde1371d-4c9d-4e91-99a5-dcd2bba4a22e\") " pod="openstack/neutron-67c3-account-create-update-2mshm" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.397299 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86gzq\" (UniqueName: \"kubernetes.io/projected/bde1371d-4c9d-4e91-99a5-dcd2bba4a22e-kube-api-access-86gzq\") pod \"neutron-67c3-account-create-update-2mshm\" (UID: \"bde1371d-4c9d-4e91-99a5-dcd2bba4a22e\") " pod="openstack/neutron-67c3-account-create-update-2mshm" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.468275 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67c3-account-create-update-2mshm" Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.642083 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-q75gz"] Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.680125 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-q75gz" event={"ID":"538190ea-f66a-408c-9623-ced05985673f","Type":"ContainerStarted","Data":"17308449644e46514d5e951d0445e2dcc996a61a9f8862bcd9992c7b3e4630a0"} Dec 05 08:54:03 crc kubenswrapper[4997]: I1205 08:54:03.984546 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-67c3-account-create-update-2mshm"] Dec 05 08:54:04 crc kubenswrapper[4997]: I1205 08:54:04.689588 4997 generic.go:334] "Generic (PLEG): container finished" podID="bde1371d-4c9d-4e91-99a5-dcd2bba4a22e" containerID="db5a163466ab65869c5e21d0cbe1357f5a2f20986d633203d918ed439454d9ce" exitCode=0 Dec 05 08:54:04 crc kubenswrapper[4997]: I1205 08:54:04.689758 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67c3-account-create-update-2mshm" event={"ID":"bde1371d-4c9d-4e91-99a5-dcd2bba4a22e","Type":"ContainerDied","Data":"db5a163466ab65869c5e21d0cbe1357f5a2f20986d633203d918ed439454d9ce"} Dec 05 08:54:04 crc kubenswrapper[4997]: I1205 08:54:04.689828 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67c3-account-create-update-2mshm" event={"ID":"bde1371d-4c9d-4e91-99a5-dcd2bba4a22e","Type":"ContainerStarted","Data":"b08cbffb300c97f9be07fc34f86f15ddc64ff2e19b459810a6b991c21d865d67"} Dec 05 08:54:04 crc kubenswrapper[4997]: I1205 08:54:04.693807 4997 generic.go:334] "Generic (PLEG): container finished" podID="538190ea-f66a-408c-9623-ced05985673f" containerID="07903e1eeb8d39db262958be291c36d6522a607de50d8c7b9e972a5b3eff1d1d" exitCode=0 Dec 05 08:54:04 crc kubenswrapper[4997]: I1205 08:54:04.693865 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-q75gz" event={"ID":"538190ea-f66a-408c-9623-ced05985673f","Type":"ContainerDied","Data":"07903e1eeb8d39db262958be291c36d6522a607de50d8c7b9e972a5b3eff1d1d"} Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.118957 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-q75gz" Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.126361 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67c3-account-create-update-2mshm" Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.234610 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538190ea-f66a-408c-9623-ced05985673f-operator-scripts\") pod \"538190ea-f66a-408c-9623-ced05985673f\" (UID: \"538190ea-f66a-408c-9623-ced05985673f\") " Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.234663 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86gzq\" (UniqueName: \"kubernetes.io/projected/bde1371d-4c9d-4e91-99a5-dcd2bba4a22e-kube-api-access-86gzq\") pod \"bde1371d-4c9d-4e91-99a5-dcd2bba4a22e\" (UID: \"bde1371d-4c9d-4e91-99a5-dcd2bba4a22e\") " Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.234788 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpbxm\" (UniqueName: \"kubernetes.io/projected/538190ea-f66a-408c-9623-ced05985673f-kube-api-access-qpbxm\") pod \"538190ea-f66a-408c-9623-ced05985673f\" (UID: \"538190ea-f66a-408c-9623-ced05985673f\") " Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.234847 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bde1371d-4c9d-4e91-99a5-dcd2bba4a22e-operator-scripts\") pod \"bde1371d-4c9d-4e91-99a5-dcd2bba4a22e\" (UID: \"bde1371d-4c9d-4e91-99a5-dcd2bba4a22e\") " Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.235944 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bde1371d-4c9d-4e91-99a5-dcd2bba4a22e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bde1371d-4c9d-4e91-99a5-dcd2bba4a22e" (UID: "bde1371d-4c9d-4e91-99a5-dcd2bba4a22e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.235936 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/538190ea-f66a-408c-9623-ced05985673f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "538190ea-f66a-408c-9623-ced05985673f" (UID: "538190ea-f66a-408c-9623-ced05985673f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.241643 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bde1371d-4c9d-4e91-99a5-dcd2bba4a22e-kube-api-access-86gzq" (OuterVolumeSpecName: "kube-api-access-86gzq") pod "bde1371d-4c9d-4e91-99a5-dcd2bba4a22e" (UID: "bde1371d-4c9d-4e91-99a5-dcd2bba4a22e"). InnerVolumeSpecName "kube-api-access-86gzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.247807 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/538190ea-f66a-408c-9623-ced05985673f-kube-api-access-qpbxm" (OuterVolumeSpecName: "kube-api-access-qpbxm") pod "538190ea-f66a-408c-9623-ced05985673f" (UID: "538190ea-f66a-408c-9623-ced05985673f"). InnerVolumeSpecName "kube-api-access-qpbxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.337080 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538190ea-f66a-408c-9623-ced05985673f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.337120 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86gzq\" (UniqueName: \"kubernetes.io/projected/bde1371d-4c9d-4e91-99a5-dcd2bba4a22e-kube-api-access-86gzq\") on node \"crc\" DevicePath \"\"" Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.337136 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpbxm\" (UniqueName: \"kubernetes.io/projected/538190ea-f66a-408c-9623-ced05985673f-kube-api-access-qpbxm\") on node \"crc\" DevicePath \"\"" Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.337148 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bde1371d-4c9d-4e91-99a5-dcd2bba4a22e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.714257 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67c3-account-create-update-2mshm" event={"ID":"bde1371d-4c9d-4e91-99a5-dcd2bba4a22e","Type":"ContainerDied","Data":"b08cbffb300c97f9be07fc34f86f15ddc64ff2e19b459810a6b991c21d865d67"} Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.714284 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67c3-account-create-update-2mshm" Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.714301 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b08cbffb300c97f9be07fc34f86f15ddc64ff2e19b459810a6b991c21d865d67" Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.715811 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-q75gz" event={"ID":"538190ea-f66a-408c-9623-ced05985673f","Type":"ContainerDied","Data":"17308449644e46514d5e951d0445e2dcc996a61a9f8862bcd9992c7b3e4630a0"} Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.715848 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17308449644e46514d5e951d0445e2dcc996a61a9f8862bcd9992c7b3e4630a0" Dec 05 08:54:06 crc kubenswrapper[4997]: I1205 08:54:06.715915 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-q75gz" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.366322 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-bcmls"] Dec 05 08:54:08 crc kubenswrapper[4997]: E1205 08:54:08.366938 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bde1371d-4c9d-4e91-99a5-dcd2bba4a22e" containerName="mariadb-account-create-update" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.366950 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="bde1371d-4c9d-4e91-99a5-dcd2bba4a22e" containerName="mariadb-account-create-update" Dec 05 08:54:08 crc kubenswrapper[4997]: E1205 08:54:08.366961 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="538190ea-f66a-408c-9623-ced05985673f" containerName="mariadb-database-create" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.366967 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="538190ea-f66a-408c-9623-ced05985673f" containerName="mariadb-database-create" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.367119 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="bde1371d-4c9d-4e91-99a5-dcd2bba4a22e" containerName="mariadb-account-create-update" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.367135 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="538190ea-f66a-408c-9623-ced05985673f" containerName="mariadb-database-create" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.367684 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bcmls" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.374584 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-bcmls"] Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.410561 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.410741 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-s6dgn" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.410804 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.471711 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0f5512ca-098c-4d06-b455-3a76338df304-config\") pod \"neutron-db-sync-bcmls\" (UID: \"0f5512ca-098c-4d06-b455-3a76338df304\") " pod="openstack/neutron-db-sync-bcmls" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.471755 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrcpt\" (UniqueName: \"kubernetes.io/projected/0f5512ca-098c-4d06-b455-3a76338df304-kube-api-access-mrcpt\") pod \"neutron-db-sync-bcmls\" (UID: \"0f5512ca-098c-4d06-b455-3a76338df304\") " pod="openstack/neutron-db-sync-bcmls" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.471787 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f5512ca-098c-4d06-b455-3a76338df304-combined-ca-bundle\") pod \"neutron-db-sync-bcmls\" (UID: \"0f5512ca-098c-4d06-b455-3a76338df304\") " pod="openstack/neutron-db-sync-bcmls" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.573934 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0f5512ca-098c-4d06-b455-3a76338df304-config\") pod \"neutron-db-sync-bcmls\" (UID: \"0f5512ca-098c-4d06-b455-3a76338df304\") " pod="openstack/neutron-db-sync-bcmls" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.573981 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrcpt\" (UniqueName: \"kubernetes.io/projected/0f5512ca-098c-4d06-b455-3a76338df304-kube-api-access-mrcpt\") pod \"neutron-db-sync-bcmls\" (UID: \"0f5512ca-098c-4d06-b455-3a76338df304\") " pod="openstack/neutron-db-sync-bcmls" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.574020 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f5512ca-098c-4d06-b455-3a76338df304-combined-ca-bundle\") pod \"neutron-db-sync-bcmls\" (UID: \"0f5512ca-098c-4d06-b455-3a76338df304\") " pod="openstack/neutron-db-sync-bcmls" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.578217 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f5512ca-098c-4d06-b455-3a76338df304-combined-ca-bundle\") pod \"neutron-db-sync-bcmls\" (UID: \"0f5512ca-098c-4d06-b455-3a76338df304\") " pod="openstack/neutron-db-sync-bcmls" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.578228 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0f5512ca-098c-4d06-b455-3a76338df304-config\") pod \"neutron-db-sync-bcmls\" (UID: \"0f5512ca-098c-4d06-b455-3a76338df304\") " pod="openstack/neutron-db-sync-bcmls" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.599560 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrcpt\" (UniqueName: \"kubernetes.io/projected/0f5512ca-098c-4d06-b455-3a76338df304-kube-api-access-mrcpt\") pod \"neutron-db-sync-bcmls\" (UID: \"0f5512ca-098c-4d06-b455-3a76338df304\") " pod="openstack/neutron-db-sync-bcmls" Dec 05 08:54:08 crc kubenswrapper[4997]: I1205 08:54:08.730602 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bcmls" Dec 05 08:54:09 crc kubenswrapper[4997]: I1205 08:54:09.209990 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-bcmls"] Dec 05 08:54:09 crc kubenswrapper[4997]: I1205 08:54:09.738441 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bcmls" event={"ID":"0f5512ca-098c-4d06-b455-3a76338df304","Type":"ContainerStarted","Data":"d3a0b14c5b38af42c3a0f352d682859e87679bf2169ad20834a221d60b888743"} Dec 05 08:54:09 crc kubenswrapper[4997]: I1205 08:54:09.739521 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bcmls" event={"ID":"0f5512ca-098c-4d06-b455-3a76338df304","Type":"ContainerStarted","Data":"8cace8ffd06440fd2c403675b0c333447580048f00c44dc784f460c90279474e"} Dec 05 08:54:09 crc kubenswrapper[4997]: I1205 08:54:09.751905 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-bcmls" podStartSLOduration=1.751885772 podStartE2EDuration="1.751885772s" podCreationTimestamp="2025-12-05 08:54:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:54:09.751098622 +0000 UTC m=+7150.280005893" watchObservedRunningTime="2025-12-05 08:54:09.751885772 +0000 UTC m=+7150.280793033" Dec 05 08:54:13 crc kubenswrapper[4997]: I1205 08:54:13.782802 4997 generic.go:334] "Generic (PLEG): container finished" podID="0f5512ca-098c-4d06-b455-3a76338df304" containerID="d3a0b14c5b38af42c3a0f352d682859e87679bf2169ad20834a221d60b888743" exitCode=0 Dec 05 08:54:13 crc kubenswrapper[4997]: I1205 08:54:13.782907 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bcmls" event={"ID":"0f5512ca-098c-4d06-b455-3a76338df304","Type":"ContainerDied","Data":"d3a0b14c5b38af42c3a0f352d682859e87679bf2169ad20834a221d60b888743"} Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.148532 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bcmls" Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.329349 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0f5512ca-098c-4d06-b455-3a76338df304-config\") pod \"0f5512ca-098c-4d06-b455-3a76338df304\" (UID: \"0f5512ca-098c-4d06-b455-3a76338df304\") " Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.329423 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrcpt\" (UniqueName: \"kubernetes.io/projected/0f5512ca-098c-4d06-b455-3a76338df304-kube-api-access-mrcpt\") pod \"0f5512ca-098c-4d06-b455-3a76338df304\" (UID: \"0f5512ca-098c-4d06-b455-3a76338df304\") " Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.329669 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f5512ca-098c-4d06-b455-3a76338df304-combined-ca-bundle\") pod \"0f5512ca-098c-4d06-b455-3a76338df304\" (UID: \"0f5512ca-098c-4d06-b455-3a76338df304\") " Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.334902 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f5512ca-098c-4d06-b455-3a76338df304-kube-api-access-mrcpt" (OuterVolumeSpecName: "kube-api-access-mrcpt") pod "0f5512ca-098c-4d06-b455-3a76338df304" (UID: "0f5512ca-098c-4d06-b455-3a76338df304"). InnerVolumeSpecName "kube-api-access-mrcpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.362439 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f5512ca-098c-4d06-b455-3a76338df304-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f5512ca-098c-4d06-b455-3a76338df304" (UID: "0f5512ca-098c-4d06-b455-3a76338df304"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.363920 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f5512ca-098c-4d06-b455-3a76338df304-config" (OuterVolumeSpecName: "config") pod "0f5512ca-098c-4d06-b455-3a76338df304" (UID: "0f5512ca-098c-4d06-b455-3a76338df304"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.431331 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f5512ca-098c-4d06-b455-3a76338df304-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.431380 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/0f5512ca-098c-4d06-b455-3a76338df304-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.431401 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrcpt\" (UniqueName: \"kubernetes.io/projected/0f5512ca-098c-4d06-b455-3a76338df304-kube-api-access-mrcpt\") on node \"crc\" DevicePath \"\"" Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.800539 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bcmls" event={"ID":"0f5512ca-098c-4d06-b455-3a76338df304","Type":"ContainerDied","Data":"8cace8ffd06440fd2c403675b0c333447580048f00c44dc784f460c90279474e"} Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.800588 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cace8ffd06440fd2c403675b0c333447580048f00c44dc784f460c90279474e" Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.800588 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bcmls" Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.927005 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67877c9485-2ncv4"] Dec 05 08:54:15 crc kubenswrapper[4997]: E1205 08:54:15.927629 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f5512ca-098c-4d06-b455-3a76338df304" containerName="neutron-db-sync" Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.927646 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f5512ca-098c-4d06-b455-3a76338df304" containerName="neutron-db-sync" Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.927856 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f5512ca-098c-4d06-b455-3a76338df304" containerName="neutron-db-sync" Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.928762 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:15 crc kubenswrapper[4997]: I1205 08:54:15.946572 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67877c9485-2ncv4"] Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.042092 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-dns-svc\") pod \"dnsmasq-dns-67877c9485-2ncv4\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.042296 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-config\") pod \"dnsmasq-dns-67877c9485-2ncv4\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.042363 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-ovsdbserver-sb\") pod \"dnsmasq-dns-67877c9485-2ncv4\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.042429 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvbg9\" (UniqueName: \"kubernetes.io/projected/c9475029-c206-4392-b272-ee631f5abf8b-kube-api-access-bvbg9\") pod \"dnsmasq-dns-67877c9485-2ncv4\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.042467 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-ovsdbserver-nb\") pod \"dnsmasq-dns-67877c9485-2ncv4\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.085165 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7f58c8c5b5-bd9p4"] Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.088997 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f58c8c5b5-bd9p4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.091784 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.091901 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-s6dgn" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.092002 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.096088 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7f58c8c5b5-bd9p4"] Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.143677 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-dns-svc\") pod \"dnsmasq-dns-67877c9485-2ncv4\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.143738 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-config\") pod \"dnsmasq-dns-67877c9485-2ncv4\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.144116 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-ovsdbserver-sb\") pod \"dnsmasq-dns-67877c9485-2ncv4\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.144213 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvbg9\" (UniqueName: \"kubernetes.io/projected/c9475029-c206-4392-b272-ee631f5abf8b-kube-api-access-bvbg9\") pod \"dnsmasq-dns-67877c9485-2ncv4\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.144235 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-ovsdbserver-nb\") pod \"dnsmasq-dns-67877c9485-2ncv4\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.144928 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-config\") pod \"dnsmasq-dns-67877c9485-2ncv4\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.144935 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-dns-svc\") pod \"dnsmasq-dns-67877c9485-2ncv4\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.145194 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-ovsdbserver-nb\") pod \"dnsmasq-dns-67877c9485-2ncv4\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.145838 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-ovsdbserver-sb\") pod \"dnsmasq-dns-67877c9485-2ncv4\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.163766 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvbg9\" (UniqueName: \"kubernetes.io/projected/c9475029-c206-4392-b272-ee631f5abf8b-kube-api-access-bvbg9\") pod \"dnsmasq-dns-67877c9485-2ncv4\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.245510 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/45add928-5cda-478a-9f33-ea2c2a09f99e-config\") pod \"neutron-7f58c8c5b5-bd9p4\" (UID: \"45add928-5cda-478a-9f33-ea2c2a09f99e\") " pod="openstack/neutron-7f58c8c5b5-bd9p4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.245565 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/45add928-5cda-478a-9f33-ea2c2a09f99e-httpd-config\") pod \"neutron-7f58c8c5b5-bd9p4\" (UID: \"45add928-5cda-478a-9f33-ea2c2a09f99e\") " pod="openstack/neutron-7f58c8c5b5-bd9p4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.245902 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg4hx\" (UniqueName: \"kubernetes.io/projected/45add928-5cda-478a-9f33-ea2c2a09f99e-kube-api-access-sg4hx\") pod \"neutron-7f58c8c5b5-bd9p4\" (UID: \"45add928-5cda-478a-9f33-ea2c2a09f99e\") " pod="openstack/neutron-7f58c8c5b5-bd9p4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.246074 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.246365 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45add928-5cda-478a-9f33-ea2c2a09f99e-combined-ca-bundle\") pod \"neutron-7f58c8c5b5-bd9p4\" (UID: \"45add928-5cda-478a-9f33-ea2c2a09f99e\") " pod="openstack/neutron-7f58c8c5b5-bd9p4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.347293 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg4hx\" (UniqueName: \"kubernetes.io/projected/45add928-5cda-478a-9f33-ea2c2a09f99e-kube-api-access-sg4hx\") pod \"neutron-7f58c8c5b5-bd9p4\" (UID: \"45add928-5cda-478a-9f33-ea2c2a09f99e\") " pod="openstack/neutron-7f58c8c5b5-bd9p4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.347650 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45add928-5cda-478a-9f33-ea2c2a09f99e-combined-ca-bundle\") pod \"neutron-7f58c8c5b5-bd9p4\" (UID: \"45add928-5cda-478a-9f33-ea2c2a09f99e\") " pod="openstack/neutron-7f58c8c5b5-bd9p4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.347691 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/45add928-5cda-478a-9f33-ea2c2a09f99e-config\") pod \"neutron-7f58c8c5b5-bd9p4\" (UID: \"45add928-5cda-478a-9f33-ea2c2a09f99e\") " pod="openstack/neutron-7f58c8c5b5-bd9p4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.347715 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/45add928-5cda-478a-9f33-ea2c2a09f99e-httpd-config\") pod \"neutron-7f58c8c5b5-bd9p4\" (UID: \"45add928-5cda-478a-9f33-ea2c2a09f99e\") " pod="openstack/neutron-7f58c8c5b5-bd9p4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.361948 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/45add928-5cda-478a-9f33-ea2c2a09f99e-httpd-config\") pod \"neutron-7f58c8c5b5-bd9p4\" (UID: \"45add928-5cda-478a-9f33-ea2c2a09f99e\") " pod="openstack/neutron-7f58c8c5b5-bd9p4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.371529 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/45add928-5cda-478a-9f33-ea2c2a09f99e-config\") pod \"neutron-7f58c8c5b5-bd9p4\" (UID: \"45add928-5cda-478a-9f33-ea2c2a09f99e\") " pod="openstack/neutron-7f58c8c5b5-bd9p4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.383115 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45add928-5cda-478a-9f33-ea2c2a09f99e-combined-ca-bundle\") pod \"neutron-7f58c8c5b5-bd9p4\" (UID: \"45add928-5cda-478a-9f33-ea2c2a09f99e\") " pod="openstack/neutron-7f58c8c5b5-bd9p4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.389441 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg4hx\" (UniqueName: \"kubernetes.io/projected/45add928-5cda-478a-9f33-ea2c2a09f99e-kube-api-access-sg4hx\") pod \"neutron-7f58c8c5b5-bd9p4\" (UID: \"45add928-5cda-478a-9f33-ea2c2a09f99e\") " pod="openstack/neutron-7f58c8c5b5-bd9p4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.413413 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f58c8c5b5-bd9p4" Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.725204 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67877c9485-2ncv4"] Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.808773 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67877c9485-2ncv4" event={"ID":"c9475029-c206-4392-b272-ee631f5abf8b","Type":"ContainerStarted","Data":"6cc899b3d18321e905f58fb7a8f7819c4d3cb8b60a91147a55d66e2cab2e35ff"} Dec 05 08:54:16 crc kubenswrapper[4997]: I1205 08:54:16.976005 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7f58c8c5b5-bd9p4"] Dec 05 08:54:16 crc kubenswrapper[4997]: W1205 08:54:16.985271 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45add928_5cda_478a_9f33_ea2c2a09f99e.slice/crio-b5522ad478ddc35a1902ef983a9c7acc12cb9d77b5a6cf60b953684e05372da2 WatchSource:0}: Error finding container b5522ad478ddc35a1902ef983a9c7acc12cb9d77b5a6cf60b953684e05372da2: Status 404 returned error can't find the container with id b5522ad478ddc35a1902ef983a9c7acc12cb9d77b5a6cf60b953684e05372da2 Dec 05 08:54:17 crc kubenswrapper[4997]: I1205 08:54:17.832097 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f58c8c5b5-bd9p4" event={"ID":"45add928-5cda-478a-9f33-ea2c2a09f99e","Type":"ContainerStarted","Data":"e9b67c9f42d6f1b74ed0e70eacc696c455a61e26f092b91f08feaff351c8f15b"} Dec 05 08:54:17 crc kubenswrapper[4997]: I1205 08:54:17.833482 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f58c8c5b5-bd9p4" event={"ID":"45add928-5cda-478a-9f33-ea2c2a09f99e","Type":"ContainerStarted","Data":"e7ddc7a4e86e3cfc9a175845e240f2a61443a1e7bc7ee3774569c33bd481d4df"} Dec 05 08:54:17 crc kubenswrapper[4997]: I1205 08:54:17.833629 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f58c8c5b5-bd9p4" event={"ID":"45add928-5cda-478a-9f33-ea2c2a09f99e","Type":"ContainerStarted","Data":"b5522ad478ddc35a1902ef983a9c7acc12cb9d77b5a6cf60b953684e05372da2"} Dec 05 08:54:17 crc kubenswrapper[4997]: I1205 08:54:17.833735 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7f58c8c5b5-bd9p4" Dec 05 08:54:17 crc kubenswrapper[4997]: I1205 08:54:17.871056 4997 generic.go:334] "Generic (PLEG): container finished" podID="c9475029-c206-4392-b272-ee631f5abf8b" containerID="1fa05a1beb6bb7ab8978174b8575098972507095532fa138ca44613b99d39083" exitCode=0 Dec 05 08:54:17 crc kubenswrapper[4997]: I1205 08:54:17.871118 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67877c9485-2ncv4" event={"ID":"c9475029-c206-4392-b272-ee631f5abf8b","Type":"ContainerDied","Data":"1fa05a1beb6bb7ab8978174b8575098972507095532fa138ca44613b99d39083"} Dec 05 08:54:17 crc kubenswrapper[4997]: I1205 08:54:17.917063 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7f58c8c5b5-bd9p4" podStartSLOduration=1.9170422980000001 podStartE2EDuration="1.917042298s" podCreationTimestamp="2025-12-05 08:54:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:54:17.877083883 +0000 UTC m=+7158.405991154" watchObservedRunningTime="2025-12-05 08:54:17.917042298 +0000 UTC m=+7158.445949549" Dec 05 08:54:18 crc kubenswrapper[4997]: I1205 08:54:18.882113 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67877c9485-2ncv4" event={"ID":"c9475029-c206-4392-b272-ee631f5abf8b","Type":"ContainerStarted","Data":"d8bbcddb7772b4d820225aa9f06e6438aca9cd7bfefda3b3a6554dbe35204231"} Dec 05 08:54:18 crc kubenswrapper[4997]: I1205 08:54:18.915549 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67877c9485-2ncv4" podStartSLOduration=3.915530918 podStartE2EDuration="3.915530918s" podCreationTimestamp="2025-12-05 08:54:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:54:18.911931612 +0000 UTC m=+7159.440838873" watchObservedRunningTime="2025-12-05 08:54:18.915530918 +0000 UTC m=+7159.444438179" Dec 05 08:54:19 crc kubenswrapper[4997]: I1205 08:54:19.770216 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:54:19 crc kubenswrapper[4997]: I1205 08:54:19.770683 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:54:19 crc kubenswrapper[4997]: I1205 08:54:19.891946 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:26 crc kubenswrapper[4997]: I1205 08:54:26.247772 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:54:26 crc kubenswrapper[4997]: I1205 08:54:26.302558 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d79c65b4f-pfwn8"] Dec 05 08:54:26 crc kubenswrapper[4997]: I1205 08:54:26.303361 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" podUID="f762f277-36aa-4ed7-8244-374a0481c789" containerName="dnsmasq-dns" containerID="cri-o://040da46b1bf79012d4d595d1f632625f98ef4b2bacadb1f74b0de36a646516b1" gracePeriod=10 Dec 05 08:54:26 crc kubenswrapper[4997]: I1205 08:54:26.876888 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:54:26 crc kubenswrapper[4997]: I1205 08:54:26.953551 4997 generic.go:334] "Generic (PLEG): container finished" podID="f762f277-36aa-4ed7-8244-374a0481c789" containerID="040da46b1bf79012d4d595d1f632625f98ef4b2bacadb1f74b0de36a646516b1" exitCode=0 Dec 05 08:54:26 crc kubenswrapper[4997]: I1205 08:54:26.953593 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" Dec 05 08:54:26 crc kubenswrapper[4997]: I1205 08:54:26.953598 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" event={"ID":"f762f277-36aa-4ed7-8244-374a0481c789","Type":"ContainerDied","Data":"040da46b1bf79012d4d595d1f632625f98ef4b2bacadb1f74b0de36a646516b1"} Dec 05 08:54:26 crc kubenswrapper[4997]: I1205 08:54:26.953641 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d79c65b4f-pfwn8" event={"ID":"f762f277-36aa-4ed7-8244-374a0481c789","Type":"ContainerDied","Data":"1dce4183c06dfb3bb6e4f0a68105ddf682a164f01e1ca0c0db9a28ca8857b766"} Dec 05 08:54:26 crc kubenswrapper[4997]: I1205 08:54:26.953669 4997 scope.go:117] "RemoveContainer" containerID="040da46b1bf79012d4d595d1f632625f98ef4b2bacadb1f74b0de36a646516b1" Dec 05 08:54:26 crc kubenswrapper[4997]: I1205 08:54:26.972055 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-ovsdbserver-nb\") pod \"f762f277-36aa-4ed7-8244-374a0481c789\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " Dec 05 08:54:26 crc kubenswrapper[4997]: I1205 08:54:26.972185 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-dns-svc\") pod \"f762f277-36aa-4ed7-8244-374a0481c789\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " Dec 05 08:54:26 crc kubenswrapper[4997]: I1205 08:54:26.972252 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnw6f\" (UniqueName: \"kubernetes.io/projected/f762f277-36aa-4ed7-8244-374a0481c789-kube-api-access-wnw6f\") pod \"f762f277-36aa-4ed7-8244-374a0481c789\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " Dec 05 08:54:26 crc kubenswrapper[4997]: I1205 08:54:26.972308 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-config\") pod \"f762f277-36aa-4ed7-8244-374a0481c789\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " Dec 05 08:54:26 crc kubenswrapper[4997]: I1205 08:54:26.972401 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-ovsdbserver-sb\") pod \"f762f277-36aa-4ed7-8244-374a0481c789\" (UID: \"f762f277-36aa-4ed7-8244-374a0481c789\") " Dec 05 08:54:26 crc kubenswrapper[4997]: I1205 08:54:26.974108 4997 scope.go:117] "RemoveContainer" containerID="88acceecb88fcb5a6a4f66c6b1648d19b3763b00a6d08872ad086d84fd8d3213" Dec 05 08:54:26 crc kubenswrapper[4997]: I1205 08:54:26.985649 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f762f277-36aa-4ed7-8244-374a0481c789-kube-api-access-wnw6f" (OuterVolumeSpecName: "kube-api-access-wnw6f") pod "f762f277-36aa-4ed7-8244-374a0481c789" (UID: "f762f277-36aa-4ed7-8244-374a0481c789"). InnerVolumeSpecName "kube-api-access-wnw6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:54:27 crc kubenswrapper[4997]: I1205 08:54:27.015668 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-config" (OuterVolumeSpecName: "config") pod "f762f277-36aa-4ed7-8244-374a0481c789" (UID: "f762f277-36aa-4ed7-8244-374a0481c789"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:54:27 crc kubenswrapper[4997]: I1205 08:54:27.018180 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f762f277-36aa-4ed7-8244-374a0481c789" (UID: "f762f277-36aa-4ed7-8244-374a0481c789"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:54:27 crc kubenswrapper[4997]: I1205 08:54:27.024222 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f762f277-36aa-4ed7-8244-374a0481c789" (UID: "f762f277-36aa-4ed7-8244-374a0481c789"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:54:27 crc kubenswrapper[4997]: I1205 08:54:27.029224 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f762f277-36aa-4ed7-8244-374a0481c789" (UID: "f762f277-36aa-4ed7-8244-374a0481c789"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:54:27 crc kubenswrapper[4997]: I1205 08:54:27.077592 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnw6f\" (UniqueName: \"kubernetes.io/projected/f762f277-36aa-4ed7-8244-374a0481c789-kube-api-access-wnw6f\") on node \"crc\" DevicePath \"\"" Dec 05 08:54:27 crc kubenswrapper[4997]: I1205 08:54:27.077688 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:54:27 crc kubenswrapper[4997]: I1205 08:54:27.077705 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:54:27 crc kubenswrapper[4997]: I1205 08:54:27.077718 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:54:27 crc kubenswrapper[4997]: I1205 08:54:27.077730 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f762f277-36aa-4ed7-8244-374a0481c789-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:54:27 crc kubenswrapper[4997]: I1205 08:54:27.082563 4997 scope.go:117] "RemoveContainer" containerID="040da46b1bf79012d4d595d1f632625f98ef4b2bacadb1f74b0de36a646516b1" Dec 05 08:54:27 crc kubenswrapper[4997]: E1205 08:54:27.083345 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"040da46b1bf79012d4d595d1f632625f98ef4b2bacadb1f74b0de36a646516b1\": container with ID starting with 040da46b1bf79012d4d595d1f632625f98ef4b2bacadb1f74b0de36a646516b1 not found: ID does not exist" containerID="040da46b1bf79012d4d595d1f632625f98ef4b2bacadb1f74b0de36a646516b1" Dec 05 08:54:27 crc kubenswrapper[4997]: I1205 08:54:27.083389 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"040da46b1bf79012d4d595d1f632625f98ef4b2bacadb1f74b0de36a646516b1"} err="failed to get container status \"040da46b1bf79012d4d595d1f632625f98ef4b2bacadb1f74b0de36a646516b1\": rpc error: code = NotFound desc = could not find container \"040da46b1bf79012d4d595d1f632625f98ef4b2bacadb1f74b0de36a646516b1\": container with ID starting with 040da46b1bf79012d4d595d1f632625f98ef4b2bacadb1f74b0de36a646516b1 not found: ID does not exist" Dec 05 08:54:27 crc kubenswrapper[4997]: I1205 08:54:27.083417 4997 scope.go:117] "RemoveContainer" containerID="88acceecb88fcb5a6a4f66c6b1648d19b3763b00a6d08872ad086d84fd8d3213" Dec 05 08:54:27 crc kubenswrapper[4997]: E1205 08:54:27.084083 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88acceecb88fcb5a6a4f66c6b1648d19b3763b00a6d08872ad086d84fd8d3213\": container with ID starting with 88acceecb88fcb5a6a4f66c6b1648d19b3763b00a6d08872ad086d84fd8d3213 not found: ID does not exist" containerID="88acceecb88fcb5a6a4f66c6b1648d19b3763b00a6d08872ad086d84fd8d3213" Dec 05 08:54:27 crc kubenswrapper[4997]: I1205 08:54:27.084134 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88acceecb88fcb5a6a4f66c6b1648d19b3763b00a6d08872ad086d84fd8d3213"} err="failed to get container status \"88acceecb88fcb5a6a4f66c6b1648d19b3763b00a6d08872ad086d84fd8d3213\": rpc error: code = NotFound desc = could not find container \"88acceecb88fcb5a6a4f66c6b1648d19b3763b00a6d08872ad086d84fd8d3213\": container with ID starting with 88acceecb88fcb5a6a4f66c6b1648d19b3763b00a6d08872ad086d84fd8d3213 not found: ID does not exist" Dec 05 08:54:27 crc kubenswrapper[4997]: I1205 08:54:27.283602 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d79c65b4f-pfwn8"] Dec 05 08:54:27 crc kubenswrapper[4997]: I1205 08:54:27.292470 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d79c65b4f-pfwn8"] Dec 05 08:54:27 crc kubenswrapper[4997]: I1205 08:54:27.759362 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f762f277-36aa-4ed7-8244-374a0481c789" path="/var/lib/kubelet/pods/f762f277-36aa-4ed7-8244-374a0481c789/volumes" Dec 05 08:54:46 crc kubenswrapper[4997]: I1205 08:54:46.428774 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7f58c8c5b5-bd9p4" Dec 05 08:54:49 crc kubenswrapper[4997]: I1205 08:54:49.770592 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:54:49 crc kubenswrapper[4997]: I1205 08:54:49.771101 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.111531 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-nlgxp"] Dec 05 08:54:53 crc kubenswrapper[4997]: E1205 08:54:53.113300 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f762f277-36aa-4ed7-8244-374a0481c789" containerName="dnsmasq-dns" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.113393 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f762f277-36aa-4ed7-8244-374a0481c789" containerName="dnsmasq-dns" Dec 05 08:54:53 crc kubenswrapper[4997]: E1205 08:54:53.113468 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f762f277-36aa-4ed7-8244-374a0481c789" containerName="init" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.113525 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f762f277-36aa-4ed7-8244-374a0481c789" containerName="init" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.113748 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f762f277-36aa-4ed7-8244-374a0481c789" containerName="dnsmasq-dns" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.114372 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-nlgxp" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.125050 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-nlgxp"] Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.212455 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-c014-account-create-update-k8fmb"] Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.213861 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c014-account-create-update-k8fmb" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.216217 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.220229 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c014-account-create-update-k8fmb"] Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.254702 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp8fp\" (UniqueName: \"kubernetes.io/projected/95769d84-8dcc-4626-8f14-e4015d3cfa5e-kube-api-access-tp8fp\") pod \"glance-db-create-nlgxp\" (UID: \"95769d84-8dcc-4626-8f14-e4015d3cfa5e\") " pod="openstack/glance-db-create-nlgxp" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.254790 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95769d84-8dcc-4626-8f14-e4015d3cfa5e-operator-scripts\") pod \"glance-db-create-nlgxp\" (UID: \"95769d84-8dcc-4626-8f14-e4015d3cfa5e\") " pod="openstack/glance-db-create-nlgxp" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.356729 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px8zp\" (UniqueName: \"kubernetes.io/projected/bc83f439-c922-4812-ba87-59c083f3885d-kube-api-access-px8zp\") pod \"glance-c014-account-create-update-k8fmb\" (UID: \"bc83f439-c922-4812-ba87-59c083f3885d\") " pod="openstack/glance-c014-account-create-update-k8fmb" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.356785 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc83f439-c922-4812-ba87-59c083f3885d-operator-scripts\") pod \"glance-c014-account-create-update-k8fmb\" (UID: \"bc83f439-c922-4812-ba87-59c083f3885d\") " pod="openstack/glance-c014-account-create-update-k8fmb" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.356822 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp8fp\" (UniqueName: \"kubernetes.io/projected/95769d84-8dcc-4626-8f14-e4015d3cfa5e-kube-api-access-tp8fp\") pod \"glance-db-create-nlgxp\" (UID: \"95769d84-8dcc-4626-8f14-e4015d3cfa5e\") " pod="openstack/glance-db-create-nlgxp" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.356883 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95769d84-8dcc-4626-8f14-e4015d3cfa5e-operator-scripts\") pod \"glance-db-create-nlgxp\" (UID: \"95769d84-8dcc-4626-8f14-e4015d3cfa5e\") " pod="openstack/glance-db-create-nlgxp" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.357567 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95769d84-8dcc-4626-8f14-e4015d3cfa5e-operator-scripts\") pod \"glance-db-create-nlgxp\" (UID: \"95769d84-8dcc-4626-8f14-e4015d3cfa5e\") " pod="openstack/glance-db-create-nlgxp" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.375716 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp8fp\" (UniqueName: \"kubernetes.io/projected/95769d84-8dcc-4626-8f14-e4015d3cfa5e-kube-api-access-tp8fp\") pod \"glance-db-create-nlgxp\" (UID: \"95769d84-8dcc-4626-8f14-e4015d3cfa5e\") " pod="openstack/glance-db-create-nlgxp" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.434562 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-nlgxp" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.458907 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px8zp\" (UniqueName: \"kubernetes.io/projected/bc83f439-c922-4812-ba87-59c083f3885d-kube-api-access-px8zp\") pod \"glance-c014-account-create-update-k8fmb\" (UID: \"bc83f439-c922-4812-ba87-59c083f3885d\") " pod="openstack/glance-c014-account-create-update-k8fmb" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.459015 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc83f439-c922-4812-ba87-59c083f3885d-operator-scripts\") pod \"glance-c014-account-create-update-k8fmb\" (UID: \"bc83f439-c922-4812-ba87-59c083f3885d\") " pod="openstack/glance-c014-account-create-update-k8fmb" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.467424 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc83f439-c922-4812-ba87-59c083f3885d-operator-scripts\") pod \"glance-c014-account-create-update-k8fmb\" (UID: \"bc83f439-c922-4812-ba87-59c083f3885d\") " pod="openstack/glance-c014-account-create-update-k8fmb" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.485229 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px8zp\" (UniqueName: \"kubernetes.io/projected/bc83f439-c922-4812-ba87-59c083f3885d-kube-api-access-px8zp\") pod \"glance-c014-account-create-update-k8fmb\" (UID: \"bc83f439-c922-4812-ba87-59c083f3885d\") " pod="openstack/glance-c014-account-create-update-k8fmb" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.527444 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c014-account-create-update-k8fmb" Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.793070 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c014-account-create-update-k8fmb"] Dec 05 08:54:53 crc kubenswrapper[4997]: W1205 08:54:53.914163 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95769d84_8dcc_4626_8f14_e4015d3cfa5e.slice/crio-495716b69e7b17898a5a83c02a5bece7a0366d0291bf55fb0560f5266b0c7b45 WatchSource:0}: Error finding container 495716b69e7b17898a5a83c02a5bece7a0366d0291bf55fb0560f5266b0c7b45: Status 404 returned error can't find the container with id 495716b69e7b17898a5a83c02a5bece7a0366d0291bf55fb0560f5266b0c7b45 Dec 05 08:54:53 crc kubenswrapper[4997]: I1205 08:54:53.915908 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-nlgxp"] Dec 05 08:54:54 crc kubenswrapper[4997]: I1205 08:54:54.426963 4997 generic.go:334] "Generic (PLEG): container finished" podID="95769d84-8dcc-4626-8f14-e4015d3cfa5e" containerID="a41934137ae05c0fe072630de94cbe2af26dfc440898cdc6d8306c4988f20cba" exitCode=0 Dec 05 08:54:54 crc kubenswrapper[4997]: I1205 08:54:54.427058 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-nlgxp" event={"ID":"95769d84-8dcc-4626-8f14-e4015d3cfa5e","Type":"ContainerDied","Data":"a41934137ae05c0fe072630de94cbe2af26dfc440898cdc6d8306c4988f20cba"} Dec 05 08:54:54 crc kubenswrapper[4997]: I1205 08:54:54.427094 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-nlgxp" event={"ID":"95769d84-8dcc-4626-8f14-e4015d3cfa5e","Type":"ContainerStarted","Data":"495716b69e7b17898a5a83c02a5bece7a0366d0291bf55fb0560f5266b0c7b45"} Dec 05 08:54:54 crc kubenswrapper[4997]: I1205 08:54:54.428485 4997 generic.go:334] "Generic (PLEG): container finished" podID="bc83f439-c922-4812-ba87-59c083f3885d" containerID="4eafa7fef7de6ce5f9bf3e3563cb746478083b69063f688fd3f54de37f880b38" exitCode=0 Dec 05 08:54:54 crc kubenswrapper[4997]: I1205 08:54:54.428508 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c014-account-create-update-k8fmb" event={"ID":"bc83f439-c922-4812-ba87-59c083f3885d","Type":"ContainerDied","Data":"4eafa7fef7de6ce5f9bf3e3563cb746478083b69063f688fd3f54de37f880b38"} Dec 05 08:54:54 crc kubenswrapper[4997]: I1205 08:54:54.428521 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c014-account-create-update-k8fmb" event={"ID":"bc83f439-c922-4812-ba87-59c083f3885d","Type":"ContainerStarted","Data":"7a376ca960cfdac04917f4765c96c36f3215d7193e502fae37f58f35b9709d21"} Dec 05 08:54:55 crc kubenswrapper[4997]: I1205 08:54:55.786761 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-nlgxp" Dec 05 08:54:55 crc kubenswrapper[4997]: I1205 08:54:55.793380 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c014-account-create-update-k8fmb" Dec 05 08:54:55 crc kubenswrapper[4997]: I1205 08:54:55.898443 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc83f439-c922-4812-ba87-59c083f3885d-operator-scripts\") pod \"bc83f439-c922-4812-ba87-59c083f3885d\" (UID: \"bc83f439-c922-4812-ba87-59c083f3885d\") " Dec 05 08:54:55 crc kubenswrapper[4997]: I1205 08:54:55.898565 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px8zp\" (UniqueName: \"kubernetes.io/projected/bc83f439-c922-4812-ba87-59c083f3885d-kube-api-access-px8zp\") pod \"bc83f439-c922-4812-ba87-59c083f3885d\" (UID: \"bc83f439-c922-4812-ba87-59c083f3885d\") " Dec 05 08:54:55 crc kubenswrapper[4997]: I1205 08:54:55.898654 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp8fp\" (UniqueName: \"kubernetes.io/projected/95769d84-8dcc-4626-8f14-e4015d3cfa5e-kube-api-access-tp8fp\") pod \"95769d84-8dcc-4626-8f14-e4015d3cfa5e\" (UID: \"95769d84-8dcc-4626-8f14-e4015d3cfa5e\") " Dec 05 08:54:55 crc kubenswrapper[4997]: I1205 08:54:55.898724 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95769d84-8dcc-4626-8f14-e4015d3cfa5e-operator-scripts\") pod \"95769d84-8dcc-4626-8f14-e4015d3cfa5e\" (UID: \"95769d84-8dcc-4626-8f14-e4015d3cfa5e\") " Dec 05 08:54:55 crc kubenswrapper[4997]: I1205 08:54:55.899527 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc83f439-c922-4812-ba87-59c083f3885d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bc83f439-c922-4812-ba87-59c083f3885d" (UID: "bc83f439-c922-4812-ba87-59c083f3885d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:54:55 crc kubenswrapper[4997]: I1205 08:54:55.899538 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95769d84-8dcc-4626-8f14-e4015d3cfa5e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "95769d84-8dcc-4626-8f14-e4015d3cfa5e" (UID: "95769d84-8dcc-4626-8f14-e4015d3cfa5e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:54:55 crc kubenswrapper[4997]: I1205 08:54:55.906134 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc83f439-c922-4812-ba87-59c083f3885d-kube-api-access-px8zp" (OuterVolumeSpecName: "kube-api-access-px8zp") pod "bc83f439-c922-4812-ba87-59c083f3885d" (UID: "bc83f439-c922-4812-ba87-59c083f3885d"). InnerVolumeSpecName "kube-api-access-px8zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:54:55 crc kubenswrapper[4997]: I1205 08:54:55.906943 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95769d84-8dcc-4626-8f14-e4015d3cfa5e-kube-api-access-tp8fp" (OuterVolumeSpecName: "kube-api-access-tp8fp") pod "95769d84-8dcc-4626-8f14-e4015d3cfa5e" (UID: "95769d84-8dcc-4626-8f14-e4015d3cfa5e"). InnerVolumeSpecName "kube-api-access-tp8fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:54:56 crc kubenswrapper[4997]: I1205 08:54:56.001550 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc83f439-c922-4812-ba87-59c083f3885d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:54:56 crc kubenswrapper[4997]: I1205 08:54:56.001651 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px8zp\" (UniqueName: \"kubernetes.io/projected/bc83f439-c922-4812-ba87-59c083f3885d-kube-api-access-px8zp\") on node \"crc\" DevicePath \"\"" Dec 05 08:54:56 crc kubenswrapper[4997]: I1205 08:54:56.001687 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp8fp\" (UniqueName: \"kubernetes.io/projected/95769d84-8dcc-4626-8f14-e4015d3cfa5e-kube-api-access-tp8fp\") on node \"crc\" DevicePath \"\"" Dec 05 08:54:56 crc kubenswrapper[4997]: I1205 08:54:56.001712 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95769d84-8dcc-4626-8f14-e4015d3cfa5e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:54:56 crc kubenswrapper[4997]: I1205 08:54:56.454186 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c014-account-create-update-k8fmb" event={"ID":"bc83f439-c922-4812-ba87-59c083f3885d","Type":"ContainerDied","Data":"7a376ca960cfdac04917f4765c96c36f3215d7193e502fae37f58f35b9709d21"} Dec 05 08:54:56 crc kubenswrapper[4997]: I1205 08:54:56.454227 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a376ca960cfdac04917f4765c96c36f3215d7193e502fae37f58f35b9709d21" Dec 05 08:54:56 crc kubenswrapper[4997]: I1205 08:54:56.454246 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c014-account-create-update-k8fmb" Dec 05 08:54:56 crc kubenswrapper[4997]: I1205 08:54:56.455815 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-nlgxp" event={"ID":"95769d84-8dcc-4626-8f14-e4015d3cfa5e","Type":"ContainerDied","Data":"495716b69e7b17898a5a83c02a5bece7a0366d0291bf55fb0560f5266b0c7b45"} Dec 05 08:54:56 crc kubenswrapper[4997]: I1205 08:54:56.455835 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="495716b69e7b17898a5a83c02a5bece7a0366d0291bf55fb0560f5266b0c7b45" Dec 05 08:54:56 crc kubenswrapper[4997]: I1205 08:54:56.455868 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-nlgxp" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.380802 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-hjb78"] Dec 05 08:54:58 crc kubenswrapper[4997]: E1205 08:54:58.381413 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc83f439-c922-4812-ba87-59c083f3885d" containerName="mariadb-account-create-update" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.381424 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc83f439-c922-4812-ba87-59c083f3885d" containerName="mariadb-account-create-update" Dec 05 08:54:58 crc kubenswrapper[4997]: E1205 08:54:58.381455 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95769d84-8dcc-4626-8f14-e4015d3cfa5e" containerName="mariadb-database-create" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.381476 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="95769d84-8dcc-4626-8f14-e4015d3cfa5e" containerName="mariadb-database-create" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.381670 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="95769d84-8dcc-4626-8f14-e4015d3cfa5e" containerName="mariadb-database-create" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.381688 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc83f439-c922-4812-ba87-59c083f3885d" containerName="mariadb-account-create-update" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.382243 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-hjb78" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.384244 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.384996 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-x844b" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.395003 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-hjb78"] Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.461953 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-db-sync-config-data\") pod \"glance-db-sync-hjb78\" (UID: \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\") " pod="openstack/glance-db-sync-hjb78" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.461994 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-config-data\") pod \"glance-db-sync-hjb78\" (UID: \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\") " pod="openstack/glance-db-sync-hjb78" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.462205 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jksfr\" (UniqueName: \"kubernetes.io/projected/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-kube-api-access-jksfr\") pod \"glance-db-sync-hjb78\" (UID: \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\") " pod="openstack/glance-db-sync-hjb78" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.462400 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-combined-ca-bundle\") pod \"glance-db-sync-hjb78\" (UID: \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\") " pod="openstack/glance-db-sync-hjb78" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.564850 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-db-sync-config-data\") pod \"glance-db-sync-hjb78\" (UID: \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\") " pod="openstack/glance-db-sync-hjb78" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.564933 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-config-data\") pod \"glance-db-sync-hjb78\" (UID: \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\") " pod="openstack/glance-db-sync-hjb78" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.564984 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jksfr\" (UniqueName: \"kubernetes.io/projected/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-kube-api-access-jksfr\") pod \"glance-db-sync-hjb78\" (UID: \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\") " pod="openstack/glance-db-sync-hjb78" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.565045 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-combined-ca-bundle\") pod \"glance-db-sync-hjb78\" (UID: \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\") " pod="openstack/glance-db-sync-hjb78" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.572394 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-combined-ca-bundle\") pod \"glance-db-sync-hjb78\" (UID: \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\") " pod="openstack/glance-db-sync-hjb78" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.572577 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-config-data\") pod \"glance-db-sync-hjb78\" (UID: \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\") " pod="openstack/glance-db-sync-hjb78" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.573044 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-db-sync-config-data\") pod \"glance-db-sync-hjb78\" (UID: \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\") " pod="openstack/glance-db-sync-hjb78" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.589163 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jksfr\" (UniqueName: \"kubernetes.io/projected/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-kube-api-access-jksfr\") pod \"glance-db-sync-hjb78\" (UID: \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\") " pod="openstack/glance-db-sync-hjb78" Dec 05 08:54:58 crc kubenswrapper[4997]: I1205 08:54:58.703361 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-hjb78" Dec 05 08:54:59 crc kubenswrapper[4997]: I1205 08:54:59.259084 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-hjb78"] Dec 05 08:54:59 crc kubenswrapper[4997]: I1205 08:54:59.485224 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-hjb78" event={"ID":"56f98ba4-c397-45b4-b6fa-3c1b2ea93543","Type":"ContainerStarted","Data":"218d8f809ba2e19b9b1da46bb9eb2b9658899b35c4b29efb1efd29fe8a27d1c0"} Dec 05 08:55:16 crc kubenswrapper[4997]: I1205 08:55:16.629982 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-hjb78" event={"ID":"56f98ba4-c397-45b4-b6fa-3c1b2ea93543","Type":"ContainerStarted","Data":"52186b60dbc9147c4a35f0872c0d85a34dda95d6728765468f540cf913486964"} Dec 05 08:55:16 crc kubenswrapper[4997]: I1205 08:55:16.655658 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-hjb78" podStartSLOduration=2.46737894 podStartE2EDuration="18.655641411s" podCreationTimestamp="2025-12-05 08:54:58 +0000 UTC" firstStartedPulling="2025-12-05 08:54:59.268600547 +0000 UTC m=+7199.797507818" lastFinishedPulling="2025-12-05 08:55:15.456863028 +0000 UTC m=+7215.985770289" observedRunningTime="2025-12-05 08:55:16.645941151 +0000 UTC m=+7217.174848442" watchObservedRunningTime="2025-12-05 08:55:16.655641411 +0000 UTC m=+7217.184548672" Dec 05 08:55:19 crc kubenswrapper[4997]: I1205 08:55:19.676402 4997 generic.go:334] "Generic (PLEG): container finished" podID="56f98ba4-c397-45b4-b6fa-3c1b2ea93543" containerID="52186b60dbc9147c4a35f0872c0d85a34dda95d6728765468f540cf913486964" exitCode=0 Dec 05 08:55:19 crc kubenswrapper[4997]: I1205 08:55:19.676471 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-hjb78" event={"ID":"56f98ba4-c397-45b4-b6fa-3c1b2ea93543","Type":"ContainerDied","Data":"52186b60dbc9147c4a35f0872c0d85a34dda95d6728765468f540cf913486964"} Dec 05 08:55:19 crc kubenswrapper[4997]: I1205 08:55:19.770336 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:55:19 crc kubenswrapper[4997]: I1205 08:55:19.770412 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:55:19 crc kubenswrapper[4997]: I1205 08:55:19.770464 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 08:55:19 crc kubenswrapper[4997]: I1205 08:55:19.771216 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c7247cb7e5929fb7303801e82d11af6c5303b8ce2c4cf98c220b4780c5e2c9f4"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:55:19 crc kubenswrapper[4997]: I1205 08:55:19.771294 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://c7247cb7e5929fb7303801e82d11af6c5303b8ce2c4cf98c220b4780c5e2c9f4" gracePeriod=600 Dec 05 08:55:20 crc kubenswrapper[4997]: I1205 08:55:20.688125 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="c7247cb7e5929fb7303801e82d11af6c5303b8ce2c4cf98c220b4780c5e2c9f4" exitCode=0 Dec 05 08:55:20 crc kubenswrapper[4997]: I1205 08:55:20.688163 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"c7247cb7e5929fb7303801e82d11af6c5303b8ce2c4cf98c220b4780c5e2c9f4"} Dec 05 08:55:20 crc kubenswrapper[4997]: I1205 08:55:20.688750 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b"} Dec 05 08:55:20 crc kubenswrapper[4997]: I1205 08:55:20.688771 4997 scope.go:117] "RemoveContainer" containerID="76c48e979de43bd5503b9e67119a68e71b8cfc214afdb394fc5584ecc0e930b9" Dec 05 08:55:21 crc kubenswrapper[4997]: I1205 08:55:21.139828 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-hjb78" Dec 05 08:55:21 crc kubenswrapper[4997]: I1205 08:55:21.270953 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-db-sync-config-data\") pod \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\" (UID: \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\") " Dec 05 08:55:21 crc kubenswrapper[4997]: I1205 08:55:21.271032 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-combined-ca-bundle\") pod \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\" (UID: \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\") " Dec 05 08:55:21 crc kubenswrapper[4997]: I1205 08:55:21.271061 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-config-data\") pod \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\" (UID: \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\") " Dec 05 08:55:21 crc kubenswrapper[4997]: I1205 08:55:21.271134 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jksfr\" (UniqueName: \"kubernetes.io/projected/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-kube-api-access-jksfr\") pod \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\" (UID: \"56f98ba4-c397-45b4-b6fa-3c1b2ea93543\") " Dec 05 08:55:21 crc kubenswrapper[4997]: I1205 08:55:21.277823 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "56f98ba4-c397-45b4-b6fa-3c1b2ea93543" (UID: "56f98ba4-c397-45b4-b6fa-3c1b2ea93543"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:55:21 crc kubenswrapper[4997]: I1205 08:55:21.277875 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-kube-api-access-jksfr" (OuterVolumeSpecName: "kube-api-access-jksfr") pod "56f98ba4-c397-45b4-b6fa-3c1b2ea93543" (UID: "56f98ba4-c397-45b4-b6fa-3c1b2ea93543"). InnerVolumeSpecName "kube-api-access-jksfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:55:21 crc kubenswrapper[4997]: I1205 08:55:21.297688 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56f98ba4-c397-45b4-b6fa-3c1b2ea93543" (UID: "56f98ba4-c397-45b4-b6fa-3c1b2ea93543"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:55:21 crc kubenswrapper[4997]: I1205 08:55:21.343094 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-config-data" (OuterVolumeSpecName: "config-data") pod "56f98ba4-c397-45b4-b6fa-3c1b2ea93543" (UID: "56f98ba4-c397-45b4-b6fa-3c1b2ea93543"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:55:21 crc kubenswrapper[4997]: I1205 08:55:21.373290 4997 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:21 crc kubenswrapper[4997]: I1205 08:55:21.373513 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:21 crc kubenswrapper[4997]: I1205 08:55:21.373680 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:21 crc kubenswrapper[4997]: I1205 08:55:21.373768 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jksfr\" (UniqueName: \"kubernetes.io/projected/56f98ba4-c397-45b4-b6fa-3c1b2ea93543-kube-api-access-jksfr\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:21 crc kubenswrapper[4997]: I1205 08:55:21.699049 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-hjb78" event={"ID":"56f98ba4-c397-45b4-b6fa-3c1b2ea93543","Type":"ContainerDied","Data":"218d8f809ba2e19b9b1da46bb9eb2b9658899b35c4b29efb1efd29fe8a27d1c0"} Dec 05 08:55:21 crc kubenswrapper[4997]: I1205 08:55:21.699306 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="218d8f809ba2e19b9b1da46bb9eb2b9658899b35c4b29efb1efd29fe8a27d1c0" Dec 05 08:55:21 crc kubenswrapper[4997]: I1205 08:55:21.699105 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-hjb78" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.001631 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:55:22 crc kubenswrapper[4997]: E1205 08:55:22.002035 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56f98ba4-c397-45b4-b6fa-3c1b2ea93543" containerName="glance-db-sync" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.002056 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="56f98ba4-c397-45b4-b6fa-3c1b2ea93543" containerName="glance-db-sync" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.002346 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="56f98ba4-c397-45b4-b6fa-3c1b2ea93543" containerName="glance-db-sync" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.003586 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.006225 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.006373 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.006902 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-x844b" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.006952 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.025364 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.149187 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-555d9b49dc-k978w"] Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.165860 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.177385 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-555d9b49dc-k978w"] Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.187634 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4474ddaa-5888-472f-9866-0b99247e6561-ceph\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.187785 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zclnx\" (UniqueName: \"kubernetes.io/projected/4474ddaa-5888-472f-9866-0b99247e6561-kube-api-access-zclnx\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.187865 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-scripts\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.187891 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4474ddaa-5888-472f-9866-0b99247e6561-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.187908 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4474ddaa-5888-472f-9866-0b99247e6561-logs\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.187945 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.188200 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-config-data\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.279650 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.281468 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.282926 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.289245 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zclnx\" (UniqueName: \"kubernetes.io/projected/4474ddaa-5888-472f-9866-0b99247e6561-kube-api-access-zclnx\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.289319 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6xxk\" (UniqueName: \"kubernetes.io/projected/131a9488-9bd7-4f21-8b0b-b11fc8e46691-kube-api-access-j6xxk\") pod \"dnsmasq-dns-555d9b49dc-k978w\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.289343 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-config\") pod \"dnsmasq-dns-555d9b49dc-k978w\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.289363 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-scripts\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.289382 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4474ddaa-5888-472f-9866-0b99247e6561-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.289400 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4474ddaa-5888-472f-9866-0b99247e6561-logs\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.289415 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.289454 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-ovsdbserver-nb\") pod \"dnsmasq-dns-555d9b49dc-k978w\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.289487 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-config-data\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.289519 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4474ddaa-5888-472f-9866-0b99247e6561-ceph\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.289541 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-ovsdbserver-sb\") pod \"dnsmasq-dns-555d9b49dc-k978w\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.289584 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-dns-svc\") pod \"dnsmasq-dns-555d9b49dc-k978w\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.290549 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4474ddaa-5888-472f-9866-0b99247e6561-logs\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.291906 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4474ddaa-5888-472f-9866-0b99247e6561-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.296339 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4474ddaa-5888-472f-9866-0b99247e6561-ceph\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.298158 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.298751 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-config-data\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.302516 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.321477 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-scripts\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.330781 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zclnx\" (UniqueName: \"kubernetes.io/projected/4474ddaa-5888-472f-9866-0b99247e6561-kube-api-access-zclnx\") pod \"glance-default-external-api-0\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.360156 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.392095 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-ovsdbserver-sb\") pod \"dnsmasq-dns-555d9b49dc-k978w\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.392166 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47x7l\" (UniqueName: \"kubernetes.io/projected/7e397301-7287-4e69-8095-c76949d79b44-kube-api-access-47x7l\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.392198 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.392234 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.392266 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-dns-svc\") pod \"dnsmasq-dns-555d9b49dc-k978w\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.392302 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e397301-7287-4e69-8095-c76949d79b44-logs\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.392331 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7e397301-7287-4e69-8095-c76949d79b44-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.392357 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6xxk\" (UniqueName: \"kubernetes.io/projected/131a9488-9bd7-4f21-8b0b-b11fc8e46691-kube-api-access-j6xxk\") pod \"dnsmasq-dns-555d9b49dc-k978w\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.392382 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-config\") pod \"dnsmasq-dns-555d9b49dc-k978w\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.392408 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.392465 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-ovsdbserver-nb\") pod \"dnsmasq-dns-555d9b49dc-k978w\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.392527 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7e397301-7287-4e69-8095-c76949d79b44-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.394006 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-ovsdbserver-sb\") pod \"dnsmasq-dns-555d9b49dc-k978w\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.394675 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-dns-svc\") pod \"dnsmasq-dns-555d9b49dc-k978w\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.396059 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-ovsdbserver-nb\") pod \"dnsmasq-dns-555d9b49dc-k978w\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.396076 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-config\") pod \"dnsmasq-dns-555d9b49dc-k978w\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.414668 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6xxk\" (UniqueName: \"kubernetes.io/projected/131a9488-9bd7-4f21-8b0b-b11fc8e46691-kube-api-access-j6xxk\") pod \"dnsmasq-dns-555d9b49dc-k978w\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.494423 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7e397301-7287-4e69-8095-c76949d79b44-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.495004 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47x7l\" (UniqueName: \"kubernetes.io/projected/7e397301-7287-4e69-8095-c76949d79b44-kube-api-access-47x7l\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.495034 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.495067 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.495107 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e397301-7287-4e69-8095-c76949d79b44-logs\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.495136 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7e397301-7287-4e69-8095-c76949d79b44-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.495172 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.502560 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.502699 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e397301-7287-4e69-8095-c76949d79b44-logs\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.502739 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7e397301-7287-4e69-8095-c76949d79b44-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.502908 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.506118 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7e397301-7287-4e69-8095-c76949d79b44-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.511345 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.538444 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47x7l\" (UniqueName: \"kubernetes.io/projected/7e397301-7287-4e69-8095-c76949d79b44-kube-api-access-47x7l\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.538544 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:22 crc kubenswrapper[4997]: I1205 08:55:22.685023 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 08:55:23 crc kubenswrapper[4997]: I1205 08:55:22.950047 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:55:23 crc kubenswrapper[4997]: I1205 08:55:23.016384 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-555d9b49dc-k978w"] Dec 05 08:55:23 crc kubenswrapper[4997]: W1205 08:55:23.017810 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod131a9488_9bd7_4f21_8b0b_b11fc8e46691.slice/crio-14d077de915174ff29a44b8bb72166390310f78cf6f2c3e2c07a538982c215a9 WatchSource:0}: Error finding container 14d077de915174ff29a44b8bb72166390310f78cf6f2c3e2c07a538982c215a9: Status 404 returned error can't find the container with id 14d077de915174ff29a44b8bb72166390310f78cf6f2c3e2c07a538982c215a9 Dec 05 08:55:23 crc kubenswrapper[4997]: I1205 08:55:23.075801 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:55:23 crc kubenswrapper[4997]: I1205 08:55:23.723577 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4474ddaa-5888-472f-9866-0b99247e6561","Type":"ContainerStarted","Data":"9a23bf378be9c98a519b50aaacf629df91c3062b8a054ee3a58cc4b2e2f15d2f"} Dec 05 08:55:23 crc kubenswrapper[4997]: I1205 08:55:23.723653 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4474ddaa-5888-472f-9866-0b99247e6561","Type":"ContainerStarted","Data":"db15160caf290c5693c6d0a1ec7e95af7f408ed38714ba3df32548b61fd27be7"} Dec 05 08:55:23 crc kubenswrapper[4997]: I1205 08:55:23.724930 4997 generic.go:334] "Generic (PLEG): container finished" podID="131a9488-9bd7-4f21-8b0b-b11fc8e46691" containerID="f4bce0aed54031d7c536ad4febdb31fba9bb353415985507c2abb3f603e3355e" exitCode=0 Dec 05 08:55:23 crc kubenswrapper[4997]: I1205 08:55:23.724957 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-555d9b49dc-k978w" event={"ID":"131a9488-9bd7-4f21-8b0b-b11fc8e46691","Type":"ContainerDied","Data":"f4bce0aed54031d7c536ad4febdb31fba9bb353415985507c2abb3f603e3355e"} Dec 05 08:55:23 crc kubenswrapper[4997]: I1205 08:55:23.724976 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-555d9b49dc-k978w" event={"ID":"131a9488-9bd7-4f21-8b0b-b11fc8e46691","Type":"ContainerStarted","Data":"14d077de915174ff29a44b8bb72166390310f78cf6f2c3e2c07a538982c215a9"} Dec 05 08:55:23 crc kubenswrapper[4997]: I1205 08:55:23.962328 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:55:23 crc kubenswrapper[4997]: W1205 08:55:23.970191 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e397301_7287_4e69_8095_c76949d79b44.slice/crio-4c124efde882ede9843f5c2f8e2b8438eff125fbfa9b50fe1e2ea8a29421b326 WatchSource:0}: Error finding container 4c124efde882ede9843f5c2f8e2b8438eff125fbfa9b50fe1e2ea8a29421b326: Status 404 returned error can't find the container with id 4c124efde882ede9843f5c2f8e2b8438eff125fbfa9b50fe1e2ea8a29421b326 Dec 05 08:55:24 crc kubenswrapper[4997]: I1205 08:55:24.742797 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4474ddaa-5888-472f-9866-0b99247e6561","Type":"ContainerStarted","Data":"16196b9fa634c6291b3ef1c3d01108aa86a9384a65a78e03450e654e554cd935"} Dec 05 08:55:24 crc kubenswrapper[4997]: I1205 08:55:24.742911 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4474ddaa-5888-472f-9866-0b99247e6561" containerName="glance-log" containerID="cri-o://9a23bf378be9c98a519b50aaacf629df91c3062b8a054ee3a58cc4b2e2f15d2f" gracePeriod=30 Dec 05 08:55:24 crc kubenswrapper[4997]: I1205 08:55:24.742996 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4474ddaa-5888-472f-9866-0b99247e6561" containerName="glance-httpd" containerID="cri-o://16196b9fa634c6291b3ef1c3d01108aa86a9384a65a78e03450e654e554cd935" gracePeriod=30 Dec 05 08:55:24 crc kubenswrapper[4997]: I1205 08:55:24.744509 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7e397301-7287-4e69-8095-c76949d79b44","Type":"ContainerStarted","Data":"08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28"} Dec 05 08:55:24 crc kubenswrapper[4997]: I1205 08:55:24.744537 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7e397301-7287-4e69-8095-c76949d79b44","Type":"ContainerStarted","Data":"4c124efde882ede9843f5c2f8e2b8438eff125fbfa9b50fe1e2ea8a29421b326"} Dec 05 08:55:24 crc kubenswrapper[4997]: I1205 08:55:24.749476 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-555d9b49dc-k978w" event={"ID":"131a9488-9bd7-4f21-8b0b-b11fc8e46691","Type":"ContainerStarted","Data":"df368418de303f483e46371634027d33a7bfe5556cf5b9ee2c03a85a10c0c30d"} Dec 05 08:55:24 crc kubenswrapper[4997]: I1205 08:55:24.749692 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:24 crc kubenswrapper[4997]: I1205 08:55:24.776590 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.776570007 podStartE2EDuration="3.776570007s" podCreationTimestamp="2025-12-05 08:55:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:55:24.771157992 +0000 UTC m=+7225.300065273" watchObservedRunningTime="2025-12-05 08:55:24.776570007 +0000 UTC m=+7225.305477268" Dec 05 08:55:24 crc kubenswrapper[4997]: I1205 08:55:24.791756 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-555d9b49dc-k978w" podStartSLOduration=2.791735965 podStartE2EDuration="2.791735965s" podCreationTimestamp="2025-12-05 08:55:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:55:24.788113758 +0000 UTC m=+7225.317021019" watchObservedRunningTime="2025-12-05 08:55:24.791735965 +0000 UTC m=+7225.320643226" Dec 05 08:55:25 crc kubenswrapper[4997]: I1205 08:55:25.661321 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:55:25 crc kubenswrapper[4997]: I1205 08:55:25.762898 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7e397301-7287-4e69-8095-c76949d79b44","Type":"ContainerStarted","Data":"a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454"} Dec 05 08:55:25 crc kubenswrapper[4997]: I1205 08:55:25.768078 4997 generic.go:334] "Generic (PLEG): container finished" podID="4474ddaa-5888-472f-9866-0b99247e6561" containerID="16196b9fa634c6291b3ef1c3d01108aa86a9384a65a78e03450e654e554cd935" exitCode=0 Dec 05 08:55:25 crc kubenswrapper[4997]: I1205 08:55:25.768127 4997 generic.go:334] "Generic (PLEG): container finished" podID="4474ddaa-5888-472f-9866-0b99247e6561" containerID="9a23bf378be9c98a519b50aaacf629df91c3062b8a054ee3a58cc4b2e2f15d2f" exitCode=143 Dec 05 08:55:25 crc kubenswrapper[4997]: I1205 08:55:25.768166 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4474ddaa-5888-472f-9866-0b99247e6561","Type":"ContainerDied","Data":"16196b9fa634c6291b3ef1c3d01108aa86a9384a65a78e03450e654e554cd935"} Dec 05 08:55:25 crc kubenswrapper[4997]: I1205 08:55:25.768212 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4474ddaa-5888-472f-9866-0b99247e6561","Type":"ContainerDied","Data":"9a23bf378be9c98a519b50aaacf629df91c3062b8a054ee3a58cc4b2e2f15d2f"} Dec 05 08:55:25 crc kubenswrapper[4997]: I1205 08:55:25.789342 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.789318351 podStartE2EDuration="3.789318351s" podCreationTimestamp="2025-12-05 08:55:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:55:25.786167486 +0000 UTC m=+7226.315074757" watchObservedRunningTime="2025-12-05 08:55:25.789318351 +0000 UTC m=+7226.318225632" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.039044 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.189564 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-scripts\") pod \"4474ddaa-5888-472f-9866-0b99247e6561\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.189631 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zclnx\" (UniqueName: \"kubernetes.io/projected/4474ddaa-5888-472f-9866-0b99247e6561-kube-api-access-zclnx\") pod \"4474ddaa-5888-472f-9866-0b99247e6561\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.189664 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4474ddaa-5888-472f-9866-0b99247e6561-ceph\") pod \"4474ddaa-5888-472f-9866-0b99247e6561\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.189782 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4474ddaa-5888-472f-9866-0b99247e6561-logs\") pod \"4474ddaa-5888-472f-9866-0b99247e6561\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.189834 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4474ddaa-5888-472f-9866-0b99247e6561-httpd-run\") pod \"4474ddaa-5888-472f-9866-0b99247e6561\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.189865 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-config-data\") pod \"4474ddaa-5888-472f-9866-0b99247e6561\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.189920 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-combined-ca-bundle\") pod \"4474ddaa-5888-472f-9866-0b99247e6561\" (UID: \"4474ddaa-5888-472f-9866-0b99247e6561\") " Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.191016 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4474ddaa-5888-472f-9866-0b99247e6561-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4474ddaa-5888-472f-9866-0b99247e6561" (UID: "4474ddaa-5888-472f-9866-0b99247e6561"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.191095 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4474ddaa-5888-472f-9866-0b99247e6561-logs" (OuterVolumeSpecName: "logs") pod "4474ddaa-5888-472f-9866-0b99247e6561" (UID: "4474ddaa-5888-472f-9866-0b99247e6561"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.195937 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4474ddaa-5888-472f-9866-0b99247e6561-ceph" (OuterVolumeSpecName: "ceph") pod "4474ddaa-5888-472f-9866-0b99247e6561" (UID: "4474ddaa-5888-472f-9866-0b99247e6561"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.196119 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4474ddaa-5888-472f-9866-0b99247e6561-kube-api-access-zclnx" (OuterVolumeSpecName: "kube-api-access-zclnx") pod "4474ddaa-5888-472f-9866-0b99247e6561" (UID: "4474ddaa-5888-472f-9866-0b99247e6561"). InnerVolumeSpecName "kube-api-access-zclnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.197730 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-scripts" (OuterVolumeSpecName: "scripts") pod "4474ddaa-5888-472f-9866-0b99247e6561" (UID: "4474ddaa-5888-472f-9866-0b99247e6561"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.221300 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4474ddaa-5888-472f-9866-0b99247e6561" (UID: "4474ddaa-5888-472f-9866-0b99247e6561"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.244459 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-config-data" (OuterVolumeSpecName: "config-data") pod "4474ddaa-5888-472f-9866-0b99247e6561" (UID: "4474ddaa-5888-472f-9866-0b99247e6561"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.292455 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.292519 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zclnx\" (UniqueName: \"kubernetes.io/projected/4474ddaa-5888-472f-9866-0b99247e6561-kube-api-access-zclnx\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.292537 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4474ddaa-5888-472f-9866-0b99247e6561-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.292549 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4474ddaa-5888-472f-9866-0b99247e6561-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.292560 4997 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4474ddaa-5888-472f-9866-0b99247e6561-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.292572 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.292583 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4474ddaa-5888-472f-9866-0b99247e6561-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.785860 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7e397301-7287-4e69-8095-c76949d79b44" containerName="glance-log" containerID="cri-o://08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28" gracePeriod=30 Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.786007 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4474ddaa-5888-472f-9866-0b99247e6561","Type":"ContainerDied","Data":"db15160caf290c5693c6d0a1ec7e95af7f408ed38714ba3df32548b61fd27be7"} Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.786112 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.786501 4997 scope.go:117] "RemoveContainer" containerID="16196b9fa634c6291b3ef1c3d01108aa86a9384a65a78e03450e654e554cd935" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.786162 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7e397301-7287-4e69-8095-c76949d79b44" containerName="glance-httpd" containerID="cri-o://a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454" gracePeriod=30 Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.831858 4997 scope.go:117] "RemoveContainer" containerID="9a23bf378be9c98a519b50aaacf629df91c3062b8a054ee3a58cc4b2e2f15d2f" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.864188 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.875945 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.909340 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:55:26 crc kubenswrapper[4997]: E1205 08:55:26.918945 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4474ddaa-5888-472f-9866-0b99247e6561" containerName="glance-httpd" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.918979 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4474ddaa-5888-472f-9866-0b99247e6561" containerName="glance-httpd" Dec 05 08:55:26 crc kubenswrapper[4997]: E1205 08:55:26.918995 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4474ddaa-5888-472f-9866-0b99247e6561" containerName="glance-log" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.919002 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4474ddaa-5888-472f-9866-0b99247e6561" containerName="glance-log" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.919180 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="4474ddaa-5888-472f-9866-0b99247e6561" containerName="glance-httpd" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.919203 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="4474ddaa-5888-472f-9866-0b99247e6561" containerName="glance-log" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.920826 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.924201 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 05 08:55:26 crc kubenswrapper[4997]: I1205 08:55:26.947485 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.004424 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-config-data\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.004497 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw87h\" (UniqueName: \"kubernetes.io/projected/733c0bd0-5736-4e03-a1f1-511bec38afa5-kube-api-access-rw87h\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.004565 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/733c0bd0-5736-4e03-a1f1-511bec38afa5-logs\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.004595 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/733c0bd0-5736-4e03-a1f1-511bec38afa5-ceph\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.004630 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/733c0bd0-5736-4e03-a1f1-511bec38afa5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.004650 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.004809 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-scripts\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.106793 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/733c0bd0-5736-4e03-a1f1-511bec38afa5-logs\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.107050 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/733c0bd0-5736-4e03-a1f1-511bec38afa5-ceph\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.107087 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/733c0bd0-5736-4e03-a1f1-511bec38afa5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.107115 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.107157 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-scripts\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.107264 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-config-data\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.107320 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw87h\" (UniqueName: \"kubernetes.io/projected/733c0bd0-5736-4e03-a1f1-511bec38afa5-kube-api-access-rw87h\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.107533 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/733c0bd0-5736-4e03-a1f1-511bec38afa5-logs\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.108479 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/733c0bd0-5736-4e03-a1f1-511bec38afa5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.112525 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-scripts\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.112706 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.120942 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/733c0bd0-5736-4e03-a1f1-511bec38afa5-ceph\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.123169 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-config-data\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.125205 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw87h\" (UniqueName: \"kubernetes.io/projected/733c0bd0-5736-4e03-a1f1-511bec38afa5-kube-api-access-rw87h\") pod \"glance-default-external-api-0\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.370202 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.380867 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.516085 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-config-data\") pod \"7e397301-7287-4e69-8095-c76949d79b44\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.516175 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7e397301-7287-4e69-8095-c76949d79b44-ceph\") pod \"7e397301-7287-4e69-8095-c76949d79b44\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.516220 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-combined-ca-bundle\") pod \"7e397301-7287-4e69-8095-c76949d79b44\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.516263 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7e397301-7287-4e69-8095-c76949d79b44-httpd-run\") pod \"7e397301-7287-4e69-8095-c76949d79b44\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.516404 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47x7l\" (UniqueName: \"kubernetes.io/projected/7e397301-7287-4e69-8095-c76949d79b44-kube-api-access-47x7l\") pod \"7e397301-7287-4e69-8095-c76949d79b44\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.516441 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-scripts\") pod \"7e397301-7287-4e69-8095-c76949d79b44\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.516548 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e397301-7287-4e69-8095-c76949d79b44-logs\") pod \"7e397301-7287-4e69-8095-c76949d79b44\" (UID: \"7e397301-7287-4e69-8095-c76949d79b44\") " Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.517392 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e397301-7287-4e69-8095-c76949d79b44-logs" (OuterVolumeSpecName: "logs") pod "7e397301-7287-4e69-8095-c76949d79b44" (UID: "7e397301-7287-4e69-8095-c76949d79b44"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.517913 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e397301-7287-4e69-8095-c76949d79b44-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7e397301-7287-4e69-8095-c76949d79b44" (UID: "7e397301-7287-4e69-8095-c76949d79b44"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.523775 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-scripts" (OuterVolumeSpecName: "scripts") pod "7e397301-7287-4e69-8095-c76949d79b44" (UID: "7e397301-7287-4e69-8095-c76949d79b44"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.523877 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e397301-7287-4e69-8095-c76949d79b44-kube-api-access-47x7l" (OuterVolumeSpecName: "kube-api-access-47x7l") pod "7e397301-7287-4e69-8095-c76949d79b44" (UID: "7e397301-7287-4e69-8095-c76949d79b44"). InnerVolumeSpecName "kube-api-access-47x7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.524091 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e397301-7287-4e69-8095-c76949d79b44-ceph" (OuterVolumeSpecName: "ceph") pod "7e397301-7287-4e69-8095-c76949d79b44" (UID: "7e397301-7287-4e69-8095-c76949d79b44"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.548215 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e397301-7287-4e69-8095-c76949d79b44" (UID: "7e397301-7287-4e69-8095-c76949d79b44"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.612937 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-config-data" (OuterVolumeSpecName: "config-data") pod "7e397301-7287-4e69-8095-c76949d79b44" (UID: "7e397301-7287-4e69-8095-c76949d79b44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.618977 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e397301-7287-4e69-8095-c76949d79b44-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.619015 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.619027 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7e397301-7287-4e69-8095-c76949d79b44-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.619038 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.619049 4997 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7e397301-7287-4e69-8095-c76949d79b44-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.619061 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47x7l\" (UniqueName: \"kubernetes.io/projected/7e397301-7287-4e69-8095-c76949d79b44-kube-api-access-47x7l\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.619072 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e397301-7287-4e69-8095-c76949d79b44-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.761206 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4474ddaa-5888-472f-9866-0b99247e6561" path="/var/lib/kubelet/pods/4474ddaa-5888-472f-9866-0b99247e6561/volumes" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.817275 4997 generic.go:334] "Generic (PLEG): container finished" podID="7e397301-7287-4e69-8095-c76949d79b44" containerID="a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454" exitCode=0 Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.817305 4997 generic.go:334] "Generic (PLEG): container finished" podID="7e397301-7287-4e69-8095-c76949d79b44" containerID="08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28" exitCode=143 Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.817322 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7e397301-7287-4e69-8095-c76949d79b44","Type":"ContainerDied","Data":"a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454"} Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.817346 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7e397301-7287-4e69-8095-c76949d79b44","Type":"ContainerDied","Data":"08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28"} Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.817359 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7e397301-7287-4e69-8095-c76949d79b44","Type":"ContainerDied","Data":"4c124efde882ede9843f5c2f8e2b8438eff125fbfa9b50fe1e2ea8a29421b326"} Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.817373 4997 scope.go:117] "RemoveContainer" containerID="a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.817466 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.846257 4997 scope.go:117] "RemoveContainer" containerID="08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.869970 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.881774 4997 scope.go:117] "RemoveContainer" containerID="a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454" Dec 05 08:55:27 crc kubenswrapper[4997]: E1205 08:55:27.884981 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454\": container with ID starting with a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454 not found: ID does not exist" containerID="a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.885027 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454"} err="failed to get container status \"a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454\": rpc error: code = NotFound desc = could not find container \"a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454\": container with ID starting with a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454 not found: ID does not exist" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.885062 4997 scope.go:117] "RemoveContainer" containerID="08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28" Dec 05 08:55:27 crc kubenswrapper[4997]: E1205 08:55:27.885334 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28\": container with ID starting with 08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28 not found: ID does not exist" containerID="08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.885364 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28"} err="failed to get container status \"08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28\": rpc error: code = NotFound desc = could not find container \"08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28\": container with ID starting with 08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28 not found: ID does not exist" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.885382 4997 scope.go:117] "RemoveContainer" containerID="a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.885600 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454"} err="failed to get container status \"a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454\": rpc error: code = NotFound desc = could not find container \"a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454\": container with ID starting with a1b23d233ded8c73c76f73c28dc73c7f96be34b60f3e00a91f09ce2349ac7454 not found: ID does not exist" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.885725 4997 scope.go:117] "RemoveContainer" containerID="08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.886002 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28"} err="failed to get container status \"08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28\": rpc error: code = NotFound desc = could not find container \"08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28\": container with ID starting with 08f457b96160e869123ccd479af28cb582c146eee6964e51ee8f8e533f471e28 not found: ID does not exist" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.886848 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.911866 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:55:27 crc kubenswrapper[4997]: E1205 08:55:27.912455 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e397301-7287-4e69-8095-c76949d79b44" containerName="glance-log" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.912486 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e397301-7287-4e69-8095-c76949d79b44" containerName="glance-log" Dec 05 08:55:27 crc kubenswrapper[4997]: E1205 08:55:27.912516 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e397301-7287-4e69-8095-c76949d79b44" containerName="glance-httpd" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.912525 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e397301-7287-4e69-8095-c76949d79b44" containerName="glance-httpd" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.912856 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e397301-7287-4e69-8095-c76949d79b44" containerName="glance-httpd" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.912905 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e397301-7287-4e69-8095-c76949d79b44" containerName="glance-log" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.914845 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.922563 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 08:55:27 crc kubenswrapper[4997]: I1205 08:55:27.961177 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.036708 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0983ed44-4f02-4751-9452-5f11f3b93f63-logs\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.036764 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.036801 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.036840 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwzn2\" (UniqueName: \"kubernetes.io/projected/0983ed44-4f02-4751-9452-5f11f3b93f63-kube-api-access-kwzn2\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.036862 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0983ed44-4f02-4751-9452-5f11f3b93f63-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.036896 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0983ed44-4f02-4751-9452-5f11f3b93f63-ceph\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.036918 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.096140 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.138860 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwzn2\" (UniqueName: \"kubernetes.io/projected/0983ed44-4f02-4751-9452-5f11f3b93f63-kube-api-access-kwzn2\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.138923 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0983ed44-4f02-4751-9452-5f11f3b93f63-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.138976 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0983ed44-4f02-4751-9452-5f11f3b93f63-ceph\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.139002 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.139075 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0983ed44-4f02-4751-9452-5f11f3b93f63-logs\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.139111 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.139148 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.139631 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0983ed44-4f02-4751-9452-5f11f3b93f63-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.139639 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0983ed44-4f02-4751-9452-5f11f3b93f63-logs\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.145015 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0983ed44-4f02-4751-9452-5f11f3b93f63-ceph\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.145193 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.145907 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.146071 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.157391 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwzn2\" (UniqueName: \"kubernetes.io/projected/0983ed44-4f02-4751-9452-5f11f3b93f63-kube-api-access-kwzn2\") pod \"glance-default-internal-api-0\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.314241 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.682389 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:55:28 crc kubenswrapper[4997]: W1205 08:55:28.683940 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0983ed44_4f02_4751_9452_5f11f3b93f63.slice/crio-4a4fe9700e9911685553c713d615525ae0f51e09fa4a569e56ea8eb330876c76 WatchSource:0}: Error finding container 4a4fe9700e9911685553c713d615525ae0f51e09fa4a569e56ea8eb330876c76: Status 404 returned error can't find the container with id 4a4fe9700e9911685553c713d615525ae0f51e09fa4a569e56ea8eb330876c76 Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.826885 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0983ed44-4f02-4751-9452-5f11f3b93f63","Type":"ContainerStarted","Data":"4a4fe9700e9911685553c713d615525ae0f51e09fa4a569e56ea8eb330876c76"} Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.830252 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"733c0bd0-5736-4e03-a1f1-511bec38afa5","Type":"ContainerStarted","Data":"8b51d4553f38927a8858bdd05f11beca552b9840963ba79d51f84d4acf38c78d"} Dec 05 08:55:28 crc kubenswrapper[4997]: I1205 08:55:28.830298 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"733c0bd0-5736-4e03-a1f1-511bec38afa5","Type":"ContainerStarted","Data":"d32bae6e0b7b66e761af3c3c095e3ce992e22a147f70f75c5879e8c61ed921d7"} Dec 05 08:55:29 crc kubenswrapper[4997]: I1205 08:55:29.763755 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e397301-7287-4e69-8095-c76949d79b44" path="/var/lib/kubelet/pods/7e397301-7287-4e69-8095-c76949d79b44/volumes" Dec 05 08:55:29 crc kubenswrapper[4997]: I1205 08:55:29.845693 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0983ed44-4f02-4751-9452-5f11f3b93f63","Type":"ContainerStarted","Data":"ee588007f28e21e91e570b04e85a69b74ab3215ab7c4d7a965388796b2558902"} Dec 05 08:55:29 crc kubenswrapper[4997]: I1205 08:55:29.845739 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0983ed44-4f02-4751-9452-5f11f3b93f63","Type":"ContainerStarted","Data":"ae9ff29232f7941f7167b41adcc8e22cc47ae1c89feeaf93f421084fd933f3c6"} Dec 05 08:55:29 crc kubenswrapper[4997]: I1205 08:55:29.850104 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"733c0bd0-5736-4e03-a1f1-511bec38afa5","Type":"ContainerStarted","Data":"90d31f83804ed13a91183dda31d25b53af7d8d17aef41209f1ad59dc5b899cd0"} Dec 05 08:55:29 crc kubenswrapper[4997]: I1205 08:55:29.873694 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.873666471 podStartE2EDuration="2.873666471s" podCreationTimestamp="2025-12-05 08:55:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:55:29.863645781 +0000 UTC m=+7230.392553062" watchObservedRunningTime="2025-12-05 08:55:29.873666471 +0000 UTC m=+7230.402573732" Dec 05 08:55:29 crc kubenswrapper[4997]: I1205 08:55:29.904784 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.904760787 podStartE2EDuration="3.904760787s" podCreationTimestamp="2025-12-05 08:55:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:55:29.897533532 +0000 UTC m=+7230.426440813" watchObservedRunningTime="2025-12-05 08:55:29.904760787 +0000 UTC m=+7230.433668048" Dec 05 08:55:32 crc kubenswrapper[4997]: I1205 08:55:32.506826 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:55:32 crc kubenswrapper[4997]: I1205 08:55:32.570406 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67877c9485-2ncv4"] Dec 05 08:55:32 crc kubenswrapper[4997]: I1205 08:55:32.575179 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67877c9485-2ncv4" podUID="c9475029-c206-4392-b272-ee631f5abf8b" containerName="dnsmasq-dns" containerID="cri-o://d8bbcddb7772b4d820225aa9f06e6438aca9cd7bfefda3b3a6554dbe35204231" gracePeriod=10 Dec 05 08:55:33 crc kubenswrapper[4997]: I1205 08:55:33.926751 4997 generic.go:334] "Generic (PLEG): container finished" podID="c9475029-c206-4392-b272-ee631f5abf8b" containerID="d8bbcddb7772b4d820225aa9f06e6438aca9cd7bfefda3b3a6554dbe35204231" exitCode=0 Dec 05 08:55:33 crc kubenswrapper[4997]: I1205 08:55:33.926864 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67877c9485-2ncv4" event={"ID":"c9475029-c206-4392-b272-ee631f5abf8b","Type":"ContainerDied","Data":"d8bbcddb7772b4d820225aa9f06e6438aca9cd7bfefda3b3a6554dbe35204231"} Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.152283 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.262034 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-ovsdbserver-sb\") pod \"c9475029-c206-4392-b272-ee631f5abf8b\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.263014 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-ovsdbserver-nb\") pod \"c9475029-c206-4392-b272-ee631f5abf8b\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.263139 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-config\") pod \"c9475029-c206-4392-b272-ee631f5abf8b\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.263208 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-dns-svc\") pod \"c9475029-c206-4392-b272-ee631f5abf8b\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.263280 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvbg9\" (UniqueName: \"kubernetes.io/projected/c9475029-c206-4392-b272-ee631f5abf8b-kube-api-access-bvbg9\") pod \"c9475029-c206-4392-b272-ee631f5abf8b\" (UID: \"c9475029-c206-4392-b272-ee631f5abf8b\") " Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.271754 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9475029-c206-4392-b272-ee631f5abf8b-kube-api-access-bvbg9" (OuterVolumeSpecName: "kube-api-access-bvbg9") pod "c9475029-c206-4392-b272-ee631f5abf8b" (UID: "c9475029-c206-4392-b272-ee631f5abf8b"). InnerVolumeSpecName "kube-api-access-bvbg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.334427 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c9475029-c206-4392-b272-ee631f5abf8b" (UID: "c9475029-c206-4392-b272-ee631f5abf8b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.340329 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c9475029-c206-4392-b272-ee631f5abf8b" (UID: "c9475029-c206-4392-b272-ee631f5abf8b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.358275 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-config" (OuterVolumeSpecName: "config") pod "c9475029-c206-4392-b272-ee631f5abf8b" (UID: "c9475029-c206-4392-b272-ee631f5abf8b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.359366 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c9475029-c206-4392-b272-ee631f5abf8b" (UID: "c9475029-c206-4392-b272-ee631f5abf8b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.365562 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.365589 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.365597 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.365608 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvbg9\" (UniqueName: \"kubernetes.io/projected/c9475029-c206-4392-b272-ee631f5abf8b-kube-api-access-bvbg9\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.365633 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9475029-c206-4392-b272-ee631f5abf8b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.937958 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67877c9485-2ncv4" event={"ID":"c9475029-c206-4392-b272-ee631f5abf8b","Type":"ContainerDied","Data":"6cc899b3d18321e905f58fb7a8f7819c4d3cb8b60a91147a55d66e2cab2e35ff"} Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.938273 4997 scope.go:117] "RemoveContainer" containerID="d8bbcddb7772b4d820225aa9f06e6438aca9cd7bfefda3b3a6554dbe35204231" Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.938056 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67877c9485-2ncv4" Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.968178 4997 scope.go:117] "RemoveContainer" containerID="1fa05a1beb6bb7ab8978174b8575098972507095532fa138ca44613b99d39083" Dec 05 08:55:34 crc kubenswrapper[4997]: I1205 08:55:34.989146 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67877c9485-2ncv4"] Dec 05 08:55:35 crc kubenswrapper[4997]: I1205 08:55:35.002121 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67877c9485-2ncv4"] Dec 05 08:55:35 crc kubenswrapper[4997]: I1205 08:55:35.759610 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9475029-c206-4392-b272-ee631f5abf8b" path="/var/lib/kubelet/pods/c9475029-c206-4392-b272-ee631f5abf8b/volumes" Dec 05 08:55:37 crc kubenswrapper[4997]: I1205 08:55:37.371819 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 08:55:37 crc kubenswrapper[4997]: I1205 08:55:37.371906 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 08:55:37 crc kubenswrapper[4997]: I1205 08:55:37.430840 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 08:55:37 crc kubenswrapper[4997]: I1205 08:55:37.443302 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 08:55:37 crc kubenswrapper[4997]: I1205 08:55:37.962037 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 08:55:37 crc kubenswrapper[4997]: I1205 08:55:37.962084 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 08:55:38 crc kubenswrapper[4997]: I1205 08:55:38.315291 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 08:55:38 crc kubenswrapper[4997]: I1205 08:55:38.316424 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 08:55:38 crc kubenswrapper[4997]: I1205 08:55:38.369651 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 08:55:38 crc kubenswrapper[4997]: I1205 08:55:38.377996 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 08:55:38 crc kubenswrapper[4997]: I1205 08:55:38.976358 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 08:55:38 crc kubenswrapper[4997]: I1205 08:55:38.976407 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 08:55:40 crc kubenswrapper[4997]: I1205 08:55:40.022899 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 08:55:40 crc kubenswrapper[4997]: I1205 08:55:40.023249 4997 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 08:55:40 crc kubenswrapper[4997]: I1205 08:55:40.029365 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 08:55:40 crc kubenswrapper[4997]: I1205 08:55:40.984884 4997 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 08:55:40 crc kubenswrapper[4997]: I1205 08:55:40.984918 4997 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 08:55:41 crc kubenswrapper[4997]: I1205 08:55:41.032524 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 08:55:41 crc kubenswrapper[4997]: I1205 08:55:41.034486 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 08:55:50 crc kubenswrapper[4997]: I1205 08:55:50.985369 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-vb5ts"] Dec 05 08:55:50 crc kubenswrapper[4997]: E1205 08:55:50.987520 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9475029-c206-4392-b272-ee631f5abf8b" containerName="dnsmasq-dns" Dec 05 08:55:50 crc kubenswrapper[4997]: I1205 08:55:50.987556 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9475029-c206-4392-b272-ee631f5abf8b" containerName="dnsmasq-dns" Dec 05 08:55:50 crc kubenswrapper[4997]: E1205 08:55:50.987577 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9475029-c206-4392-b272-ee631f5abf8b" containerName="init" Dec 05 08:55:50 crc kubenswrapper[4997]: I1205 08:55:50.987584 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9475029-c206-4392-b272-ee631f5abf8b" containerName="init" Dec 05 08:55:50 crc kubenswrapper[4997]: I1205 08:55:50.988004 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9475029-c206-4392-b272-ee631f5abf8b" containerName="dnsmasq-dns" Dec 05 08:55:50 crc kubenswrapper[4997]: I1205 08:55:50.988884 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vb5ts" Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.002512 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-3d5f-account-create-update-v4cs5"] Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.005505 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3d5f-account-create-update-v4cs5" Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.007186 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.012992 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-vb5ts"] Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.022067 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-3d5f-account-create-update-v4cs5"] Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.130250 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgqdc\" (UniqueName: \"kubernetes.io/projected/4ccc1288-f7d9-4963-8efa-1e2392030be5-kube-api-access-kgqdc\") pod \"placement-db-create-vb5ts\" (UID: \"4ccc1288-f7d9-4963-8efa-1e2392030be5\") " pod="openstack/placement-db-create-vb5ts" Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.130456 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ccc1288-f7d9-4963-8efa-1e2392030be5-operator-scripts\") pod \"placement-db-create-vb5ts\" (UID: \"4ccc1288-f7d9-4963-8efa-1e2392030be5\") " pod="openstack/placement-db-create-vb5ts" Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.130690 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldbk2\" (UniqueName: \"kubernetes.io/projected/e40fa138-3d33-42be-b39e-6b8c1533323b-kube-api-access-ldbk2\") pod \"placement-3d5f-account-create-update-v4cs5\" (UID: \"e40fa138-3d33-42be-b39e-6b8c1533323b\") " pod="openstack/placement-3d5f-account-create-update-v4cs5" Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.130857 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e40fa138-3d33-42be-b39e-6b8c1533323b-operator-scripts\") pod \"placement-3d5f-account-create-update-v4cs5\" (UID: \"e40fa138-3d33-42be-b39e-6b8c1533323b\") " pod="openstack/placement-3d5f-account-create-update-v4cs5" Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.232572 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e40fa138-3d33-42be-b39e-6b8c1533323b-operator-scripts\") pod \"placement-3d5f-account-create-update-v4cs5\" (UID: \"e40fa138-3d33-42be-b39e-6b8c1533323b\") " pod="openstack/placement-3d5f-account-create-update-v4cs5" Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.232717 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgqdc\" (UniqueName: \"kubernetes.io/projected/4ccc1288-f7d9-4963-8efa-1e2392030be5-kube-api-access-kgqdc\") pod \"placement-db-create-vb5ts\" (UID: \"4ccc1288-f7d9-4963-8efa-1e2392030be5\") " pod="openstack/placement-db-create-vb5ts" Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.232801 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ccc1288-f7d9-4963-8efa-1e2392030be5-operator-scripts\") pod \"placement-db-create-vb5ts\" (UID: \"4ccc1288-f7d9-4963-8efa-1e2392030be5\") " pod="openstack/placement-db-create-vb5ts" Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.232858 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldbk2\" (UniqueName: \"kubernetes.io/projected/e40fa138-3d33-42be-b39e-6b8c1533323b-kube-api-access-ldbk2\") pod \"placement-3d5f-account-create-update-v4cs5\" (UID: \"e40fa138-3d33-42be-b39e-6b8c1533323b\") " pod="openstack/placement-3d5f-account-create-update-v4cs5" Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.233462 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e40fa138-3d33-42be-b39e-6b8c1533323b-operator-scripts\") pod \"placement-3d5f-account-create-update-v4cs5\" (UID: \"e40fa138-3d33-42be-b39e-6b8c1533323b\") " pod="openstack/placement-3d5f-account-create-update-v4cs5" Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.233743 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ccc1288-f7d9-4963-8efa-1e2392030be5-operator-scripts\") pod \"placement-db-create-vb5ts\" (UID: \"4ccc1288-f7d9-4963-8efa-1e2392030be5\") " pod="openstack/placement-db-create-vb5ts" Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.256979 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldbk2\" (UniqueName: \"kubernetes.io/projected/e40fa138-3d33-42be-b39e-6b8c1533323b-kube-api-access-ldbk2\") pod \"placement-3d5f-account-create-update-v4cs5\" (UID: \"e40fa138-3d33-42be-b39e-6b8c1533323b\") " pod="openstack/placement-3d5f-account-create-update-v4cs5" Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.258247 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgqdc\" (UniqueName: \"kubernetes.io/projected/4ccc1288-f7d9-4963-8efa-1e2392030be5-kube-api-access-kgqdc\") pod \"placement-db-create-vb5ts\" (UID: \"4ccc1288-f7d9-4963-8efa-1e2392030be5\") " pod="openstack/placement-db-create-vb5ts" Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.354369 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vb5ts" Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.367349 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3d5f-account-create-update-v4cs5" Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.714335 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-3d5f-account-create-update-v4cs5"] Dec 05 08:55:51 crc kubenswrapper[4997]: W1205 08:55:51.717106 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode40fa138_3d33_42be_b39e_6b8c1533323b.slice/crio-4e725bdcd136d6a626b2656c4e89c3391a29de99c2f5487ee8c95862e2bcb6e3 WatchSource:0}: Error finding container 4e725bdcd136d6a626b2656c4e89c3391a29de99c2f5487ee8c95862e2bcb6e3: Status 404 returned error can't find the container with id 4e725bdcd136d6a626b2656c4e89c3391a29de99c2f5487ee8c95862e2bcb6e3 Dec 05 08:55:51 crc kubenswrapper[4997]: I1205 08:55:51.840357 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-vb5ts"] Dec 05 08:55:51 crc kubenswrapper[4997]: W1205 08:55:51.840740 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ccc1288_f7d9_4963_8efa_1e2392030be5.slice/crio-1a2fb7c9b8ebf17a97c73afbb6149e1c6c86d4561f63c2b823ab903dc7d0a590 WatchSource:0}: Error finding container 1a2fb7c9b8ebf17a97c73afbb6149e1c6c86d4561f63c2b823ab903dc7d0a590: Status 404 returned error can't find the container with id 1a2fb7c9b8ebf17a97c73afbb6149e1c6c86d4561f63c2b823ab903dc7d0a590 Dec 05 08:55:52 crc kubenswrapper[4997]: I1205 08:55:52.075071 4997 generic.go:334] "Generic (PLEG): container finished" podID="e40fa138-3d33-42be-b39e-6b8c1533323b" containerID="dfd4f3305734811a291c363bcfb98fae4e5dc21b8cac4081c8f81957ac03ec08" exitCode=0 Dec 05 08:55:52 crc kubenswrapper[4997]: I1205 08:55:52.075153 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3d5f-account-create-update-v4cs5" event={"ID":"e40fa138-3d33-42be-b39e-6b8c1533323b","Type":"ContainerDied","Data":"dfd4f3305734811a291c363bcfb98fae4e5dc21b8cac4081c8f81957ac03ec08"} Dec 05 08:55:52 crc kubenswrapper[4997]: I1205 08:55:52.075177 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3d5f-account-create-update-v4cs5" event={"ID":"e40fa138-3d33-42be-b39e-6b8c1533323b","Type":"ContainerStarted","Data":"4e725bdcd136d6a626b2656c4e89c3391a29de99c2f5487ee8c95862e2bcb6e3"} Dec 05 08:55:52 crc kubenswrapper[4997]: I1205 08:55:52.078431 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vb5ts" event={"ID":"4ccc1288-f7d9-4963-8efa-1e2392030be5","Type":"ContainerStarted","Data":"968e6edc631cecf7c42ee56b31588268f8277c6db305828b4e3f368c78a688d2"} Dec 05 08:55:52 crc kubenswrapper[4997]: I1205 08:55:52.078582 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vb5ts" event={"ID":"4ccc1288-f7d9-4963-8efa-1e2392030be5","Type":"ContainerStarted","Data":"1a2fb7c9b8ebf17a97c73afbb6149e1c6c86d4561f63c2b823ab903dc7d0a590"} Dec 05 08:55:52 crc kubenswrapper[4997]: I1205 08:55:52.104059 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-vb5ts" podStartSLOduration=2.10402181 podStartE2EDuration="2.10402181s" podCreationTimestamp="2025-12-05 08:55:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:55:52.103211107 +0000 UTC m=+7252.632118388" watchObservedRunningTime="2025-12-05 08:55:52.10402181 +0000 UTC m=+7252.632929071" Dec 05 08:55:53 crc kubenswrapper[4997]: I1205 08:55:53.089057 4997 generic.go:334] "Generic (PLEG): container finished" podID="4ccc1288-f7d9-4963-8efa-1e2392030be5" containerID="968e6edc631cecf7c42ee56b31588268f8277c6db305828b4e3f368c78a688d2" exitCode=0 Dec 05 08:55:53 crc kubenswrapper[4997]: I1205 08:55:53.089128 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vb5ts" event={"ID":"4ccc1288-f7d9-4963-8efa-1e2392030be5","Type":"ContainerDied","Data":"968e6edc631cecf7c42ee56b31588268f8277c6db305828b4e3f368c78a688d2"} Dec 05 08:55:53 crc kubenswrapper[4997]: I1205 08:55:53.452031 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3d5f-account-create-update-v4cs5" Dec 05 08:55:53 crc kubenswrapper[4997]: I1205 08:55:53.573412 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldbk2\" (UniqueName: \"kubernetes.io/projected/e40fa138-3d33-42be-b39e-6b8c1533323b-kube-api-access-ldbk2\") pod \"e40fa138-3d33-42be-b39e-6b8c1533323b\" (UID: \"e40fa138-3d33-42be-b39e-6b8c1533323b\") " Dec 05 08:55:53 crc kubenswrapper[4997]: I1205 08:55:53.573782 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e40fa138-3d33-42be-b39e-6b8c1533323b-operator-scripts\") pod \"e40fa138-3d33-42be-b39e-6b8c1533323b\" (UID: \"e40fa138-3d33-42be-b39e-6b8c1533323b\") " Dec 05 08:55:53 crc kubenswrapper[4997]: I1205 08:55:53.574311 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e40fa138-3d33-42be-b39e-6b8c1533323b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e40fa138-3d33-42be-b39e-6b8c1533323b" (UID: "e40fa138-3d33-42be-b39e-6b8c1533323b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:55:53 crc kubenswrapper[4997]: I1205 08:55:53.578854 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e40fa138-3d33-42be-b39e-6b8c1533323b-kube-api-access-ldbk2" (OuterVolumeSpecName: "kube-api-access-ldbk2") pod "e40fa138-3d33-42be-b39e-6b8c1533323b" (UID: "e40fa138-3d33-42be-b39e-6b8c1533323b"). InnerVolumeSpecName "kube-api-access-ldbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:55:53 crc kubenswrapper[4997]: I1205 08:55:53.676409 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldbk2\" (UniqueName: \"kubernetes.io/projected/e40fa138-3d33-42be-b39e-6b8c1533323b-kube-api-access-ldbk2\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:53 crc kubenswrapper[4997]: I1205 08:55:53.676445 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e40fa138-3d33-42be-b39e-6b8c1533323b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:54 crc kubenswrapper[4997]: I1205 08:55:54.102266 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3d5f-account-create-update-v4cs5" Dec 05 08:55:54 crc kubenswrapper[4997]: I1205 08:55:54.103445 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3d5f-account-create-update-v4cs5" event={"ID":"e40fa138-3d33-42be-b39e-6b8c1533323b","Type":"ContainerDied","Data":"4e725bdcd136d6a626b2656c4e89c3391a29de99c2f5487ee8c95862e2bcb6e3"} Dec 05 08:55:54 crc kubenswrapper[4997]: I1205 08:55:54.103534 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e725bdcd136d6a626b2656c4e89c3391a29de99c2f5487ee8c95862e2bcb6e3" Dec 05 08:55:54 crc kubenswrapper[4997]: I1205 08:55:54.403647 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vb5ts" Dec 05 08:55:54 crc kubenswrapper[4997]: I1205 08:55:54.491667 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ccc1288-f7d9-4963-8efa-1e2392030be5-operator-scripts\") pod \"4ccc1288-f7d9-4963-8efa-1e2392030be5\" (UID: \"4ccc1288-f7d9-4963-8efa-1e2392030be5\") " Dec 05 08:55:54 crc kubenswrapper[4997]: I1205 08:55:54.491768 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgqdc\" (UniqueName: \"kubernetes.io/projected/4ccc1288-f7d9-4963-8efa-1e2392030be5-kube-api-access-kgqdc\") pod \"4ccc1288-f7d9-4963-8efa-1e2392030be5\" (UID: \"4ccc1288-f7d9-4963-8efa-1e2392030be5\") " Dec 05 08:55:54 crc kubenswrapper[4997]: I1205 08:55:54.492823 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ccc1288-f7d9-4963-8efa-1e2392030be5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4ccc1288-f7d9-4963-8efa-1e2392030be5" (UID: "4ccc1288-f7d9-4963-8efa-1e2392030be5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:55:54 crc kubenswrapper[4997]: I1205 08:55:54.496912 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ccc1288-f7d9-4963-8efa-1e2392030be5-kube-api-access-kgqdc" (OuterVolumeSpecName: "kube-api-access-kgqdc") pod "4ccc1288-f7d9-4963-8efa-1e2392030be5" (UID: "4ccc1288-f7d9-4963-8efa-1e2392030be5"). InnerVolumeSpecName "kube-api-access-kgqdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:55:54 crc kubenswrapper[4997]: I1205 08:55:54.593774 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ccc1288-f7d9-4963-8efa-1e2392030be5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:54 crc kubenswrapper[4997]: I1205 08:55:54.593810 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgqdc\" (UniqueName: \"kubernetes.io/projected/4ccc1288-f7d9-4963-8efa-1e2392030be5-kube-api-access-kgqdc\") on node \"crc\" DevicePath \"\"" Dec 05 08:55:55 crc kubenswrapper[4997]: I1205 08:55:55.115898 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vb5ts" event={"ID":"4ccc1288-f7d9-4963-8efa-1e2392030be5","Type":"ContainerDied","Data":"1a2fb7c9b8ebf17a97c73afbb6149e1c6c86d4561f63c2b823ab903dc7d0a590"} Dec 05 08:55:55 crc kubenswrapper[4997]: I1205 08:55:55.115962 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a2fb7c9b8ebf17a97c73afbb6149e1c6c86d4561f63c2b823ab903dc7d0a590" Dec 05 08:55:55 crc kubenswrapper[4997]: I1205 08:55:55.115989 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vb5ts" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.402657 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d66fd5885-nb4nr"] Dec 05 08:55:56 crc kubenswrapper[4997]: E1205 08:55:56.404142 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e40fa138-3d33-42be-b39e-6b8c1533323b" containerName="mariadb-account-create-update" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.404236 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e40fa138-3d33-42be-b39e-6b8c1533323b" containerName="mariadb-account-create-update" Dec 05 08:55:56 crc kubenswrapper[4997]: E1205 08:55:56.404331 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ccc1288-f7d9-4963-8efa-1e2392030be5" containerName="mariadb-database-create" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.404389 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ccc1288-f7d9-4963-8efa-1e2392030be5" containerName="mariadb-database-create" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.406232 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ccc1288-f7d9-4963-8efa-1e2392030be5" containerName="mariadb-database-create" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.406365 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e40fa138-3d33-42be-b39e-6b8c1533323b" containerName="mariadb-account-create-update" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.407444 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.422295 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d66fd5885-nb4nr"] Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.437464 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-7959t"] Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.439126 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7959t" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.441736 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.444663 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-2m8xp" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.444988 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.465704 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-7959t"] Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.530462 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-ovsdbserver-sb\") pod \"dnsmasq-dns-d66fd5885-nb4nr\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.530878 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd6h5\" (UniqueName: \"kubernetes.io/projected/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-kube-api-access-bd6h5\") pod \"placement-db-sync-7959t\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " pod="openstack/placement-db-sync-7959t" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.530937 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-logs\") pod \"placement-db-sync-7959t\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " pod="openstack/placement-db-sync-7959t" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.530996 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzddl\" (UniqueName: \"kubernetes.io/projected/d9d96a8f-1144-4680-952a-37f0580bfec0-kube-api-access-wzddl\") pod \"dnsmasq-dns-d66fd5885-nb4nr\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.531079 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-config\") pod \"dnsmasq-dns-d66fd5885-nb4nr\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.531128 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-dns-svc\") pod \"dnsmasq-dns-d66fd5885-nb4nr\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.531270 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-scripts\") pod \"placement-db-sync-7959t\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " pod="openstack/placement-db-sync-7959t" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.531336 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-combined-ca-bundle\") pod \"placement-db-sync-7959t\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " pod="openstack/placement-db-sync-7959t" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.531392 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-config-data\") pod \"placement-db-sync-7959t\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " pod="openstack/placement-db-sync-7959t" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.531474 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-ovsdbserver-nb\") pod \"dnsmasq-dns-d66fd5885-nb4nr\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.633245 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-config\") pod \"dnsmasq-dns-d66fd5885-nb4nr\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.633289 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-dns-svc\") pod \"dnsmasq-dns-d66fd5885-nb4nr\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.633314 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-scripts\") pod \"placement-db-sync-7959t\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " pod="openstack/placement-db-sync-7959t" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.633334 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-combined-ca-bundle\") pod \"placement-db-sync-7959t\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " pod="openstack/placement-db-sync-7959t" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.633357 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-config-data\") pod \"placement-db-sync-7959t\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " pod="openstack/placement-db-sync-7959t" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.633380 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-ovsdbserver-nb\") pod \"dnsmasq-dns-d66fd5885-nb4nr\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.633474 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-ovsdbserver-sb\") pod \"dnsmasq-dns-d66fd5885-nb4nr\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.633536 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd6h5\" (UniqueName: \"kubernetes.io/projected/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-kube-api-access-bd6h5\") pod \"placement-db-sync-7959t\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " pod="openstack/placement-db-sync-7959t" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.633559 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-logs\") pod \"placement-db-sync-7959t\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " pod="openstack/placement-db-sync-7959t" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.633580 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzddl\" (UniqueName: \"kubernetes.io/projected/d9d96a8f-1144-4680-952a-37f0580bfec0-kube-api-access-wzddl\") pod \"dnsmasq-dns-d66fd5885-nb4nr\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.634304 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-logs\") pod \"placement-db-sync-7959t\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " pod="openstack/placement-db-sync-7959t" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.634375 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-config\") pod \"dnsmasq-dns-d66fd5885-nb4nr\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.634422 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-dns-svc\") pod \"dnsmasq-dns-d66fd5885-nb4nr\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.634548 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-ovsdbserver-sb\") pod \"dnsmasq-dns-d66fd5885-nb4nr\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.635116 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-ovsdbserver-nb\") pod \"dnsmasq-dns-d66fd5885-nb4nr\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.639206 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-scripts\") pod \"placement-db-sync-7959t\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " pod="openstack/placement-db-sync-7959t" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.640313 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-combined-ca-bundle\") pod \"placement-db-sync-7959t\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " pod="openstack/placement-db-sync-7959t" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.647118 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-config-data\") pod \"placement-db-sync-7959t\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " pod="openstack/placement-db-sync-7959t" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.649819 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzddl\" (UniqueName: \"kubernetes.io/projected/d9d96a8f-1144-4680-952a-37f0580bfec0-kube-api-access-wzddl\") pod \"dnsmasq-dns-d66fd5885-nb4nr\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.650277 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd6h5\" (UniqueName: \"kubernetes.io/projected/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-kube-api-access-bd6h5\") pod \"placement-db-sync-7959t\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " pod="openstack/placement-db-sync-7959t" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.730037 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:55:56 crc kubenswrapper[4997]: I1205 08:55:56.770798 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7959t" Dec 05 08:55:57 crc kubenswrapper[4997]: I1205 08:55:57.062947 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-7959t"] Dec 05 08:55:57 crc kubenswrapper[4997]: I1205 08:55:57.142826 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7959t" event={"ID":"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924","Type":"ContainerStarted","Data":"3abd5b18767da80299ba0a74ba33544364598c5361a934eb5d7d33693953093a"} Dec 05 08:55:57 crc kubenswrapper[4997]: I1205 08:55:57.221058 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d66fd5885-nb4nr"] Dec 05 08:55:58 crc kubenswrapper[4997]: I1205 08:55:58.153414 4997 generic.go:334] "Generic (PLEG): container finished" podID="d9d96a8f-1144-4680-952a-37f0580bfec0" containerID="d79d83516589ac4159831c7598ca2713d4b370dd4a870b8f97f91f9777d638c3" exitCode=0 Dec 05 08:55:58 crc kubenswrapper[4997]: I1205 08:55:58.153502 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" event={"ID":"d9d96a8f-1144-4680-952a-37f0580bfec0","Type":"ContainerDied","Data":"d79d83516589ac4159831c7598ca2713d4b370dd4a870b8f97f91f9777d638c3"} Dec 05 08:55:58 crc kubenswrapper[4997]: I1205 08:55:58.154028 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" event={"ID":"d9d96a8f-1144-4680-952a-37f0580bfec0","Type":"ContainerStarted","Data":"03f42fff35f42b8a91a40558d6ff22427bf83037b938383342863bcdd2ba74e0"} Dec 05 08:56:02 crc kubenswrapper[4997]: I1205 08:56:02.188934 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" event={"ID":"d9d96a8f-1144-4680-952a-37f0580bfec0","Type":"ContainerStarted","Data":"392ec1087f285d4ec9ff0f0f251b0edb60f288b2189931b51a95c7c4e534f891"} Dec 05 08:56:02 crc kubenswrapper[4997]: I1205 08:56:02.189499 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:56:02 crc kubenswrapper[4997]: I1205 08:56:02.222468 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" podStartSLOduration=6.222446928 podStartE2EDuration="6.222446928s" podCreationTimestamp="2025-12-05 08:55:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:56:02.215978765 +0000 UTC m=+7262.744886046" watchObservedRunningTime="2025-12-05 08:56:02.222446928 +0000 UTC m=+7262.751354189" Dec 05 08:56:05 crc kubenswrapper[4997]: I1205 08:56:05.219703 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7959t" event={"ID":"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924","Type":"ContainerStarted","Data":"bd5e7d9c0313564c3df38c8b5a3e5f433d878996a240eda09e395ae2b708f35c"} Dec 05 08:56:05 crc kubenswrapper[4997]: I1205 08:56:05.249111 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-7959t" podStartSLOduration=2.672408461 podStartE2EDuration="9.249089607s" podCreationTimestamp="2025-12-05 08:55:56 +0000 UTC" firstStartedPulling="2025-12-05 08:55:57.068144887 +0000 UTC m=+7257.597052148" lastFinishedPulling="2025-12-05 08:56:03.644826033 +0000 UTC m=+7264.173733294" observedRunningTime="2025-12-05 08:56:05.243292471 +0000 UTC m=+7265.772199742" watchObservedRunningTime="2025-12-05 08:56:05.249089607 +0000 UTC m=+7265.777996878" Dec 05 08:56:06 crc kubenswrapper[4997]: I1205 08:56:06.731771 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:56:06 crc kubenswrapper[4997]: I1205 08:56:06.781603 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-555d9b49dc-k978w"] Dec 05 08:56:06 crc kubenswrapper[4997]: I1205 08:56:06.781910 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-555d9b49dc-k978w" podUID="131a9488-9bd7-4f21-8b0b-b11fc8e46691" containerName="dnsmasq-dns" containerID="cri-o://df368418de303f483e46371634027d33a7bfe5556cf5b9ee2c03a85a10c0c30d" gracePeriod=10 Dec 05 08:56:07 crc kubenswrapper[4997]: I1205 08:56:07.236838 4997 generic.go:334] "Generic (PLEG): container finished" podID="131a9488-9bd7-4f21-8b0b-b11fc8e46691" containerID="df368418de303f483e46371634027d33a7bfe5556cf5b9ee2c03a85a10c0c30d" exitCode=0 Dec 05 08:56:07 crc kubenswrapper[4997]: I1205 08:56:07.236938 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-555d9b49dc-k978w" event={"ID":"131a9488-9bd7-4f21-8b0b-b11fc8e46691","Type":"ContainerDied","Data":"df368418de303f483e46371634027d33a7bfe5556cf5b9ee2c03a85a10c0c30d"} Dec 05 08:56:07 crc kubenswrapper[4997]: I1205 08:56:07.505240 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-555d9b49dc-k978w" podUID="131a9488-9bd7-4f21-8b0b-b11fc8e46691" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.56:5353: connect: connection refused" Dec 05 08:56:08 crc kubenswrapper[4997]: I1205 08:56:08.558317 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:56:08 crc kubenswrapper[4997]: I1205 08:56:08.651492 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-dns-svc\") pod \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " Dec 05 08:56:08 crc kubenswrapper[4997]: I1205 08:56:08.651584 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-config\") pod \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " Dec 05 08:56:08 crc kubenswrapper[4997]: I1205 08:56:08.651658 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-ovsdbserver-nb\") pod \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " Dec 05 08:56:08 crc kubenswrapper[4997]: I1205 08:56:08.651703 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-ovsdbserver-sb\") pod \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " Dec 05 08:56:08 crc kubenswrapper[4997]: I1205 08:56:08.651733 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6xxk\" (UniqueName: \"kubernetes.io/projected/131a9488-9bd7-4f21-8b0b-b11fc8e46691-kube-api-access-j6xxk\") pod \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\" (UID: \"131a9488-9bd7-4f21-8b0b-b11fc8e46691\") " Dec 05 08:56:08 crc kubenswrapper[4997]: I1205 08:56:08.656812 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/131a9488-9bd7-4f21-8b0b-b11fc8e46691-kube-api-access-j6xxk" (OuterVolumeSpecName: "kube-api-access-j6xxk") pod "131a9488-9bd7-4f21-8b0b-b11fc8e46691" (UID: "131a9488-9bd7-4f21-8b0b-b11fc8e46691"). InnerVolumeSpecName "kube-api-access-j6xxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:56:08 crc kubenswrapper[4997]: I1205 08:56:08.697983 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "131a9488-9bd7-4f21-8b0b-b11fc8e46691" (UID: "131a9488-9bd7-4f21-8b0b-b11fc8e46691"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:56:08 crc kubenswrapper[4997]: I1205 08:56:08.706761 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "131a9488-9bd7-4f21-8b0b-b11fc8e46691" (UID: "131a9488-9bd7-4f21-8b0b-b11fc8e46691"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:56:08 crc kubenswrapper[4997]: I1205 08:56:08.706999 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "131a9488-9bd7-4f21-8b0b-b11fc8e46691" (UID: "131a9488-9bd7-4f21-8b0b-b11fc8e46691"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:56:08 crc kubenswrapper[4997]: I1205 08:56:08.721391 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-config" (OuterVolumeSpecName: "config") pod "131a9488-9bd7-4f21-8b0b-b11fc8e46691" (UID: "131a9488-9bd7-4f21-8b0b-b11fc8e46691"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:56:08 crc kubenswrapper[4997]: I1205 08:56:08.754172 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:56:08 crc kubenswrapper[4997]: I1205 08:56:08.754210 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:56:08 crc kubenswrapper[4997]: I1205 08:56:08.754220 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:56:08 crc kubenswrapper[4997]: I1205 08:56:08.754235 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/131a9488-9bd7-4f21-8b0b-b11fc8e46691-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:56:08 crc kubenswrapper[4997]: I1205 08:56:08.754246 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6xxk\" (UniqueName: \"kubernetes.io/projected/131a9488-9bd7-4f21-8b0b-b11fc8e46691-kube-api-access-j6xxk\") on node \"crc\" DevicePath \"\"" Dec 05 08:56:09 crc kubenswrapper[4997]: I1205 08:56:09.256994 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-555d9b49dc-k978w" event={"ID":"131a9488-9bd7-4f21-8b0b-b11fc8e46691","Type":"ContainerDied","Data":"14d077de915174ff29a44b8bb72166390310f78cf6f2c3e2c07a538982c215a9"} Dec 05 08:56:09 crc kubenswrapper[4997]: I1205 08:56:09.257057 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-555d9b49dc-k978w" Dec 05 08:56:09 crc kubenswrapper[4997]: I1205 08:56:09.257069 4997 scope.go:117] "RemoveContainer" containerID="df368418de303f483e46371634027d33a7bfe5556cf5b9ee2c03a85a10c0c30d" Dec 05 08:56:09 crc kubenswrapper[4997]: I1205 08:56:09.290231 4997 scope.go:117] "RemoveContainer" containerID="f4bce0aed54031d7c536ad4febdb31fba9bb353415985507c2abb3f603e3355e" Dec 05 08:56:09 crc kubenswrapper[4997]: I1205 08:56:09.305041 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-555d9b49dc-k978w"] Dec 05 08:56:09 crc kubenswrapper[4997]: I1205 08:56:09.312643 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-555d9b49dc-k978w"] Dec 05 08:56:09 crc kubenswrapper[4997]: I1205 08:56:09.761236 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="131a9488-9bd7-4f21-8b0b-b11fc8e46691" path="/var/lib/kubelet/pods/131a9488-9bd7-4f21-8b0b-b11fc8e46691/volumes" Dec 05 08:56:10 crc kubenswrapper[4997]: I1205 08:56:10.270091 4997 generic.go:334] "Generic (PLEG): container finished" podID="0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924" containerID="bd5e7d9c0313564c3df38c8b5a3e5f433d878996a240eda09e395ae2b708f35c" exitCode=0 Dec 05 08:56:10 crc kubenswrapper[4997]: I1205 08:56:10.270173 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7959t" event={"ID":"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924","Type":"ContainerDied","Data":"bd5e7d9c0313564c3df38c8b5a3e5f433d878996a240eda09e395ae2b708f35c"} Dec 05 08:56:11 crc kubenswrapper[4997]: I1205 08:56:11.609240 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7959t" Dec 05 08:56:11 crc kubenswrapper[4997]: I1205 08:56:11.704729 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bd6h5\" (UniqueName: \"kubernetes.io/projected/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-kube-api-access-bd6h5\") pod \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " Dec 05 08:56:11 crc kubenswrapper[4997]: I1205 08:56:11.704899 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-config-data\") pod \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " Dec 05 08:56:11 crc kubenswrapper[4997]: I1205 08:56:11.705049 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-logs\") pod \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " Dec 05 08:56:11 crc kubenswrapper[4997]: I1205 08:56:11.705225 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-scripts\") pod \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " Dec 05 08:56:11 crc kubenswrapper[4997]: I1205 08:56:11.705324 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-combined-ca-bundle\") pod \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\" (UID: \"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924\") " Dec 05 08:56:11 crc kubenswrapper[4997]: I1205 08:56:11.705665 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-logs" (OuterVolumeSpecName: "logs") pod "0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924" (UID: "0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:56:11 crc kubenswrapper[4997]: I1205 08:56:11.710213 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-scripts" (OuterVolumeSpecName: "scripts") pod "0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924" (UID: "0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:56:11 crc kubenswrapper[4997]: I1205 08:56:11.710597 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-kube-api-access-bd6h5" (OuterVolumeSpecName: "kube-api-access-bd6h5") pod "0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924" (UID: "0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924"). InnerVolumeSpecName "kube-api-access-bd6h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:56:11 crc kubenswrapper[4997]: I1205 08:56:11.730194 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924" (UID: "0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:56:11 crc kubenswrapper[4997]: I1205 08:56:11.733964 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-config-data" (OuterVolumeSpecName: "config-data") pod "0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924" (UID: "0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:56:11 crc kubenswrapper[4997]: I1205 08:56:11.812705 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:56:11 crc kubenswrapper[4997]: I1205 08:56:11.813427 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:56:11 crc kubenswrapper[4997]: I1205 08:56:11.813480 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bd6h5\" (UniqueName: \"kubernetes.io/projected/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-kube-api-access-bd6h5\") on node \"crc\" DevicePath \"\"" Dec 05 08:56:11 crc kubenswrapper[4997]: I1205 08:56:11.813513 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:56:11 crc kubenswrapper[4997]: I1205 08:56:11.813527 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.286838 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7959t" event={"ID":"0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924","Type":"ContainerDied","Data":"3abd5b18767da80299ba0a74ba33544364598c5361a934eb5d7d33693953093a"} Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.287061 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3abd5b18767da80299ba0a74ba33544364598c5361a934eb5d7d33693953093a" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.286890 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7959t" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.382881 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-565c9dc47b-g8s9d"] Dec 05 08:56:12 crc kubenswrapper[4997]: E1205 08:56:12.383961 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924" containerName="placement-db-sync" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.385331 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924" containerName="placement-db-sync" Dec 05 08:56:12 crc kubenswrapper[4997]: E1205 08:56:12.385704 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131a9488-9bd7-4f21-8b0b-b11fc8e46691" containerName="init" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.385790 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="131a9488-9bd7-4f21-8b0b-b11fc8e46691" containerName="init" Dec 05 08:56:12 crc kubenswrapper[4997]: E1205 08:56:12.385890 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131a9488-9bd7-4f21-8b0b-b11fc8e46691" containerName="dnsmasq-dns" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.385960 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="131a9488-9bd7-4f21-8b0b-b11fc8e46691" containerName="dnsmasq-dns" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.386339 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="131a9488-9bd7-4f21-8b0b-b11fc8e46691" containerName="dnsmasq-dns" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.386434 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924" containerName="placement-db-sync" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.387862 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.390561 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-565c9dc47b-g8s9d"] Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.391773 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.391982 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-2m8xp" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.398545 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.523911 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns6h2\" (UniqueName: \"kubernetes.io/projected/bd651c2b-8f0e-4caf-967b-36b0ccdbe54a-kube-api-access-ns6h2\") pod \"placement-565c9dc47b-g8s9d\" (UID: \"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a\") " pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.523963 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd651c2b-8f0e-4caf-967b-36b0ccdbe54a-combined-ca-bundle\") pod \"placement-565c9dc47b-g8s9d\" (UID: \"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a\") " pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.524012 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd651c2b-8f0e-4caf-967b-36b0ccdbe54a-config-data\") pod \"placement-565c9dc47b-g8s9d\" (UID: \"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a\") " pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.524083 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd651c2b-8f0e-4caf-967b-36b0ccdbe54a-logs\") pod \"placement-565c9dc47b-g8s9d\" (UID: \"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a\") " pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.524142 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd651c2b-8f0e-4caf-967b-36b0ccdbe54a-scripts\") pod \"placement-565c9dc47b-g8s9d\" (UID: \"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a\") " pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.631777 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns6h2\" (UniqueName: \"kubernetes.io/projected/bd651c2b-8f0e-4caf-967b-36b0ccdbe54a-kube-api-access-ns6h2\") pod \"placement-565c9dc47b-g8s9d\" (UID: \"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a\") " pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.631844 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd651c2b-8f0e-4caf-967b-36b0ccdbe54a-combined-ca-bundle\") pod \"placement-565c9dc47b-g8s9d\" (UID: \"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a\") " pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.631911 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd651c2b-8f0e-4caf-967b-36b0ccdbe54a-config-data\") pod \"placement-565c9dc47b-g8s9d\" (UID: \"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a\") " pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.631963 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd651c2b-8f0e-4caf-967b-36b0ccdbe54a-logs\") pod \"placement-565c9dc47b-g8s9d\" (UID: \"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a\") " pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.632046 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd651c2b-8f0e-4caf-967b-36b0ccdbe54a-scripts\") pod \"placement-565c9dc47b-g8s9d\" (UID: \"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a\") " pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.632911 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd651c2b-8f0e-4caf-967b-36b0ccdbe54a-logs\") pod \"placement-565c9dc47b-g8s9d\" (UID: \"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a\") " pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.646530 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd651c2b-8f0e-4caf-967b-36b0ccdbe54a-config-data\") pod \"placement-565c9dc47b-g8s9d\" (UID: \"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a\") " pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.657111 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd651c2b-8f0e-4caf-967b-36b0ccdbe54a-scripts\") pod \"placement-565c9dc47b-g8s9d\" (UID: \"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a\") " pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.663215 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd651c2b-8f0e-4caf-967b-36b0ccdbe54a-combined-ca-bundle\") pod \"placement-565c9dc47b-g8s9d\" (UID: \"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a\") " pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.665633 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns6h2\" (UniqueName: \"kubernetes.io/projected/bd651c2b-8f0e-4caf-967b-36b0ccdbe54a-kube-api-access-ns6h2\") pod \"placement-565c9dc47b-g8s9d\" (UID: \"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a\") " pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:12 crc kubenswrapper[4997]: I1205 08:56:12.716046 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:13 crc kubenswrapper[4997]: I1205 08:56:13.199971 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-565c9dc47b-g8s9d"] Dec 05 08:56:13 crc kubenswrapper[4997]: W1205 08:56:13.203647 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd651c2b_8f0e_4caf_967b_36b0ccdbe54a.slice/crio-967c3218fc1c2d42c6c67504e75467aa67e80b3d93df81c025a68217f1af6f61 WatchSource:0}: Error finding container 967c3218fc1c2d42c6c67504e75467aa67e80b3d93df81c025a68217f1af6f61: Status 404 returned error can't find the container with id 967c3218fc1c2d42c6c67504e75467aa67e80b3d93df81c025a68217f1af6f61 Dec 05 08:56:13 crc kubenswrapper[4997]: I1205 08:56:13.295281 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-565c9dc47b-g8s9d" event={"ID":"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a","Type":"ContainerStarted","Data":"967c3218fc1c2d42c6c67504e75467aa67e80b3d93df81c025a68217f1af6f61"} Dec 05 08:56:14 crc kubenswrapper[4997]: I1205 08:56:14.305694 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-565c9dc47b-g8s9d" event={"ID":"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a","Type":"ContainerStarted","Data":"0440f8ebbdac8457d8f1a0972a79005842e2e81acc53872bea46ed010a9e6397"} Dec 05 08:56:14 crc kubenswrapper[4997]: I1205 08:56:14.306080 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-565c9dc47b-g8s9d" event={"ID":"bd651c2b-8f0e-4caf-967b-36b0ccdbe54a","Type":"ContainerStarted","Data":"a255c382c15b1e5e08f2714a57822729e0108887b826e291cf70598da17728c1"} Dec 05 08:56:14 crc kubenswrapper[4997]: I1205 08:56:14.308052 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:14 crc kubenswrapper[4997]: I1205 08:56:14.308104 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:14 crc kubenswrapper[4997]: I1205 08:56:14.333635 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-565c9dc47b-g8s9d" podStartSLOduration=2.333601645 podStartE2EDuration="2.333601645s" podCreationTimestamp="2025-12-05 08:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:56:14.332048413 +0000 UTC m=+7274.860955694" watchObservedRunningTime="2025-12-05 08:56:14.333601645 +0000 UTC m=+7274.862508916" Dec 05 08:56:15 crc kubenswrapper[4997]: I1205 08:56:15.386506 4997 scope.go:117] "RemoveContainer" containerID="83bd574bee13eea7f1a9ff7aa3d514e06ed543014dee9ba67759d04ac052ee59" Dec 05 08:56:43 crc kubenswrapper[4997]: I1205 08:56:43.777158 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:56:43 crc kubenswrapper[4997]: I1205 08:56:43.777725 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-565c9dc47b-g8s9d" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.094024 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-mkz6j"] Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.095916 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mkz6j" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.106417 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-mkz6j"] Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.194965 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-w6nhg"] Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.196380 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-w6nhg" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.204361 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-w6nhg"] Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.244736 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3987fcc4-0e92-4006-a1f5-ad2c937f057e-operator-scripts\") pod \"nova-api-db-create-mkz6j\" (UID: \"3987fcc4-0e92-4006-a1f5-ad2c937f057e\") " pod="openstack/nova-api-db-create-mkz6j" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.244836 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swqxq\" (UniqueName: \"kubernetes.io/projected/3987fcc4-0e92-4006-a1f5-ad2c937f057e-kube-api-access-swqxq\") pod \"nova-api-db-create-mkz6j\" (UID: \"3987fcc4-0e92-4006-a1f5-ad2c937f057e\") " pod="openstack/nova-api-db-create-mkz6j" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.302688 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-5tctv"] Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.304071 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5tctv" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.311228 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-750c-account-create-update-jd4sj"] Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.312514 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-750c-account-create-update-jd4sj" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.314884 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.328332 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5tctv"] Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.341691 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-750c-account-create-update-jd4sj"] Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.347268 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3987fcc4-0e92-4006-a1f5-ad2c937f057e-operator-scripts\") pod \"nova-api-db-create-mkz6j\" (UID: \"3987fcc4-0e92-4006-a1f5-ad2c937f057e\") " pod="openstack/nova-api-db-create-mkz6j" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.347334 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ec1f8f1-691f-4750-ad3b-8da18a706133-operator-scripts\") pod \"nova-cell0-db-create-w6nhg\" (UID: \"3ec1f8f1-691f-4750-ad3b-8da18a706133\") " pod="openstack/nova-cell0-db-create-w6nhg" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.347364 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swqxq\" (UniqueName: \"kubernetes.io/projected/3987fcc4-0e92-4006-a1f5-ad2c937f057e-kube-api-access-swqxq\") pod \"nova-api-db-create-mkz6j\" (UID: \"3987fcc4-0e92-4006-a1f5-ad2c937f057e\") " pod="openstack/nova-api-db-create-mkz6j" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.347431 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cflnh\" (UniqueName: \"kubernetes.io/projected/3ec1f8f1-691f-4750-ad3b-8da18a706133-kube-api-access-cflnh\") pod \"nova-cell0-db-create-w6nhg\" (UID: \"3ec1f8f1-691f-4750-ad3b-8da18a706133\") " pod="openstack/nova-cell0-db-create-w6nhg" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.347738 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3987fcc4-0e92-4006-a1f5-ad2c937f057e-operator-scripts\") pod \"nova-api-db-create-mkz6j\" (UID: \"3987fcc4-0e92-4006-a1f5-ad2c937f057e\") " pod="openstack/nova-api-db-create-mkz6j" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.377736 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swqxq\" (UniqueName: \"kubernetes.io/projected/3987fcc4-0e92-4006-a1f5-ad2c937f057e-kube-api-access-swqxq\") pod \"nova-api-db-create-mkz6j\" (UID: \"3987fcc4-0e92-4006-a1f5-ad2c937f057e\") " pod="openstack/nova-api-db-create-mkz6j" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.415157 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mkz6j" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.449145 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wszmr\" (UniqueName: \"kubernetes.io/projected/c3ea1a78-5866-44a2-b8ce-5203e199f4f9-kube-api-access-wszmr\") pod \"nova-cell1-db-create-5tctv\" (UID: \"c3ea1a78-5866-44a2-b8ce-5203e199f4f9\") " pod="openstack/nova-cell1-db-create-5tctv" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.449197 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cflnh\" (UniqueName: \"kubernetes.io/projected/3ec1f8f1-691f-4750-ad3b-8da18a706133-kube-api-access-cflnh\") pod \"nova-cell0-db-create-w6nhg\" (UID: \"3ec1f8f1-691f-4750-ad3b-8da18a706133\") " pod="openstack/nova-cell0-db-create-w6nhg" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.449473 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rrmz\" (UniqueName: \"kubernetes.io/projected/ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f-kube-api-access-9rrmz\") pod \"nova-api-750c-account-create-update-jd4sj\" (UID: \"ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f\") " pod="openstack/nova-api-750c-account-create-update-jd4sj" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.449524 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ec1f8f1-691f-4750-ad3b-8da18a706133-operator-scripts\") pod \"nova-cell0-db-create-w6nhg\" (UID: \"3ec1f8f1-691f-4750-ad3b-8da18a706133\") " pod="openstack/nova-cell0-db-create-w6nhg" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.449546 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f-operator-scripts\") pod \"nova-api-750c-account-create-update-jd4sj\" (UID: \"ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f\") " pod="openstack/nova-api-750c-account-create-update-jd4sj" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.449665 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3ea1a78-5866-44a2-b8ce-5203e199f4f9-operator-scripts\") pod \"nova-cell1-db-create-5tctv\" (UID: \"c3ea1a78-5866-44a2-b8ce-5203e199f4f9\") " pod="openstack/nova-cell1-db-create-5tctv" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.450399 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ec1f8f1-691f-4750-ad3b-8da18a706133-operator-scripts\") pod \"nova-cell0-db-create-w6nhg\" (UID: \"3ec1f8f1-691f-4750-ad3b-8da18a706133\") " pod="openstack/nova-cell0-db-create-w6nhg" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.473220 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cflnh\" (UniqueName: \"kubernetes.io/projected/3ec1f8f1-691f-4750-ad3b-8da18a706133-kube-api-access-cflnh\") pod \"nova-cell0-db-create-w6nhg\" (UID: \"3ec1f8f1-691f-4750-ad3b-8da18a706133\") " pod="openstack/nova-cell0-db-create-w6nhg" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.501554 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-fbe8-account-create-update-p8p6z"] Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.502777 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fbe8-account-create-update-p8p6z" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.507061 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.513518 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-w6nhg" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.527655 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-fbe8-account-create-update-p8p6z"] Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.551518 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rrmz\" (UniqueName: \"kubernetes.io/projected/ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f-kube-api-access-9rrmz\") pod \"nova-api-750c-account-create-update-jd4sj\" (UID: \"ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f\") " pod="openstack/nova-api-750c-account-create-update-jd4sj" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.551559 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f-operator-scripts\") pod \"nova-api-750c-account-create-update-jd4sj\" (UID: \"ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f\") " pod="openstack/nova-api-750c-account-create-update-jd4sj" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.551598 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3ea1a78-5866-44a2-b8ce-5203e199f4f9-operator-scripts\") pod \"nova-cell1-db-create-5tctv\" (UID: \"c3ea1a78-5866-44a2-b8ce-5203e199f4f9\") " pod="openstack/nova-cell1-db-create-5tctv" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.551666 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wszmr\" (UniqueName: \"kubernetes.io/projected/c3ea1a78-5866-44a2-b8ce-5203e199f4f9-kube-api-access-wszmr\") pod \"nova-cell1-db-create-5tctv\" (UID: \"c3ea1a78-5866-44a2-b8ce-5203e199f4f9\") " pod="openstack/nova-cell1-db-create-5tctv" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.552778 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3ea1a78-5866-44a2-b8ce-5203e199f4f9-operator-scripts\") pod \"nova-cell1-db-create-5tctv\" (UID: \"c3ea1a78-5866-44a2-b8ce-5203e199f4f9\") " pod="openstack/nova-cell1-db-create-5tctv" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.553059 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f-operator-scripts\") pod \"nova-api-750c-account-create-update-jd4sj\" (UID: \"ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f\") " pod="openstack/nova-api-750c-account-create-update-jd4sj" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.575468 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wszmr\" (UniqueName: \"kubernetes.io/projected/c3ea1a78-5866-44a2-b8ce-5203e199f4f9-kube-api-access-wszmr\") pod \"nova-cell1-db-create-5tctv\" (UID: \"c3ea1a78-5866-44a2-b8ce-5203e199f4f9\") " pod="openstack/nova-cell1-db-create-5tctv" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.575036 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rrmz\" (UniqueName: \"kubernetes.io/projected/ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f-kube-api-access-9rrmz\") pod \"nova-api-750c-account-create-update-jd4sj\" (UID: \"ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f\") " pod="openstack/nova-api-750c-account-create-update-jd4sj" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.631324 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5tctv" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.650023 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-750c-account-create-update-jd4sj" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.653257 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlzhv\" (UniqueName: \"kubernetes.io/projected/e23dfad8-111d-415a-bff0-2b8796cc951d-kube-api-access-nlzhv\") pod \"nova-cell0-fbe8-account-create-update-p8p6z\" (UID: \"e23dfad8-111d-415a-bff0-2b8796cc951d\") " pod="openstack/nova-cell0-fbe8-account-create-update-p8p6z" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.653390 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e23dfad8-111d-415a-bff0-2b8796cc951d-operator-scripts\") pod \"nova-cell0-fbe8-account-create-update-p8p6z\" (UID: \"e23dfad8-111d-415a-bff0-2b8796cc951d\") " pod="openstack/nova-cell0-fbe8-account-create-update-p8p6z" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.712534 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-a76d-account-create-update-rn7tm"] Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.713975 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a76d-account-create-update-rn7tm" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.717129 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.722955 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-a76d-account-create-update-rn7tm"] Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.754914 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlzhv\" (UniqueName: \"kubernetes.io/projected/e23dfad8-111d-415a-bff0-2b8796cc951d-kube-api-access-nlzhv\") pod \"nova-cell0-fbe8-account-create-update-p8p6z\" (UID: \"e23dfad8-111d-415a-bff0-2b8796cc951d\") " pod="openstack/nova-cell0-fbe8-account-create-update-p8p6z" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.755028 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e23dfad8-111d-415a-bff0-2b8796cc951d-operator-scripts\") pod \"nova-cell0-fbe8-account-create-update-p8p6z\" (UID: \"e23dfad8-111d-415a-bff0-2b8796cc951d\") " pod="openstack/nova-cell0-fbe8-account-create-update-p8p6z" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.755921 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e23dfad8-111d-415a-bff0-2b8796cc951d-operator-scripts\") pod \"nova-cell0-fbe8-account-create-update-p8p6z\" (UID: \"e23dfad8-111d-415a-bff0-2b8796cc951d\") " pod="openstack/nova-cell0-fbe8-account-create-update-p8p6z" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.783984 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlzhv\" (UniqueName: \"kubernetes.io/projected/e23dfad8-111d-415a-bff0-2b8796cc951d-kube-api-access-nlzhv\") pod \"nova-cell0-fbe8-account-create-update-p8p6z\" (UID: \"e23dfad8-111d-415a-bff0-2b8796cc951d\") " pod="openstack/nova-cell0-fbe8-account-create-update-p8p6z" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.814675 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-w6nhg"] Dec 05 08:57:07 crc kubenswrapper[4997]: W1205 08:57:07.823716 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ec1f8f1_691f_4750_ad3b_8da18a706133.slice/crio-1fbaf0aadba3fa7c6f6214efcde571332a7ed28784045c884191658e174c3565 WatchSource:0}: Error finding container 1fbaf0aadba3fa7c6f6214efcde571332a7ed28784045c884191658e174c3565: Status 404 returned error can't find the container with id 1fbaf0aadba3fa7c6f6214efcde571332a7ed28784045c884191658e174c3565 Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.856092 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fbe8-account-create-update-p8p6z" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.856755 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0fb2077-9180-483e-94ce-c73bee41dd1b-operator-scripts\") pod \"nova-cell1-a76d-account-create-update-rn7tm\" (UID: \"d0fb2077-9180-483e-94ce-c73bee41dd1b\") " pod="openstack/nova-cell1-a76d-account-create-update-rn7tm" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.856832 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd6x7\" (UniqueName: \"kubernetes.io/projected/d0fb2077-9180-483e-94ce-c73bee41dd1b-kube-api-access-pd6x7\") pod \"nova-cell1-a76d-account-create-update-rn7tm\" (UID: \"d0fb2077-9180-483e-94ce-c73bee41dd1b\") " pod="openstack/nova-cell1-a76d-account-create-update-rn7tm" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.906135 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-mkz6j"] Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.940370 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5tctv"] Dec 05 08:57:07 crc kubenswrapper[4997]: W1205 08:57:07.955348 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3ea1a78_5866_44a2_b8ce_5203e199f4f9.slice/crio-4a3f9df2626cff8780c8ba94292cbff5834fcb9aa03b06ef6fbc884e8286c61f WatchSource:0}: Error finding container 4a3f9df2626cff8780c8ba94292cbff5834fcb9aa03b06ef6fbc884e8286c61f: Status 404 returned error can't find the container with id 4a3f9df2626cff8780c8ba94292cbff5834fcb9aa03b06ef6fbc884e8286c61f Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.962105 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0fb2077-9180-483e-94ce-c73bee41dd1b-operator-scripts\") pod \"nova-cell1-a76d-account-create-update-rn7tm\" (UID: \"d0fb2077-9180-483e-94ce-c73bee41dd1b\") " pod="openstack/nova-cell1-a76d-account-create-update-rn7tm" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.962188 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd6x7\" (UniqueName: \"kubernetes.io/projected/d0fb2077-9180-483e-94ce-c73bee41dd1b-kube-api-access-pd6x7\") pod \"nova-cell1-a76d-account-create-update-rn7tm\" (UID: \"d0fb2077-9180-483e-94ce-c73bee41dd1b\") " pod="openstack/nova-cell1-a76d-account-create-update-rn7tm" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.962917 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0fb2077-9180-483e-94ce-c73bee41dd1b-operator-scripts\") pod \"nova-cell1-a76d-account-create-update-rn7tm\" (UID: \"d0fb2077-9180-483e-94ce-c73bee41dd1b\") " pod="openstack/nova-cell1-a76d-account-create-update-rn7tm" Dec 05 08:57:07 crc kubenswrapper[4997]: I1205 08:57:07.986913 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd6x7\" (UniqueName: \"kubernetes.io/projected/d0fb2077-9180-483e-94ce-c73bee41dd1b-kube-api-access-pd6x7\") pod \"nova-cell1-a76d-account-create-update-rn7tm\" (UID: \"d0fb2077-9180-483e-94ce-c73bee41dd1b\") " pod="openstack/nova-cell1-a76d-account-create-update-rn7tm" Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.013693 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-750c-account-create-update-jd4sj"] Dec 05 08:57:08 crc kubenswrapper[4997]: W1205 08:57:08.014041 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccfb9b2b_0899_4f23_a7a2_601fd5a6a66f.slice/crio-dcde11ccfa9a1f34196c1a2313d93449653d51c3ecff2216f0525d9b0a3fcc34 WatchSource:0}: Error finding container dcde11ccfa9a1f34196c1a2313d93449653d51c3ecff2216f0525d9b0a3fcc34: Status 404 returned error can't find the container with id dcde11ccfa9a1f34196c1a2313d93449653d51c3ecff2216f0525d9b0a3fcc34 Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.049998 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a76d-account-create-update-rn7tm" Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.312909 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-fbe8-account-create-update-p8p6z"] Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.543303 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-a76d-account-create-update-rn7tm"] Dec 05 08:57:08 crc kubenswrapper[4997]: W1205 08:57:08.579503 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0fb2077_9180_483e_94ce_c73bee41dd1b.slice/crio-a8c1d23743f2a9cdaaf1c674dabb2f64281b84578779f828570070307926a310 WatchSource:0}: Error finding container a8c1d23743f2a9cdaaf1c674dabb2f64281b84578779f828570070307926a310: Status 404 returned error can't find the container with id a8c1d23743f2a9cdaaf1c674dabb2f64281b84578779f828570070307926a310 Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.806774 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a76d-account-create-update-rn7tm" event={"ID":"d0fb2077-9180-483e-94ce-c73bee41dd1b","Type":"ContainerStarted","Data":"afbdaf466a44206a95ee91f0f74bdc6cb1a2fcff639c2147026e6faf413d2157"} Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.806817 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a76d-account-create-update-rn7tm" event={"ID":"d0fb2077-9180-483e-94ce-c73bee41dd1b","Type":"ContainerStarted","Data":"a8c1d23743f2a9cdaaf1c674dabb2f64281b84578779f828570070307926a310"} Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.809293 4997 generic.go:334] "Generic (PLEG): container finished" podID="c3ea1a78-5866-44a2-b8ce-5203e199f4f9" containerID="a782bd326b4eb5a7bc1dcf43d39a57cb83d62c4c33ec62697bc83ad8d3e9fe23" exitCode=0 Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.809369 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5tctv" event={"ID":"c3ea1a78-5866-44a2-b8ce-5203e199f4f9","Type":"ContainerDied","Data":"a782bd326b4eb5a7bc1dcf43d39a57cb83d62c4c33ec62697bc83ad8d3e9fe23"} Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.809393 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5tctv" event={"ID":"c3ea1a78-5866-44a2-b8ce-5203e199f4f9","Type":"ContainerStarted","Data":"4a3f9df2626cff8780c8ba94292cbff5834fcb9aa03b06ef6fbc884e8286c61f"} Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.811330 4997 generic.go:334] "Generic (PLEG): container finished" podID="ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f" containerID="b25f9149e65b3ad1ff2e3498fea6cd8ea1e00287214d430556e09c7021437fa4" exitCode=0 Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.811383 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-750c-account-create-update-jd4sj" event={"ID":"ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f","Type":"ContainerDied","Data":"b25f9149e65b3ad1ff2e3498fea6cd8ea1e00287214d430556e09c7021437fa4"} Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.811403 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-750c-account-create-update-jd4sj" event={"ID":"ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f","Type":"ContainerStarted","Data":"dcde11ccfa9a1f34196c1a2313d93449653d51c3ecff2216f0525d9b0a3fcc34"} Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.814509 4997 generic.go:334] "Generic (PLEG): container finished" podID="3ec1f8f1-691f-4750-ad3b-8da18a706133" containerID="5d80064eba66e65f9b836496de33afded37525a73aa1631ca6a871a1fc465fd2" exitCode=0 Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.814562 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-w6nhg" event={"ID":"3ec1f8f1-691f-4750-ad3b-8da18a706133","Type":"ContainerDied","Data":"5d80064eba66e65f9b836496de33afded37525a73aa1631ca6a871a1fc465fd2"} Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.814584 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-w6nhg" event={"ID":"3ec1f8f1-691f-4750-ad3b-8da18a706133","Type":"ContainerStarted","Data":"1fbaf0aadba3fa7c6f6214efcde571332a7ed28784045c884191658e174c3565"} Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.816717 4997 generic.go:334] "Generic (PLEG): container finished" podID="3987fcc4-0e92-4006-a1f5-ad2c937f057e" containerID="7832ca93ad1b7e818faff5771a9058448de46d6009bdc24638f6f51ca37d74a1" exitCode=0 Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.816783 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mkz6j" event={"ID":"3987fcc4-0e92-4006-a1f5-ad2c937f057e","Type":"ContainerDied","Data":"7832ca93ad1b7e818faff5771a9058448de46d6009bdc24638f6f51ca37d74a1"} Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.816808 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mkz6j" event={"ID":"3987fcc4-0e92-4006-a1f5-ad2c937f057e","Type":"ContainerStarted","Data":"76cc18e607f5b90a1bd3cce0d518cb50562c4dd06d0737c3024dd0855488ecb6"} Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.818509 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-fbe8-account-create-update-p8p6z" event={"ID":"e23dfad8-111d-415a-bff0-2b8796cc951d","Type":"ContainerStarted","Data":"3ee65aa68794c7532078e07e4859042a969598949ca778893474cbf07e1236e9"} Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.818540 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-fbe8-account-create-update-p8p6z" event={"ID":"e23dfad8-111d-415a-bff0-2b8796cc951d","Type":"ContainerStarted","Data":"443adfebeb27a11b7ebe8ed9a7c8268781c6aa10c35a6517845ab2e72affd559"} Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.827462 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-a76d-account-create-update-rn7tm" podStartSLOduration=1.827432827 podStartE2EDuration="1.827432827s" podCreationTimestamp="2025-12-05 08:57:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:57:08.823954324 +0000 UTC m=+7329.352861595" watchObservedRunningTime="2025-12-05 08:57:08.827432827 +0000 UTC m=+7329.356340088" Dec 05 08:57:08 crc kubenswrapper[4997]: I1205 08:57:08.888575 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-fbe8-account-create-update-p8p6z" podStartSLOduration=1.88855695 podStartE2EDuration="1.88855695s" podCreationTimestamp="2025-12-05 08:57:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:57:08.879288561 +0000 UTC m=+7329.408195842" watchObservedRunningTime="2025-12-05 08:57:08.88855695 +0000 UTC m=+7329.417464211" Dec 05 08:57:09 crc kubenswrapper[4997]: I1205 08:57:09.848134 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-fbe8-account-create-update-p8p6z" event={"ID":"e23dfad8-111d-415a-bff0-2b8796cc951d","Type":"ContainerDied","Data":"3ee65aa68794c7532078e07e4859042a969598949ca778893474cbf07e1236e9"} Dec 05 08:57:09 crc kubenswrapper[4997]: I1205 08:57:09.848006 4997 generic.go:334] "Generic (PLEG): container finished" podID="e23dfad8-111d-415a-bff0-2b8796cc951d" containerID="3ee65aa68794c7532078e07e4859042a969598949ca778893474cbf07e1236e9" exitCode=0 Dec 05 08:57:09 crc kubenswrapper[4997]: I1205 08:57:09.853303 4997 generic.go:334] "Generic (PLEG): container finished" podID="d0fb2077-9180-483e-94ce-c73bee41dd1b" containerID="afbdaf466a44206a95ee91f0f74bdc6cb1a2fcff639c2147026e6faf413d2157" exitCode=0 Dec 05 08:57:09 crc kubenswrapper[4997]: I1205 08:57:09.853396 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a76d-account-create-update-rn7tm" event={"ID":"d0fb2077-9180-483e-94ce-c73bee41dd1b","Type":"ContainerDied","Data":"afbdaf466a44206a95ee91f0f74bdc6cb1a2fcff639c2147026e6faf413d2157"} Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.242216 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-750c-account-create-update-jd4sj" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.324698 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rrmz\" (UniqueName: \"kubernetes.io/projected/ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f-kube-api-access-9rrmz\") pod \"ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f\" (UID: \"ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f\") " Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.324807 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f-operator-scripts\") pod \"ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f\" (UID: \"ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f\") " Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.326224 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f" (UID: "ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.333102 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f-kube-api-access-9rrmz" (OuterVolumeSpecName: "kube-api-access-9rrmz") pod "ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f" (UID: "ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f"). InnerVolumeSpecName "kube-api-access-9rrmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.382421 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5tctv" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.393148 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mkz6j" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.398154 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-w6nhg" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.427407 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.427866 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rrmz\" (UniqueName: \"kubernetes.io/projected/ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f-kube-api-access-9rrmz\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.529259 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3ea1a78-5866-44a2-b8ce-5203e199f4f9-operator-scripts\") pod \"c3ea1a78-5866-44a2-b8ce-5203e199f4f9\" (UID: \"c3ea1a78-5866-44a2-b8ce-5203e199f4f9\") " Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.529334 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wszmr\" (UniqueName: \"kubernetes.io/projected/c3ea1a78-5866-44a2-b8ce-5203e199f4f9-kube-api-access-wszmr\") pod \"c3ea1a78-5866-44a2-b8ce-5203e199f4f9\" (UID: \"c3ea1a78-5866-44a2-b8ce-5203e199f4f9\") " Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.529420 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3987fcc4-0e92-4006-a1f5-ad2c937f057e-operator-scripts\") pod \"3987fcc4-0e92-4006-a1f5-ad2c937f057e\" (UID: \"3987fcc4-0e92-4006-a1f5-ad2c937f057e\") " Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.529523 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cflnh\" (UniqueName: \"kubernetes.io/projected/3ec1f8f1-691f-4750-ad3b-8da18a706133-kube-api-access-cflnh\") pod \"3ec1f8f1-691f-4750-ad3b-8da18a706133\" (UID: \"3ec1f8f1-691f-4750-ad3b-8da18a706133\") " Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.529559 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ec1f8f1-691f-4750-ad3b-8da18a706133-operator-scripts\") pod \"3ec1f8f1-691f-4750-ad3b-8da18a706133\" (UID: \"3ec1f8f1-691f-4750-ad3b-8da18a706133\") " Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.529663 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swqxq\" (UniqueName: \"kubernetes.io/projected/3987fcc4-0e92-4006-a1f5-ad2c937f057e-kube-api-access-swqxq\") pod \"3987fcc4-0e92-4006-a1f5-ad2c937f057e\" (UID: \"3987fcc4-0e92-4006-a1f5-ad2c937f057e\") " Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.530368 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ec1f8f1-691f-4750-ad3b-8da18a706133-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3ec1f8f1-691f-4750-ad3b-8da18a706133" (UID: "3ec1f8f1-691f-4750-ad3b-8da18a706133"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.530884 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3987fcc4-0e92-4006-a1f5-ad2c937f057e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3987fcc4-0e92-4006-a1f5-ad2c937f057e" (UID: "3987fcc4-0e92-4006-a1f5-ad2c937f057e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.532289 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3ea1a78-5866-44a2-b8ce-5203e199f4f9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c3ea1a78-5866-44a2-b8ce-5203e199f4f9" (UID: "c3ea1a78-5866-44a2-b8ce-5203e199f4f9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.534190 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3ea1a78-5866-44a2-b8ce-5203e199f4f9-kube-api-access-wszmr" (OuterVolumeSpecName: "kube-api-access-wszmr") pod "c3ea1a78-5866-44a2-b8ce-5203e199f4f9" (UID: "c3ea1a78-5866-44a2-b8ce-5203e199f4f9"). InnerVolumeSpecName "kube-api-access-wszmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.534720 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ec1f8f1-691f-4750-ad3b-8da18a706133-kube-api-access-cflnh" (OuterVolumeSpecName: "kube-api-access-cflnh") pod "3ec1f8f1-691f-4750-ad3b-8da18a706133" (UID: "3ec1f8f1-691f-4750-ad3b-8da18a706133"). InnerVolumeSpecName "kube-api-access-cflnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.534789 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3987fcc4-0e92-4006-a1f5-ad2c937f057e-kube-api-access-swqxq" (OuterVolumeSpecName: "kube-api-access-swqxq") pod "3987fcc4-0e92-4006-a1f5-ad2c937f057e" (UID: "3987fcc4-0e92-4006-a1f5-ad2c937f057e"). InnerVolumeSpecName "kube-api-access-swqxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.631573 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swqxq\" (UniqueName: \"kubernetes.io/projected/3987fcc4-0e92-4006-a1f5-ad2c937f057e-kube-api-access-swqxq\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.631646 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3ea1a78-5866-44a2-b8ce-5203e199f4f9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.631662 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wszmr\" (UniqueName: \"kubernetes.io/projected/c3ea1a78-5866-44a2-b8ce-5203e199f4f9-kube-api-access-wszmr\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.631671 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3987fcc4-0e92-4006-a1f5-ad2c937f057e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.631684 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cflnh\" (UniqueName: \"kubernetes.io/projected/3ec1f8f1-691f-4750-ad3b-8da18a706133-kube-api-access-cflnh\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.631693 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ec1f8f1-691f-4750-ad3b-8da18a706133-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.862998 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5tctv" event={"ID":"c3ea1a78-5866-44a2-b8ce-5203e199f4f9","Type":"ContainerDied","Data":"4a3f9df2626cff8780c8ba94292cbff5834fcb9aa03b06ef6fbc884e8286c61f"} Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.863383 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a3f9df2626cff8780c8ba94292cbff5834fcb9aa03b06ef6fbc884e8286c61f" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.863046 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5tctv" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.865337 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-750c-account-create-update-jd4sj" event={"ID":"ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f","Type":"ContainerDied","Data":"dcde11ccfa9a1f34196c1a2313d93449653d51c3ecff2216f0525d9b0a3fcc34"} Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.865373 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-750c-account-create-update-jd4sj" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.865389 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcde11ccfa9a1f34196c1a2313d93449653d51c3ecff2216f0525d9b0a3fcc34" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.867216 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-w6nhg" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.867246 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-w6nhg" event={"ID":"3ec1f8f1-691f-4750-ad3b-8da18a706133","Type":"ContainerDied","Data":"1fbaf0aadba3fa7c6f6214efcde571332a7ed28784045c884191658e174c3565"} Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.867285 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fbaf0aadba3fa7c6f6214efcde571332a7ed28784045c884191658e174c3565" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.869075 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mkz6j" Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.869084 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mkz6j" event={"ID":"3987fcc4-0e92-4006-a1f5-ad2c937f057e","Type":"ContainerDied","Data":"76cc18e607f5b90a1bd3cce0d518cb50562c4dd06d0737c3024dd0855488ecb6"} Dec 05 08:57:10 crc kubenswrapper[4997]: I1205 08:57:10.869117 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76cc18e607f5b90a1bd3cce0d518cb50562c4dd06d0737c3024dd0855488ecb6" Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.250128 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fbe8-account-create-update-p8p6z" Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.257024 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a76d-account-create-update-rn7tm" Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.344186 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlzhv\" (UniqueName: \"kubernetes.io/projected/e23dfad8-111d-415a-bff0-2b8796cc951d-kube-api-access-nlzhv\") pod \"e23dfad8-111d-415a-bff0-2b8796cc951d\" (UID: \"e23dfad8-111d-415a-bff0-2b8796cc951d\") " Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.344294 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e23dfad8-111d-415a-bff0-2b8796cc951d-operator-scripts\") pod \"e23dfad8-111d-415a-bff0-2b8796cc951d\" (UID: \"e23dfad8-111d-415a-bff0-2b8796cc951d\") " Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.344428 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0fb2077-9180-483e-94ce-c73bee41dd1b-operator-scripts\") pod \"d0fb2077-9180-483e-94ce-c73bee41dd1b\" (UID: \"d0fb2077-9180-483e-94ce-c73bee41dd1b\") " Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.344511 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pd6x7\" (UniqueName: \"kubernetes.io/projected/d0fb2077-9180-483e-94ce-c73bee41dd1b-kube-api-access-pd6x7\") pod \"d0fb2077-9180-483e-94ce-c73bee41dd1b\" (UID: \"d0fb2077-9180-483e-94ce-c73bee41dd1b\") " Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.344752 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0fb2077-9180-483e-94ce-c73bee41dd1b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d0fb2077-9180-483e-94ce-c73bee41dd1b" (UID: "d0fb2077-9180-483e-94ce-c73bee41dd1b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.344759 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e23dfad8-111d-415a-bff0-2b8796cc951d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e23dfad8-111d-415a-bff0-2b8796cc951d" (UID: "e23dfad8-111d-415a-bff0-2b8796cc951d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.344989 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0fb2077-9180-483e-94ce-c73bee41dd1b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.345024 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e23dfad8-111d-415a-bff0-2b8796cc951d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.346865 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e23dfad8-111d-415a-bff0-2b8796cc951d-kube-api-access-nlzhv" (OuterVolumeSpecName: "kube-api-access-nlzhv") pod "e23dfad8-111d-415a-bff0-2b8796cc951d" (UID: "e23dfad8-111d-415a-bff0-2b8796cc951d"). InnerVolumeSpecName "kube-api-access-nlzhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.347194 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0fb2077-9180-483e-94ce-c73bee41dd1b-kube-api-access-pd6x7" (OuterVolumeSpecName: "kube-api-access-pd6x7") pod "d0fb2077-9180-483e-94ce-c73bee41dd1b" (UID: "d0fb2077-9180-483e-94ce-c73bee41dd1b"). InnerVolumeSpecName "kube-api-access-pd6x7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.447320 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pd6x7\" (UniqueName: \"kubernetes.io/projected/d0fb2077-9180-483e-94ce-c73bee41dd1b-kube-api-access-pd6x7\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.447382 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlzhv\" (UniqueName: \"kubernetes.io/projected/e23dfad8-111d-415a-bff0-2b8796cc951d-kube-api-access-nlzhv\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.879761 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-fbe8-account-create-update-p8p6z" event={"ID":"e23dfad8-111d-415a-bff0-2b8796cc951d","Type":"ContainerDied","Data":"443adfebeb27a11b7ebe8ed9a7c8268781c6aa10c35a6517845ab2e72affd559"} Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.879788 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fbe8-account-create-update-p8p6z" Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.879803 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="443adfebeb27a11b7ebe8ed9a7c8268781c6aa10c35a6517845ab2e72affd559" Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.881113 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a76d-account-create-update-rn7tm" event={"ID":"d0fb2077-9180-483e-94ce-c73bee41dd1b","Type":"ContainerDied","Data":"a8c1d23743f2a9cdaaf1c674dabb2f64281b84578779f828570070307926a310"} Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.881240 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8c1d23743f2a9cdaaf1c674dabb2f64281b84578779f828570070307926a310" Dec 05 08:57:11 crc kubenswrapper[4997]: I1205 08:57:11.881146 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a76d-account-create-update-rn7tm" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.690175 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bpbv8"] Dec 05 08:57:12 crc kubenswrapper[4997]: E1205 08:57:12.690841 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f" containerName="mariadb-account-create-update" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.690858 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f" containerName="mariadb-account-create-update" Dec 05 08:57:12 crc kubenswrapper[4997]: E1205 08:57:12.690871 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0fb2077-9180-483e-94ce-c73bee41dd1b" containerName="mariadb-account-create-update" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.690878 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0fb2077-9180-483e-94ce-c73bee41dd1b" containerName="mariadb-account-create-update" Dec 05 08:57:12 crc kubenswrapper[4997]: E1205 08:57:12.690898 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3987fcc4-0e92-4006-a1f5-ad2c937f057e" containerName="mariadb-database-create" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.690904 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3987fcc4-0e92-4006-a1f5-ad2c937f057e" containerName="mariadb-database-create" Dec 05 08:57:12 crc kubenswrapper[4997]: E1205 08:57:12.690914 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3ea1a78-5866-44a2-b8ce-5203e199f4f9" containerName="mariadb-database-create" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.690920 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3ea1a78-5866-44a2-b8ce-5203e199f4f9" containerName="mariadb-database-create" Dec 05 08:57:12 crc kubenswrapper[4997]: E1205 08:57:12.690928 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ec1f8f1-691f-4750-ad3b-8da18a706133" containerName="mariadb-database-create" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.690934 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ec1f8f1-691f-4750-ad3b-8da18a706133" containerName="mariadb-database-create" Dec 05 08:57:12 crc kubenswrapper[4997]: E1205 08:57:12.690949 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e23dfad8-111d-415a-bff0-2b8796cc951d" containerName="mariadb-account-create-update" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.690955 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e23dfad8-111d-415a-bff0-2b8796cc951d" containerName="mariadb-account-create-update" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.691178 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f" containerName="mariadb-account-create-update" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.691198 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3ea1a78-5866-44a2-b8ce-5203e199f4f9" containerName="mariadb-database-create" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.691208 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="3987fcc4-0e92-4006-a1f5-ad2c937f057e" containerName="mariadb-database-create" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.691218 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e23dfad8-111d-415a-bff0-2b8796cc951d" containerName="mariadb-account-create-update" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.691231 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ec1f8f1-691f-4750-ad3b-8da18a706133" containerName="mariadb-database-create" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.691240 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0fb2077-9180-483e-94ce-c73bee41dd1b" containerName="mariadb-account-create-update" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.691828 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bpbv8" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.698127 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.698559 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2jlfs" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.698729 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.713077 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bpbv8"] Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.774812 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-config-data\") pod \"nova-cell0-conductor-db-sync-bpbv8\" (UID: \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\") " pod="openstack/nova-cell0-conductor-db-sync-bpbv8" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.774876 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdd4v\" (UniqueName: \"kubernetes.io/projected/66bea143-bb89-4a9c-82cb-d0cff893d8ad-kube-api-access-cdd4v\") pod \"nova-cell0-conductor-db-sync-bpbv8\" (UID: \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\") " pod="openstack/nova-cell0-conductor-db-sync-bpbv8" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.774917 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-scripts\") pod \"nova-cell0-conductor-db-sync-bpbv8\" (UID: \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\") " pod="openstack/nova-cell0-conductor-db-sync-bpbv8" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.774962 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bpbv8\" (UID: \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\") " pod="openstack/nova-cell0-conductor-db-sync-bpbv8" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.876158 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-config-data\") pod \"nova-cell0-conductor-db-sync-bpbv8\" (UID: \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\") " pod="openstack/nova-cell0-conductor-db-sync-bpbv8" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.876238 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdd4v\" (UniqueName: \"kubernetes.io/projected/66bea143-bb89-4a9c-82cb-d0cff893d8ad-kube-api-access-cdd4v\") pod \"nova-cell0-conductor-db-sync-bpbv8\" (UID: \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\") " pod="openstack/nova-cell0-conductor-db-sync-bpbv8" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.876291 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-scripts\") pod \"nova-cell0-conductor-db-sync-bpbv8\" (UID: \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\") " pod="openstack/nova-cell0-conductor-db-sync-bpbv8" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.876363 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bpbv8\" (UID: \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\") " pod="openstack/nova-cell0-conductor-db-sync-bpbv8" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.881421 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bpbv8\" (UID: \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\") " pod="openstack/nova-cell0-conductor-db-sync-bpbv8" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.881459 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-config-data\") pod \"nova-cell0-conductor-db-sync-bpbv8\" (UID: \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\") " pod="openstack/nova-cell0-conductor-db-sync-bpbv8" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.889210 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-scripts\") pod \"nova-cell0-conductor-db-sync-bpbv8\" (UID: \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\") " pod="openstack/nova-cell0-conductor-db-sync-bpbv8" Dec 05 08:57:12 crc kubenswrapper[4997]: I1205 08:57:12.898689 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdd4v\" (UniqueName: \"kubernetes.io/projected/66bea143-bb89-4a9c-82cb-d0cff893d8ad-kube-api-access-cdd4v\") pod \"nova-cell0-conductor-db-sync-bpbv8\" (UID: \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\") " pod="openstack/nova-cell0-conductor-db-sync-bpbv8" Dec 05 08:57:13 crc kubenswrapper[4997]: I1205 08:57:13.031953 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bpbv8" Dec 05 08:57:13 crc kubenswrapper[4997]: I1205 08:57:13.270233 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bpbv8"] Dec 05 08:57:13 crc kubenswrapper[4997]: W1205 08:57:13.273953 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66bea143_bb89_4a9c_82cb_d0cff893d8ad.slice/crio-0520b831857d86e565750f200495a55ee031d8e7bf2984014bbacc5663ba7481 WatchSource:0}: Error finding container 0520b831857d86e565750f200495a55ee031d8e7bf2984014bbacc5663ba7481: Status 404 returned error can't find the container with id 0520b831857d86e565750f200495a55ee031d8e7bf2984014bbacc5663ba7481 Dec 05 08:57:13 crc kubenswrapper[4997]: I1205 08:57:13.898655 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bpbv8" event={"ID":"66bea143-bb89-4a9c-82cb-d0cff893d8ad","Type":"ContainerStarted","Data":"0520b831857d86e565750f200495a55ee031d8e7bf2984014bbacc5663ba7481"} Dec 05 08:57:15 crc kubenswrapper[4997]: I1205 08:57:15.493307 4997 scope.go:117] "RemoveContainer" containerID="26e04d5978d8454a5e28dbfd786bc8d7c723835268518aac180c43f3aa839922" Dec 05 08:57:23 crc kubenswrapper[4997]: I1205 08:57:23.984738 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bpbv8" event={"ID":"66bea143-bb89-4a9c-82cb-d0cff893d8ad","Type":"ContainerStarted","Data":"b978475c065f58bcfa921adf4ea2d5e99a71fadae7eccc3992831e6da1066e04"} Dec 05 08:57:24 crc kubenswrapper[4997]: I1205 08:57:24.006560 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-bpbv8" podStartSLOduration=2.319072676 podStartE2EDuration="12.006540802s" podCreationTimestamp="2025-12-05 08:57:12 +0000 UTC" firstStartedPulling="2025-12-05 08:57:13.278231768 +0000 UTC m=+7333.807139029" lastFinishedPulling="2025-12-05 08:57:22.965699894 +0000 UTC m=+7343.494607155" observedRunningTime="2025-12-05 08:57:23.99974512 +0000 UTC m=+7344.528652401" watchObservedRunningTime="2025-12-05 08:57:24.006540802 +0000 UTC m=+7344.535448063" Dec 05 08:57:30 crc kubenswrapper[4997]: I1205 08:57:30.053769 4997 generic.go:334] "Generic (PLEG): container finished" podID="66bea143-bb89-4a9c-82cb-d0cff893d8ad" containerID="b978475c065f58bcfa921adf4ea2d5e99a71fadae7eccc3992831e6da1066e04" exitCode=0 Dec 05 08:57:30 crc kubenswrapper[4997]: I1205 08:57:30.053907 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bpbv8" event={"ID":"66bea143-bb89-4a9c-82cb-d0cff893d8ad","Type":"ContainerDied","Data":"b978475c065f58bcfa921adf4ea2d5e99a71fadae7eccc3992831e6da1066e04"} Dec 05 08:57:31 crc kubenswrapper[4997]: I1205 08:57:31.454156 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bpbv8" Dec 05 08:57:31 crc kubenswrapper[4997]: I1205 08:57:31.519466 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdd4v\" (UniqueName: \"kubernetes.io/projected/66bea143-bb89-4a9c-82cb-d0cff893d8ad-kube-api-access-cdd4v\") pod \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\" (UID: \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\") " Dec 05 08:57:31 crc kubenswrapper[4997]: I1205 08:57:31.519511 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-combined-ca-bundle\") pod \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\" (UID: \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\") " Dec 05 08:57:31 crc kubenswrapper[4997]: I1205 08:57:31.519536 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-scripts\") pod \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\" (UID: \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\") " Dec 05 08:57:31 crc kubenswrapper[4997]: I1205 08:57:31.519631 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-config-data\") pod \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\" (UID: \"66bea143-bb89-4a9c-82cb-d0cff893d8ad\") " Dec 05 08:57:31 crc kubenswrapper[4997]: I1205 08:57:31.525768 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66bea143-bb89-4a9c-82cb-d0cff893d8ad-kube-api-access-cdd4v" (OuterVolumeSpecName: "kube-api-access-cdd4v") pod "66bea143-bb89-4a9c-82cb-d0cff893d8ad" (UID: "66bea143-bb89-4a9c-82cb-d0cff893d8ad"). InnerVolumeSpecName "kube-api-access-cdd4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:57:31 crc kubenswrapper[4997]: I1205 08:57:31.526068 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-scripts" (OuterVolumeSpecName: "scripts") pod "66bea143-bb89-4a9c-82cb-d0cff893d8ad" (UID: "66bea143-bb89-4a9c-82cb-d0cff893d8ad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:31 crc kubenswrapper[4997]: I1205 08:57:31.555297 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66bea143-bb89-4a9c-82cb-d0cff893d8ad" (UID: "66bea143-bb89-4a9c-82cb-d0cff893d8ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:31 crc kubenswrapper[4997]: I1205 08:57:31.566046 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-config-data" (OuterVolumeSpecName: "config-data") pod "66bea143-bb89-4a9c-82cb-d0cff893d8ad" (UID: "66bea143-bb89-4a9c-82cb-d0cff893d8ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:31 crc kubenswrapper[4997]: I1205 08:57:31.622278 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:31 crc kubenswrapper[4997]: I1205 08:57:31.622315 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdd4v\" (UniqueName: \"kubernetes.io/projected/66bea143-bb89-4a9c-82cb-d0cff893d8ad-kube-api-access-cdd4v\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:31 crc kubenswrapper[4997]: I1205 08:57:31.622331 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:31 crc kubenswrapper[4997]: I1205 08:57:31.622344 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66bea143-bb89-4a9c-82cb-d0cff893d8ad-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.074679 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bpbv8" event={"ID":"66bea143-bb89-4a9c-82cb-d0cff893d8ad","Type":"ContainerDied","Data":"0520b831857d86e565750f200495a55ee031d8e7bf2984014bbacc5663ba7481"} Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.074717 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0520b831857d86e565750f200495a55ee031d8e7bf2984014bbacc5663ba7481" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.074834 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bpbv8" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.168063 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:57:32 crc kubenswrapper[4997]: E1205 08:57:32.168421 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66bea143-bb89-4a9c-82cb-d0cff893d8ad" containerName="nova-cell0-conductor-db-sync" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.168438 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="66bea143-bb89-4a9c-82cb-d0cff893d8ad" containerName="nova-cell0-conductor-db-sync" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.168637 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="66bea143-bb89-4a9c-82cb-d0cff893d8ad" containerName="nova-cell0-conductor-db-sync" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.169208 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.172657 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2jlfs" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.174679 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.180443 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.233676 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/187e8113-1011-4622-bd2b-a582a9450a6f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"187e8113-1011-4622-bd2b-a582a9450a6f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.233813 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l25k\" (UniqueName: \"kubernetes.io/projected/187e8113-1011-4622-bd2b-a582a9450a6f-kube-api-access-8l25k\") pod \"nova-cell0-conductor-0\" (UID: \"187e8113-1011-4622-bd2b-a582a9450a6f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.233931 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/187e8113-1011-4622-bd2b-a582a9450a6f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"187e8113-1011-4622-bd2b-a582a9450a6f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.335605 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l25k\" (UniqueName: \"kubernetes.io/projected/187e8113-1011-4622-bd2b-a582a9450a6f-kube-api-access-8l25k\") pod \"nova-cell0-conductor-0\" (UID: \"187e8113-1011-4622-bd2b-a582a9450a6f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.335822 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/187e8113-1011-4622-bd2b-a582a9450a6f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"187e8113-1011-4622-bd2b-a582a9450a6f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.336855 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/187e8113-1011-4622-bd2b-a582a9450a6f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"187e8113-1011-4622-bd2b-a582a9450a6f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.340435 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/187e8113-1011-4622-bd2b-a582a9450a6f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"187e8113-1011-4622-bd2b-a582a9450a6f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.341414 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/187e8113-1011-4622-bd2b-a582a9450a6f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"187e8113-1011-4622-bd2b-a582a9450a6f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.351323 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l25k\" (UniqueName: \"kubernetes.io/projected/187e8113-1011-4622-bd2b-a582a9450a6f-kube-api-access-8l25k\") pod \"nova-cell0-conductor-0\" (UID: \"187e8113-1011-4622-bd2b-a582a9450a6f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.499184 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 08:57:32 crc kubenswrapper[4997]: I1205 08:57:32.917954 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:57:33 crc kubenswrapper[4997]: I1205 08:57:33.086441 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"187e8113-1011-4622-bd2b-a582a9450a6f","Type":"ContainerStarted","Data":"4f8a15ececd1a65272d44ea356426700d9be18b5b1fc57b73a61d3252cd00a69"} Dec 05 08:57:34 crc kubenswrapper[4997]: I1205 08:57:34.098182 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"187e8113-1011-4622-bd2b-a582a9450a6f","Type":"ContainerStarted","Data":"0c4171ada7a2365ff05d143c62c6b5d2ff029f4808c2fc08d95ee1888d1b7557"} Dec 05 08:57:34 crc kubenswrapper[4997]: I1205 08:57:34.099494 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 05 08:57:34 crc kubenswrapper[4997]: I1205 08:57:34.127085 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.127044299 podStartE2EDuration="2.127044299s" podCreationTimestamp="2025-12-05 08:57:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:57:34.114672096 +0000 UTC m=+7354.643579367" watchObservedRunningTime="2025-12-05 08:57:34.127044299 +0000 UTC m=+7354.655951560" Dec 05 08:57:42 crc kubenswrapper[4997]: I1205 08:57:42.201098 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ddtzs"] Dec 05 08:57:42 crc kubenswrapper[4997]: I1205 08:57:42.204597 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ddtzs" Dec 05 08:57:42 crc kubenswrapper[4997]: I1205 08:57:42.222762 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ddtzs"] Dec 05 08:57:42 crc kubenswrapper[4997]: I1205 08:57:42.253666 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb204e34-3079-463a-833b-fcd2e5d2dd29-catalog-content\") pod \"redhat-marketplace-ddtzs\" (UID: \"bb204e34-3079-463a-833b-fcd2e5d2dd29\") " pod="openshift-marketplace/redhat-marketplace-ddtzs" Dec 05 08:57:42 crc kubenswrapper[4997]: I1205 08:57:42.253762 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb204e34-3079-463a-833b-fcd2e5d2dd29-utilities\") pod \"redhat-marketplace-ddtzs\" (UID: \"bb204e34-3079-463a-833b-fcd2e5d2dd29\") " pod="openshift-marketplace/redhat-marketplace-ddtzs" Dec 05 08:57:42 crc kubenswrapper[4997]: I1205 08:57:42.253997 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hsng\" (UniqueName: \"kubernetes.io/projected/bb204e34-3079-463a-833b-fcd2e5d2dd29-kube-api-access-4hsng\") pod \"redhat-marketplace-ddtzs\" (UID: \"bb204e34-3079-463a-833b-fcd2e5d2dd29\") " pod="openshift-marketplace/redhat-marketplace-ddtzs" Dec 05 08:57:42 crc kubenswrapper[4997]: I1205 08:57:42.355305 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hsng\" (UniqueName: \"kubernetes.io/projected/bb204e34-3079-463a-833b-fcd2e5d2dd29-kube-api-access-4hsng\") pod \"redhat-marketplace-ddtzs\" (UID: \"bb204e34-3079-463a-833b-fcd2e5d2dd29\") " pod="openshift-marketplace/redhat-marketplace-ddtzs" Dec 05 08:57:42 crc kubenswrapper[4997]: I1205 08:57:42.355372 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb204e34-3079-463a-833b-fcd2e5d2dd29-catalog-content\") pod \"redhat-marketplace-ddtzs\" (UID: \"bb204e34-3079-463a-833b-fcd2e5d2dd29\") " pod="openshift-marketplace/redhat-marketplace-ddtzs" Dec 05 08:57:42 crc kubenswrapper[4997]: I1205 08:57:42.355407 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb204e34-3079-463a-833b-fcd2e5d2dd29-utilities\") pod \"redhat-marketplace-ddtzs\" (UID: \"bb204e34-3079-463a-833b-fcd2e5d2dd29\") " pod="openshift-marketplace/redhat-marketplace-ddtzs" Dec 05 08:57:42 crc kubenswrapper[4997]: I1205 08:57:42.355815 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb204e34-3079-463a-833b-fcd2e5d2dd29-utilities\") pod \"redhat-marketplace-ddtzs\" (UID: \"bb204e34-3079-463a-833b-fcd2e5d2dd29\") " pod="openshift-marketplace/redhat-marketplace-ddtzs" Dec 05 08:57:42 crc kubenswrapper[4997]: I1205 08:57:42.356307 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb204e34-3079-463a-833b-fcd2e5d2dd29-catalog-content\") pod \"redhat-marketplace-ddtzs\" (UID: \"bb204e34-3079-463a-833b-fcd2e5d2dd29\") " pod="openshift-marketplace/redhat-marketplace-ddtzs" Dec 05 08:57:42 crc kubenswrapper[4997]: I1205 08:57:42.375324 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hsng\" (UniqueName: \"kubernetes.io/projected/bb204e34-3079-463a-833b-fcd2e5d2dd29-kube-api-access-4hsng\") pod \"redhat-marketplace-ddtzs\" (UID: \"bb204e34-3079-463a-833b-fcd2e5d2dd29\") " pod="openshift-marketplace/redhat-marketplace-ddtzs" Dec 05 08:57:42 crc kubenswrapper[4997]: I1205 08:57:42.529266 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 05 08:57:42 crc kubenswrapper[4997]: I1205 08:57:42.537457 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ddtzs" Dec 05 08:57:42 crc kubenswrapper[4997]: I1205 08:57:42.811382 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ddtzs"] Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.008000 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-n4vrb"] Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.009358 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-n4vrb" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.012061 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.012511 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.026805 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-n4vrb"] Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.067860 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-n4vrb\" (UID: \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\") " pod="openstack/nova-cell0-cell-mapping-n4vrb" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.068209 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-scripts\") pod \"nova-cell0-cell-mapping-n4vrb\" (UID: \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\") " pod="openstack/nova-cell0-cell-mapping-n4vrb" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.068244 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-config-data\") pod \"nova-cell0-cell-mapping-n4vrb\" (UID: \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\") " pod="openstack/nova-cell0-cell-mapping-n4vrb" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.068391 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmqcc\" (UniqueName: \"kubernetes.io/projected/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-kube-api-access-dmqcc\") pod \"nova-cell0-cell-mapping-n4vrb\" (UID: \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\") " pod="openstack/nova-cell0-cell-mapping-n4vrb" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.130379 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.132261 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.134748 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.145603 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.172724 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3154810c-312d-4ac4-872f-7becd4fad42c-config-data\") pod \"nova-api-0\" (UID: \"3154810c-312d-4ac4-872f-7becd4fad42c\") " pod="openstack/nova-api-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.172773 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdfd9\" (UniqueName: \"kubernetes.io/projected/3154810c-312d-4ac4-872f-7becd4fad42c-kube-api-access-sdfd9\") pod \"nova-api-0\" (UID: \"3154810c-312d-4ac4-872f-7becd4fad42c\") " pod="openstack/nova-api-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.172845 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-n4vrb\" (UID: \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\") " pod="openstack/nova-cell0-cell-mapping-n4vrb" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.172873 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-scripts\") pod \"nova-cell0-cell-mapping-n4vrb\" (UID: \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\") " pod="openstack/nova-cell0-cell-mapping-n4vrb" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.172898 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-config-data\") pod \"nova-cell0-cell-mapping-n4vrb\" (UID: \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\") " pod="openstack/nova-cell0-cell-mapping-n4vrb" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.172956 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3154810c-312d-4ac4-872f-7becd4fad42c-logs\") pod \"nova-api-0\" (UID: \"3154810c-312d-4ac4-872f-7becd4fad42c\") " pod="openstack/nova-api-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.172974 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3154810c-312d-4ac4-872f-7becd4fad42c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3154810c-312d-4ac4-872f-7becd4fad42c\") " pod="openstack/nova-api-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.173002 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmqcc\" (UniqueName: \"kubernetes.io/projected/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-kube-api-access-dmqcc\") pod \"nova-cell0-cell-mapping-n4vrb\" (UID: \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\") " pod="openstack/nova-cell0-cell-mapping-n4vrb" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.179747 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-scripts\") pod \"nova-cell0-cell-mapping-n4vrb\" (UID: \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\") " pod="openstack/nova-cell0-cell-mapping-n4vrb" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.182049 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-n4vrb\" (UID: \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\") " pod="openstack/nova-cell0-cell-mapping-n4vrb" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.182828 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-config-data\") pod \"nova-cell0-cell-mapping-n4vrb\" (UID: \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\") " pod="openstack/nova-cell0-cell-mapping-n4vrb" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.207253 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmqcc\" (UniqueName: \"kubernetes.io/projected/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-kube-api-access-dmqcc\") pod \"nova-cell0-cell-mapping-n4vrb\" (UID: \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\") " pod="openstack/nova-cell0-cell-mapping-n4vrb" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.210694 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.212048 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.218918 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.224294 4997 generic.go:334] "Generic (PLEG): container finished" podID="bb204e34-3079-463a-833b-fcd2e5d2dd29" containerID="9e8cf71b7f0898961ed6b3b9f800f532557dbab642ccdb154c3862a8e2e2d2cc" exitCode=0 Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.224334 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddtzs" event={"ID":"bb204e34-3079-463a-833b-fcd2e5d2dd29","Type":"ContainerDied","Data":"9e8cf71b7f0898961ed6b3b9f800f532557dbab642ccdb154c3862a8e2e2d2cc"} Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.224364 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddtzs" event={"ID":"bb204e34-3079-463a-833b-fcd2e5d2dd29","Type":"ContainerStarted","Data":"b3a80eb07f1f3c03b44fe3c090c6a96ed779d5c5858a0950ac9dbedfaaecd76f"} Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.277946 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.279721 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-config-data\") pod \"nova-metadata-0\" (UID: \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.279854 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz4x7\" (UniqueName: \"kubernetes.io/projected/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-kube-api-access-pz4x7\") pod \"nova-metadata-0\" (UID: \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.280107 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.280140 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3154810c-312d-4ac4-872f-7becd4fad42c-logs\") pod \"nova-api-0\" (UID: \"3154810c-312d-4ac4-872f-7becd4fad42c\") " pod="openstack/nova-api-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.280163 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3154810c-312d-4ac4-872f-7becd4fad42c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3154810c-312d-4ac4-872f-7becd4fad42c\") " pod="openstack/nova-api-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.280212 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3154810c-312d-4ac4-872f-7becd4fad42c-config-data\") pod \"nova-api-0\" (UID: \"3154810c-312d-4ac4-872f-7becd4fad42c\") " pod="openstack/nova-api-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.280239 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdfd9\" (UniqueName: \"kubernetes.io/projected/3154810c-312d-4ac4-872f-7becd4fad42c-kube-api-access-sdfd9\") pod \"nova-api-0\" (UID: \"3154810c-312d-4ac4-872f-7becd4fad42c\") " pod="openstack/nova-api-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.280266 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-logs\") pod \"nova-metadata-0\" (UID: \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.281203 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3154810c-312d-4ac4-872f-7becd4fad42c-logs\") pod \"nova-api-0\" (UID: \"3154810c-312d-4ac4-872f-7becd4fad42c\") " pod="openstack/nova-api-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.290262 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3154810c-312d-4ac4-872f-7becd4fad42c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3154810c-312d-4ac4-872f-7becd4fad42c\") " pod="openstack/nova-api-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.293326 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3154810c-312d-4ac4-872f-7becd4fad42c-config-data\") pod \"nova-api-0\" (UID: \"3154810c-312d-4ac4-872f-7becd4fad42c\") " pod="openstack/nova-api-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.300812 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.302043 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.303910 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdfd9\" (UniqueName: \"kubernetes.io/projected/3154810c-312d-4ac4-872f-7becd4fad42c-kube-api-access-sdfd9\") pod \"nova-api-0\" (UID: \"3154810c-312d-4ac4-872f-7becd4fad42c\") " pod="openstack/nova-api-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.307849 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.322150 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.348039 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-n4vrb" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.384295 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.385651 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-logs\") pod \"nova-metadata-0\" (UID: \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.385711 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b731c82f-0c89-4897-ab96-bf5fe10adbde-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b731c82f-0c89-4897-ab96-bf5fe10adbde\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.385760 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-config-data\") pod \"nova-metadata-0\" (UID: \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.385769 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.385815 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b731c82f-0c89-4897-ab96-bf5fe10adbde-config-data\") pod \"nova-scheduler-0\" (UID: \"b731c82f-0c89-4897-ab96-bf5fe10adbde\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.385843 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz4x7\" (UniqueName: \"kubernetes.io/projected/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-kube-api-access-pz4x7\") pod \"nova-metadata-0\" (UID: \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.385862 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdmz2\" (UniqueName: \"kubernetes.io/projected/b731c82f-0c89-4897-ab96-bf5fe10adbde-kube-api-access-tdmz2\") pod \"nova-scheduler-0\" (UID: \"b731c82f-0c89-4897-ab96-bf5fe10adbde\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.385889 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.386759 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-logs\") pod \"nova-metadata-0\" (UID: \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.389697 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.389889 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.391582 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-config-data\") pod \"nova-metadata-0\" (UID: \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.425004 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz4x7\" (UniqueName: \"kubernetes.io/projected/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-kube-api-access-pz4x7\") pod \"nova-metadata-0\" (UID: \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.449692 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.451592 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cd6b446c-ncp6v"] Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.453183 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.487077 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e1c50e-f246-4d92-ae8c-857da4eb1c10-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"58e1c50e-f246-4d92-ae8c-857da4eb1c10\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.487416 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz668\" (UniqueName: \"kubernetes.io/projected/d66c0d05-f098-459a-9398-cf7b1ae4edf1-kube-api-access-bz668\") pod \"dnsmasq-dns-7cd6b446c-ncp6v\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.487455 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-ovsdbserver-nb\") pod \"dnsmasq-dns-7cd6b446c-ncp6v\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.487799 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zfr7\" (UniqueName: \"kubernetes.io/projected/58e1c50e-f246-4d92-ae8c-857da4eb1c10-kube-api-access-6zfr7\") pod \"nova-cell1-novncproxy-0\" (UID: \"58e1c50e-f246-4d92-ae8c-857da4eb1c10\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.487834 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-config\") pod \"dnsmasq-dns-7cd6b446c-ncp6v\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.487887 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b731c82f-0c89-4897-ab96-bf5fe10adbde-config-data\") pod \"nova-scheduler-0\" (UID: \"b731c82f-0c89-4897-ab96-bf5fe10adbde\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.487928 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdmz2\" (UniqueName: \"kubernetes.io/projected/b731c82f-0c89-4897-ab96-bf5fe10adbde-kube-api-access-tdmz2\") pod \"nova-scheduler-0\" (UID: \"b731c82f-0c89-4897-ab96-bf5fe10adbde\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.487965 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-ovsdbserver-sb\") pod \"dnsmasq-dns-7cd6b446c-ncp6v\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.488061 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-dns-svc\") pod \"dnsmasq-dns-7cd6b446c-ncp6v\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.488087 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b731c82f-0c89-4897-ab96-bf5fe10adbde-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b731c82f-0c89-4897-ab96-bf5fe10adbde\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.488119 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58e1c50e-f246-4d92-ae8c-857da4eb1c10-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"58e1c50e-f246-4d92-ae8c-857da4eb1c10\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.494227 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b731c82f-0c89-4897-ab96-bf5fe10adbde-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b731c82f-0c89-4897-ab96-bf5fe10adbde\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.498289 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b731c82f-0c89-4897-ab96-bf5fe10adbde-config-data\") pod \"nova-scheduler-0\" (UID: \"b731c82f-0c89-4897-ab96-bf5fe10adbde\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.519371 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.528758 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cd6b446c-ncp6v"] Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.537253 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.545348 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdmz2\" (UniqueName: \"kubernetes.io/projected/b731c82f-0c89-4897-ab96-bf5fe10adbde-kube-api-access-tdmz2\") pod \"nova-scheduler-0\" (UID: \"b731c82f-0c89-4897-ab96-bf5fe10adbde\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.589811 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-dns-svc\") pod \"dnsmasq-dns-7cd6b446c-ncp6v\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.589855 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58e1c50e-f246-4d92-ae8c-857da4eb1c10-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"58e1c50e-f246-4d92-ae8c-857da4eb1c10\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.589901 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e1c50e-f246-4d92-ae8c-857da4eb1c10-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"58e1c50e-f246-4d92-ae8c-857da4eb1c10\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.589917 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz668\" (UniqueName: \"kubernetes.io/projected/d66c0d05-f098-459a-9398-cf7b1ae4edf1-kube-api-access-bz668\") pod \"dnsmasq-dns-7cd6b446c-ncp6v\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.589945 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-ovsdbserver-nb\") pod \"dnsmasq-dns-7cd6b446c-ncp6v\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.589974 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zfr7\" (UniqueName: \"kubernetes.io/projected/58e1c50e-f246-4d92-ae8c-857da4eb1c10-kube-api-access-6zfr7\") pod \"nova-cell1-novncproxy-0\" (UID: \"58e1c50e-f246-4d92-ae8c-857da4eb1c10\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.589994 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-config\") pod \"dnsmasq-dns-7cd6b446c-ncp6v\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.590056 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-ovsdbserver-sb\") pod \"dnsmasq-dns-7cd6b446c-ncp6v\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.591197 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-ovsdbserver-sb\") pod \"dnsmasq-dns-7cd6b446c-ncp6v\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.591242 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-dns-svc\") pod \"dnsmasq-dns-7cd6b446c-ncp6v\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.592172 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-ovsdbserver-nb\") pod \"dnsmasq-dns-7cd6b446c-ncp6v\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.594279 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-config\") pod \"dnsmasq-dns-7cd6b446c-ncp6v\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.596771 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58e1c50e-f246-4d92-ae8c-857da4eb1c10-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"58e1c50e-f246-4d92-ae8c-857da4eb1c10\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.598252 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e1c50e-f246-4d92-ae8c-857da4eb1c10-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"58e1c50e-f246-4d92-ae8c-857da4eb1c10\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.614486 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz668\" (UniqueName: \"kubernetes.io/projected/d66c0d05-f098-459a-9398-cf7b1ae4edf1-kube-api-access-bz668\") pod \"dnsmasq-dns-7cd6b446c-ncp6v\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.615793 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zfr7\" (UniqueName: \"kubernetes.io/projected/58e1c50e-f246-4d92-ae8c-857da4eb1c10-kube-api-access-6zfr7\") pod \"nova-cell1-novncproxy-0\" (UID: \"58e1c50e-f246-4d92-ae8c-857da4eb1c10\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.629042 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.846383 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:57:43 crc kubenswrapper[4997]: W1205 08:57:43.879500 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3154810c_312d_4ac4_872f_7becd4fad42c.slice/crio-fcb0f0afbae2301f5c2042a77b29abf1a31c260214b623f40b35aa4ce79498b7 WatchSource:0}: Error finding container fcb0f0afbae2301f5c2042a77b29abf1a31c260214b623f40b35aa4ce79498b7: Status 404 returned error can't find the container with id fcb0f0afbae2301f5c2042a77b29abf1a31c260214b623f40b35aa4ce79498b7 Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.882721 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.896380 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.931411 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-n4vrb"] Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.939526 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6vfwt"] Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.950457 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6vfwt"] Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.950539 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6vfwt" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.953083 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 08:57:43 crc kubenswrapper[4997]: I1205 08:57:43.953230 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.013667 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-scripts\") pod \"nova-cell1-conductor-db-sync-6vfwt\" (UID: \"b195cd20-21ad-4b84-b5dd-911265d4a823\") " pod="openstack/nova-cell1-conductor-db-sync-6vfwt" Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.014769 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-config-data\") pod \"nova-cell1-conductor-db-sync-6vfwt\" (UID: \"b195cd20-21ad-4b84-b5dd-911265d4a823\") " pod="openstack/nova-cell1-conductor-db-sync-6vfwt" Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.014861 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf2jf\" (UniqueName: \"kubernetes.io/projected/b195cd20-21ad-4b84-b5dd-911265d4a823-kube-api-access-zf2jf\") pod \"nova-cell1-conductor-db-sync-6vfwt\" (UID: \"b195cd20-21ad-4b84-b5dd-911265d4a823\") " pod="openstack/nova-cell1-conductor-db-sync-6vfwt" Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.014893 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6vfwt\" (UID: \"b195cd20-21ad-4b84-b5dd-911265d4a823\") " pod="openstack/nova-cell1-conductor-db-sync-6vfwt" Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.116501 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-config-data\") pod \"nova-cell1-conductor-db-sync-6vfwt\" (UID: \"b195cd20-21ad-4b84-b5dd-911265d4a823\") " pod="openstack/nova-cell1-conductor-db-sync-6vfwt" Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.116816 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf2jf\" (UniqueName: \"kubernetes.io/projected/b195cd20-21ad-4b84-b5dd-911265d4a823-kube-api-access-zf2jf\") pod \"nova-cell1-conductor-db-sync-6vfwt\" (UID: \"b195cd20-21ad-4b84-b5dd-911265d4a823\") " pod="openstack/nova-cell1-conductor-db-sync-6vfwt" Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.116841 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6vfwt\" (UID: \"b195cd20-21ad-4b84-b5dd-911265d4a823\") " pod="openstack/nova-cell1-conductor-db-sync-6vfwt" Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.116904 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-scripts\") pod \"nova-cell1-conductor-db-sync-6vfwt\" (UID: \"b195cd20-21ad-4b84-b5dd-911265d4a823\") " pod="openstack/nova-cell1-conductor-db-sync-6vfwt" Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.123637 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-scripts\") pod \"nova-cell1-conductor-db-sync-6vfwt\" (UID: \"b195cd20-21ad-4b84-b5dd-911265d4a823\") " pod="openstack/nova-cell1-conductor-db-sync-6vfwt" Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.123950 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-config-data\") pod \"nova-cell1-conductor-db-sync-6vfwt\" (UID: \"b195cd20-21ad-4b84-b5dd-911265d4a823\") " pod="openstack/nova-cell1-conductor-db-sync-6vfwt" Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.129529 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6vfwt\" (UID: \"b195cd20-21ad-4b84-b5dd-911265d4a823\") " pod="openstack/nova-cell1-conductor-db-sync-6vfwt" Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.136525 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf2jf\" (UniqueName: \"kubernetes.io/projected/b195cd20-21ad-4b84-b5dd-911265d4a823-kube-api-access-zf2jf\") pod \"nova-cell1-conductor-db-sync-6vfwt\" (UID: \"b195cd20-21ad-4b84-b5dd-911265d4a823\") " pod="openstack/nova-cell1-conductor-db-sync-6vfwt" Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.223598 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.243657 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-n4vrb" event={"ID":"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3","Type":"ContainerStarted","Data":"bb93c1917706fc42241924f8d7bd1c0dcd0b6e200158acbeadac262575c2fcef"} Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.243698 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-n4vrb" event={"ID":"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3","Type":"ContainerStarted","Data":"5487e3d1d30f7c87006eb06fc3f15c8ae1152bf14c677af9e7d07b1d1d1ce769"} Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.246471 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3154810c-312d-4ac4-872f-7becd4fad42c","Type":"ContainerStarted","Data":"fcb0f0afbae2301f5c2042a77b29abf1a31c260214b623f40b35aa4ce79498b7"} Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.263930 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-n4vrb" podStartSLOduration=2.263912205 podStartE2EDuration="2.263912205s" podCreationTimestamp="2025-12-05 08:57:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:57:44.263864534 +0000 UTC m=+7364.792771805" watchObservedRunningTime="2025-12-05 08:57:44.263912205 +0000 UTC m=+7364.792819466" Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.272956 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6vfwt" Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.303891 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.426886 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:57:44 crc kubenswrapper[4997]: W1205 08:57:44.443942 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58e1c50e_f246_4d92_ae8c_857da4eb1c10.slice/crio-a5b80aa9aabc8d3d5cca45a8140c21cdb883930555095ebe0daf78b36f64b324 WatchSource:0}: Error finding container a5b80aa9aabc8d3d5cca45a8140c21cdb883930555095ebe0daf78b36f64b324: Status 404 returned error can't find the container with id a5b80aa9aabc8d3d5cca45a8140c21cdb883930555095ebe0daf78b36f64b324 Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.444047 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cd6b446c-ncp6v"] Dec 05 08:57:44 crc kubenswrapper[4997]: W1205 08:57:44.447775 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd66c0d05_f098_459a_9398_cf7b1ae4edf1.slice/crio-ea78f35038212f2000fb6b17cdb8e025a9e91643e1e936e1c7e49fe2147783ac WatchSource:0}: Error finding container ea78f35038212f2000fb6b17cdb8e025a9e91643e1e936e1c7e49fe2147783ac: Status 404 returned error can't find the container with id ea78f35038212f2000fb6b17cdb8e025a9e91643e1e936e1c7e49fe2147783ac Dec 05 08:57:44 crc kubenswrapper[4997]: I1205 08:57:44.593573 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6vfwt"] Dec 05 08:57:44 crc kubenswrapper[4997]: W1205 08:57:44.600650 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb195cd20_21ad_4b84_b5dd_911265d4a823.slice/crio-1fe552ddb0103d73e734cc57bddbdfedae2a5a20b3b8a9b84d6e0f2324b0c2ee WatchSource:0}: Error finding container 1fe552ddb0103d73e734cc57bddbdfedae2a5a20b3b8a9b84d6e0f2324b0c2ee: Status 404 returned error can't find the container with id 1fe552ddb0103d73e734cc57bddbdfedae2a5a20b3b8a9b84d6e0f2324b0c2ee Dec 05 08:57:45 crc kubenswrapper[4997]: I1205 08:57:45.277757 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e2dd2984-e4c8-406c-91a0-b41df0b17b8f","Type":"ContainerStarted","Data":"85d9ed6f3a21e2b45253121ab9e0eb357db2bad939658f3829a92a9e499dc737"} Dec 05 08:57:45 crc kubenswrapper[4997]: I1205 08:57:45.283513 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"58e1c50e-f246-4d92-ae8c-857da4eb1c10","Type":"ContainerStarted","Data":"a5b80aa9aabc8d3d5cca45a8140c21cdb883930555095ebe0daf78b36f64b324"} Dec 05 08:57:45 crc kubenswrapper[4997]: I1205 08:57:45.286114 4997 generic.go:334] "Generic (PLEG): container finished" podID="bb204e34-3079-463a-833b-fcd2e5d2dd29" containerID="5de72f4aa8c1f52947b1965ec66a5e532f6d3c3355fef628f5b9b7aeae7ab210" exitCode=0 Dec 05 08:57:45 crc kubenswrapper[4997]: I1205 08:57:45.286225 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddtzs" event={"ID":"bb204e34-3079-463a-833b-fcd2e5d2dd29","Type":"ContainerDied","Data":"5de72f4aa8c1f52947b1965ec66a5e532f6d3c3355fef628f5b9b7aeae7ab210"} Dec 05 08:57:45 crc kubenswrapper[4997]: I1205 08:57:45.289103 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6vfwt" event={"ID":"b195cd20-21ad-4b84-b5dd-911265d4a823","Type":"ContainerStarted","Data":"971f750d4d3626d8b4a1c5eb133f6f45e6bae2d49b9d31f1f3b26d326de5403f"} Dec 05 08:57:45 crc kubenswrapper[4997]: I1205 08:57:45.289465 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6vfwt" event={"ID":"b195cd20-21ad-4b84-b5dd-911265d4a823","Type":"ContainerStarted","Data":"1fe552ddb0103d73e734cc57bddbdfedae2a5a20b3b8a9b84d6e0f2324b0c2ee"} Dec 05 08:57:45 crc kubenswrapper[4997]: I1205 08:57:45.291795 4997 generic.go:334] "Generic (PLEG): container finished" podID="d66c0d05-f098-459a-9398-cf7b1ae4edf1" containerID="393d8d1969cce96ea0817d1a04cb970a551bb7554378965786e81c0c048d659b" exitCode=0 Dec 05 08:57:45 crc kubenswrapper[4997]: I1205 08:57:45.291881 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" event={"ID":"d66c0d05-f098-459a-9398-cf7b1ae4edf1","Type":"ContainerDied","Data":"393d8d1969cce96ea0817d1a04cb970a551bb7554378965786e81c0c048d659b"} Dec 05 08:57:45 crc kubenswrapper[4997]: I1205 08:57:45.291925 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" event={"ID":"d66c0d05-f098-459a-9398-cf7b1ae4edf1","Type":"ContainerStarted","Data":"ea78f35038212f2000fb6b17cdb8e025a9e91643e1e936e1c7e49fe2147783ac"} Dec 05 08:57:45 crc kubenswrapper[4997]: I1205 08:57:45.293780 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b731c82f-0c89-4897-ab96-bf5fe10adbde","Type":"ContainerStarted","Data":"4c8cdc3c7f0d42ec4adfa2e042179ebb4d28b5e08b674d0e3311725a3f6ea367"} Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.321202 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b731c82f-0c89-4897-ab96-bf5fe10adbde","Type":"ContainerStarted","Data":"a65e0964ec99b944af43aadba0c305b732dbe20f3cad10b804dbe225467077f0"} Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.324186 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e2dd2984-e4c8-406c-91a0-b41df0b17b8f","Type":"ContainerStarted","Data":"efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685"} Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.324212 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e2dd2984-e4c8-406c-91a0-b41df0b17b8f","Type":"ContainerStarted","Data":"3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407"} Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.326355 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"58e1c50e-f246-4d92-ae8c-857da4eb1c10","Type":"ContainerStarted","Data":"0125569bd24894dec6abb10d0d93c20cd62c994116636351ae75d86287e14989"} Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.329064 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddtzs" event={"ID":"bb204e34-3079-463a-833b-fcd2e5d2dd29","Type":"ContainerStarted","Data":"3d4ec069a69e22ac7bdab98e76e5b8d29b174359374d39b6d056c58745188140"} Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.332132 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3154810c-312d-4ac4-872f-7becd4fad42c","Type":"ContainerStarted","Data":"a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4"} Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.332168 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3154810c-312d-4ac4-872f-7becd4fad42c","Type":"ContainerStarted","Data":"c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4"} Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.334138 4997 generic.go:334] "Generic (PLEG): container finished" podID="b195cd20-21ad-4b84-b5dd-911265d4a823" containerID="971f750d4d3626d8b4a1c5eb133f6f45e6bae2d49b9d31f1f3b26d326de5403f" exitCode=0 Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.334200 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6vfwt" event={"ID":"b195cd20-21ad-4b84-b5dd-911265d4a823","Type":"ContainerDied","Data":"971f750d4d3626d8b4a1c5eb133f6f45e6bae2d49b9d31f1f3b26d326de5403f"} Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.337360 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" event={"ID":"d66c0d05-f098-459a-9398-cf7b1ae4edf1","Type":"ContainerStarted","Data":"afe2705efd72071cb13623dcd699f8cd8b042f47f2111908cff7f90a10632d8f"} Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.337587 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.347924 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.121839547 podStartE2EDuration="5.347897376s" podCreationTimestamp="2025-12-05 08:57:43 +0000 UTC" firstStartedPulling="2025-12-05 08:57:44.306602443 +0000 UTC m=+7364.835509714" lastFinishedPulling="2025-12-05 08:57:47.532660282 +0000 UTC m=+7368.061567543" observedRunningTime="2025-12-05 08:57:48.340796225 +0000 UTC m=+7368.869703496" watchObservedRunningTime="2025-12-05 08:57:48.347897376 +0000 UTC m=+7368.876804637" Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.350489 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-6vfwt" podStartSLOduration=5.350478025 podStartE2EDuration="5.350478025s" podCreationTimestamp="2025-12-05 08:57:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:57:45.35556113 +0000 UTC m=+7365.884468391" watchObservedRunningTime="2025-12-05 08:57:48.350478025 +0000 UTC m=+7368.879385286" Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.363201 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" podStartSLOduration=5.363183427 podStartE2EDuration="5.363183427s" podCreationTimestamp="2025-12-05 08:57:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:57:48.357692609 +0000 UTC m=+7368.886599880" watchObservedRunningTime="2025-12-05 08:57:48.363183427 +0000 UTC m=+7368.892090678" Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.420570 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ddtzs" podStartSLOduration=2.110696697 podStartE2EDuration="6.420548919s" podCreationTimestamp="2025-12-05 08:57:42 +0000 UTC" firstStartedPulling="2025-12-05 08:57:43.226236502 +0000 UTC m=+7363.755143763" lastFinishedPulling="2025-12-05 08:57:47.536088714 +0000 UTC m=+7368.064995985" observedRunningTime="2025-12-05 08:57:48.417580519 +0000 UTC m=+7368.946487780" watchObservedRunningTime="2025-12-05 08:57:48.420548919 +0000 UTC m=+7368.949456180" Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.427153 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.347299377 podStartE2EDuration="5.427132856s" podCreationTimestamp="2025-12-05 08:57:43 +0000 UTC" firstStartedPulling="2025-12-05 08:57:44.448540548 +0000 UTC m=+7364.977447809" lastFinishedPulling="2025-12-05 08:57:47.528374027 +0000 UTC m=+7368.057281288" observedRunningTime="2025-12-05 08:57:48.39752258 +0000 UTC m=+7368.926429921" watchObservedRunningTime="2025-12-05 08:57:48.427132856 +0000 UTC m=+7368.956040117" Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.437185 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.7928490940000001 podStartE2EDuration="5.437165006s" podCreationTimestamp="2025-12-05 08:57:43 +0000 UTC" firstStartedPulling="2025-12-05 08:57:43.890058636 +0000 UTC m=+7364.418965897" lastFinishedPulling="2025-12-05 08:57:47.534374528 +0000 UTC m=+7368.063281809" observedRunningTime="2025-12-05 08:57:48.432090349 +0000 UTC m=+7368.960997620" watchObservedRunningTime="2025-12-05 08:57:48.437165006 +0000 UTC m=+7368.966072267" Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.455178 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.159300224 podStartE2EDuration="5.45516163s" podCreationTimestamp="2025-12-05 08:57:43 +0000 UTC" firstStartedPulling="2025-12-05 08:57:44.238200534 +0000 UTC m=+7364.767107805" lastFinishedPulling="2025-12-05 08:57:47.53406194 +0000 UTC m=+7368.062969211" observedRunningTime="2025-12-05 08:57:48.448289175 +0000 UTC m=+7368.977196426" watchObservedRunningTime="2025-12-05 08:57:48.45516163 +0000 UTC m=+7368.984068891" Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.538442 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.538539 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.629652 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 08:57:48 crc kubenswrapper[4997]: I1205 08:57:48.882954 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:57:49 crc kubenswrapper[4997]: I1205 08:57:49.345846 4997 generic.go:334] "Generic (PLEG): container finished" podID="6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3" containerID="bb93c1917706fc42241924f8d7bd1c0dcd0b6e200158acbeadac262575c2fcef" exitCode=0 Dec 05 08:57:49 crc kubenswrapper[4997]: I1205 08:57:49.345996 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-n4vrb" event={"ID":"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3","Type":"ContainerDied","Data":"bb93c1917706fc42241924f8d7bd1c0dcd0b6e200158acbeadac262575c2fcef"} Dec 05 08:57:49 crc kubenswrapper[4997]: I1205 08:57:49.755325 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6vfwt" Dec 05 08:57:49 crc kubenswrapper[4997]: I1205 08:57:49.770666 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:57:49 crc kubenswrapper[4997]: I1205 08:57:49.770729 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:57:49 crc kubenswrapper[4997]: I1205 08:57:49.867833 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-scripts\") pod \"b195cd20-21ad-4b84-b5dd-911265d4a823\" (UID: \"b195cd20-21ad-4b84-b5dd-911265d4a823\") " Dec 05 08:57:49 crc kubenswrapper[4997]: I1205 08:57:49.867886 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-combined-ca-bundle\") pod \"b195cd20-21ad-4b84-b5dd-911265d4a823\" (UID: \"b195cd20-21ad-4b84-b5dd-911265d4a823\") " Dec 05 08:57:49 crc kubenswrapper[4997]: I1205 08:57:49.867911 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-config-data\") pod \"b195cd20-21ad-4b84-b5dd-911265d4a823\" (UID: \"b195cd20-21ad-4b84-b5dd-911265d4a823\") " Dec 05 08:57:49 crc kubenswrapper[4997]: I1205 08:57:49.868068 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zf2jf\" (UniqueName: \"kubernetes.io/projected/b195cd20-21ad-4b84-b5dd-911265d4a823-kube-api-access-zf2jf\") pod \"b195cd20-21ad-4b84-b5dd-911265d4a823\" (UID: \"b195cd20-21ad-4b84-b5dd-911265d4a823\") " Dec 05 08:57:49 crc kubenswrapper[4997]: I1205 08:57:49.872885 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-scripts" (OuterVolumeSpecName: "scripts") pod "b195cd20-21ad-4b84-b5dd-911265d4a823" (UID: "b195cd20-21ad-4b84-b5dd-911265d4a823"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:49 crc kubenswrapper[4997]: I1205 08:57:49.873720 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b195cd20-21ad-4b84-b5dd-911265d4a823-kube-api-access-zf2jf" (OuterVolumeSpecName: "kube-api-access-zf2jf") pod "b195cd20-21ad-4b84-b5dd-911265d4a823" (UID: "b195cd20-21ad-4b84-b5dd-911265d4a823"). InnerVolumeSpecName "kube-api-access-zf2jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:57:49 crc kubenswrapper[4997]: I1205 08:57:49.898567 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-config-data" (OuterVolumeSpecName: "config-data") pod "b195cd20-21ad-4b84-b5dd-911265d4a823" (UID: "b195cd20-21ad-4b84-b5dd-911265d4a823"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:49 crc kubenswrapper[4997]: I1205 08:57:49.900165 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b195cd20-21ad-4b84-b5dd-911265d4a823" (UID: "b195cd20-21ad-4b84-b5dd-911265d4a823"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:49 crc kubenswrapper[4997]: I1205 08:57:49.970685 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:49 crc kubenswrapper[4997]: I1205 08:57:49.970718 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:49 crc kubenswrapper[4997]: I1205 08:57:49.970729 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b195cd20-21ad-4b84-b5dd-911265d4a823-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:49 crc kubenswrapper[4997]: I1205 08:57:49.970738 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zf2jf\" (UniqueName: \"kubernetes.io/projected/b195cd20-21ad-4b84-b5dd-911265d4a823-kube-api-access-zf2jf\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.358274 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6vfwt" event={"ID":"b195cd20-21ad-4b84-b5dd-911265d4a823","Type":"ContainerDied","Data":"1fe552ddb0103d73e734cc57bddbdfedae2a5a20b3b8a9b84d6e0f2324b0c2ee"} Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.358318 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fe552ddb0103d73e734cc57bddbdfedae2a5a20b3b8a9b84d6e0f2324b0c2ee" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.358430 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6vfwt" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.470191 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 08:57:50 crc kubenswrapper[4997]: E1205 08:57:50.473876 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b195cd20-21ad-4b84-b5dd-911265d4a823" containerName="nova-cell1-conductor-db-sync" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.473900 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b195cd20-21ad-4b84-b5dd-911265d4a823" containerName="nova-cell1-conductor-db-sync" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.474087 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="b195cd20-21ad-4b84-b5dd-911265d4a823" containerName="nova-cell1-conductor-db-sync" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.474857 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.478332 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.478599 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.579160 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadd4122-9bc4-4e70-97d4-7e062feed5de-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"cadd4122-9bc4-4e70-97d4-7e062feed5de\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.579243 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadd4122-9bc4-4e70-97d4-7e062feed5de-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"cadd4122-9bc4-4e70-97d4-7e062feed5de\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.579399 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj6zn\" (UniqueName: \"kubernetes.io/projected/cadd4122-9bc4-4e70-97d4-7e062feed5de-kube-api-access-zj6zn\") pod \"nova-cell1-conductor-0\" (UID: \"cadd4122-9bc4-4e70-97d4-7e062feed5de\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.681387 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj6zn\" (UniqueName: \"kubernetes.io/projected/cadd4122-9bc4-4e70-97d4-7e062feed5de-kube-api-access-zj6zn\") pod \"nova-cell1-conductor-0\" (UID: \"cadd4122-9bc4-4e70-97d4-7e062feed5de\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.681506 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadd4122-9bc4-4e70-97d4-7e062feed5de-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"cadd4122-9bc4-4e70-97d4-7e062feed5de\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.681581 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadd4122-9bc4-4e70-97d4-7e062feed5de-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"cadd4122-9bc4-4e70-97d4-7e062feed5de\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.686934 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadd4122-9bc4-4e70-97d4-7e062feed5de-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"cadd4122-9bc4-4e70-97d4-7e062feed5de\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.686976 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadd4122-9bc4-4e70-97d4-7e062feed5de-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"cadd4122-9bc4-4e70-97d4-7e062feed5de\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.705799 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj6zn\" (UniqueName: \"kubernetes.io/projected/cadd4122-9bc4-4e70-97d4-7e062feed5de-kube-api-access-zj6zn\") pod \"nova-cell1-conductor-0\" (UID: \"cadd4122-9bc4-4e70-97d4-7e062feed5de\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.775348 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-n4vrb" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.798880 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.885113 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-scripts\") pod \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\" (UID: \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\") " Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.885522 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-config-data\") pod \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\" (UID: \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\") " Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.885560 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmqcc\" (UniqueName: \"kubernetes.io/projected/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-kube-api-access-dmqcc\") pod \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\" (UID: \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\") " Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.885606 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-combined-ca-bundle\") pod \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\" (UID: \"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3\") " Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.895811 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-scripts" (OuterVolumeSpecName: "scripts") pod "6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3" (UID: "6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.895871 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-kube-api-access-dmqcc" (OuterVolumeSpecName: "kube-api-access-dmqcc") pod "6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3" (UID: "6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3"). InnerVolumeSpecName "kube-api-access-dmqcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.915995 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3" (UID: "6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.918959 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-config-data" (OuterVolumeSpecName: "config-data") pod "6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3" (UID: "6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.988923 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmqcc\" (UniqueName: \"kubernetes.io/projected/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-kube-api-access-dmqcc\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.988963 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.988975 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:50 crc kubenswrapper[4997]: I1205 08:57:50.988988 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:51 crc kubenswrapper[4997]: W1205 08:57:51.225460 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcadd4122_9bc4_4e70_97d4_7e062feed5de.slice/crio-9af4c1582127e78dc66395947ff9cb62c444c1d10f40abb30a04767bef12f8d8 WatchSource:0}: Error finding container 9af4c1582127e78dc66395947ff9cb62c444c1d10f40abb30a04767bef12f8d8: Status 404 returned error can't find the container with id 9af4c1582127e78dc66395947ff9cb62c444c1d10f40abb30a04767bef12f8d8 Dec 05 08:57:51 crc kubenswrapper[4997]: I1205 08:57:51.225687 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 08:57:51 crc kubenswrapper[4997]: I1205 08:57:51.369382 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"cadd4122-9bc4-4e70-97d4-7e062feed5de","Type":"ContainerStarted","Data":"9af4c1582127e78dc66395947ff9cb62c444c1d10f40abb30a04767bef12f8d8"} Dec 05 08:57:51 crc kubenswrapper[4997]: I1205 08:57:51.372158 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-n4vrb" event={"ID":"6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3","Type":"ContainerDied","Data":"5487e3d1d30f7c87006eb06fc3f15c8ae1152bf14c677af9e7d07b1d1d1ce769"} Dec 05 08:57:51 crc kubenswrapper[4997]: I1205 08:57:51.372181 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5487e3d1d30f7c87006eb06fc3f15c8ae1152bf14c677af9e7d07b1d1d1ce769" Dec 05 08:57:51 crc kubenswrapper[4997]: I1205 08:57:51.372224 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-n4vrb" Dec 05 08:57:51 crc kubenswrapper[4997]: I1205 08:57:51.546796 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:57:51 crc kubenswrapper[4997]: I1205 08:57:51.547124 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="b731c82f-0c89-4897-ab96-bf5fe10adbde" containerName="nova-scheduler-scheduler" containerID="cri-o://a65e0964ec99b944af43aadba0c305b732dbe20f3cad10b804dbe225467077f0" gracePeriod=30 Dec 05 08:57:51 crc kubenswrapper[4997]: I1205 08:57:51.556593 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:57:51 crc kubenswrapper[4997]: I1205 08:57:51.556893 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3154810c-312d-4ac4-872f-7becd4fad42c" containerName="nova-api-log" containerID="cri-o://c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4" gracePeriod=30 Dec 05 08:57:51 crc kubenswrapper[4997]: I1205 08:57:51.557409 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3154810c-312d-4ac4-872f-7becd4fad42c" containerName="nova-api-api" containerID="cri-o://a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4" gracePeriod=30 Dec 05 08:57:51 crc kubenswrapper[4997]: I1205 08:57:51.563455 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:57:51 crc kubenswrapper[4997]: I1205 08:57:51.563720 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e2dd2984-e4c8-406c-91a0-b41df0b17b8f" containerName="nova-metadata-log" containerID="cri-o://3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407" gracePeriod=30 Dec 05 08:57:51 crc kubenswrapper[4997]: I1205 08:57:51.563834 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e2dd2984-e4c8-406c-91a0-b41df0b17b8f" containerName="nova-metadata-metadata" containerID="cri-o://efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685" gracePeriod=30 Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.197348 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.205966 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.315845 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdfd9\" (UniqueName: \"kubernetes.io/projected/3154810c-312d-4ac4-872f-7becd4fad42c-kube-api-access-sdfd9\") pod \"3154810c-312d-4ac4-872f-7becd4fad42c\" (UID: \"3154810c-312d-4ac4-872f-7becd4fad42c\") " Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.317139 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3154810c-312d-4ac4-872f-7becd4fad42c-config-data\") pod \"3154810c-312d-4ac4-872f-7becd4fad42c\" (UID: \"3154810c-312d-4ac4-872f-7becd4fad42c\") " Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.317563 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3154810c-312d-4ac4-872f-7becd4fad42c-logs\") pod \"3154810c-312d-4ac4-872f-7becd4fad42c\" (UID: \"3154810c-312d-4ac4-872f-7becd4fad42c\") " Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.317678 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pz4x7\" (UniqueName: \"kubernetes.io/projected/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-kube-api-access-pz4x7\") pod \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\" (UID: \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\") " Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.317785 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-logs\") pod \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\" (UID: \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\") " Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.317884 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3154810c-312d-4ac4-872f-7becd4fad42c-combined-ca-bundle\") pod \"3154810c-312d-4ac4-872f-7becd4fad42c\" (UID: \"3154810c-312d-4ac4-872f-7becd4fad42c\") " Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.318021 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-combined-ca-bundle\") pod \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\" (UID: \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\") " Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.318013 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3154810c-312d-4ac4-872f-7becd4fad42c-logs" (OuterVolumeSpecName: "logs") pod "3154810c-312d-4ac4-872f-7becd4fad42c" (UID: "3154810c-312d-4ac4-872f-7becd4fad42c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.318142 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-config-data\") pod \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\" (UID: \"e2dd2984-e4c8-406c-91a0-b41df0b17b8f\") " Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.318575 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-logs" (OuterVolumeSpecName: "logs") pod "e2dd2984-e4c8-406c-91a0-b41df0b17b8f" (UID: "e2dd2984-e4c8-406c-91a0-b41df0b17b8f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.319146 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3154810c-312d-4ac4-872f-7becd4fad42c-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.319179 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.321843 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-kube-api-access-pz4x7" (OuterVolumeSpecName: "kube-api-access-pz4x7") pod "e2dd2984-e4c8-406c-91a0-b41df0b17b8f" (UID: "e2dd2984-e4c8-406c-91a0-b41df0b17b8f"). InnerVolumeSpecName "kube-api-access-pz4x7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.321986 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3154810c-312d-4ac4-872f-7becd4fad42c-kube-api-access-sdfd9" (OuterVolumeSpecName: "kube-api-access-sdfd9") pod "3154810c-312d-4ac4-872f-7becd4fad42c" (UID: "3154810c-312d-4ac4-872f-7becd4fad42c"). InnerVolumeSpecName "kube-api-access-sdfd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.341134 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3154810c-312d-4ac4-872f-7becd4fad42c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3154810c-312d-4ac4-872f-7becd4fad42c" (UID: "3154810c-312d-4ac4-872f-7becd4fad42c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.341413 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3154810c-312d-4ac4-872f-7becd4fad42c-config-data" (OuterVolumeSpecName: "config-data") pod "3154810c-312d-4ac4-872f-7becd4fad42c" (UID: "3154810c-312d-4ac4-872f-7becd4fad42c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.347172 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2dd2984-e4c8-406c-91a0-b41df0b17b8f" (UID: "e2dd2984-e4c8-406c-91a0-b41df0b17b8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.350813 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-config-data" (OuterVolumeSpecName: "config-data") pod "e2dd2984-e4c8-406c-91a0-b41df0b17b8f" (UID: "e2dd2984-e4c8-406c-91a0-b41df0b17b8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.382522 4997 generic.go:334] "Generic (PLEG): container finished" podID="3154810c-312d-4ac4-872f-7becd4fad42c" containerID="a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4" exitCode=0 Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.383501 4997 generic.go:334] "Generic (PLEG): container finished" podID="3154810c-312d-4ac4-872f-7becd4fad42c" containerID="c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4" exitCode=143 Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.383654 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3154810c-312d-4ac4-872f-7becd4fad42c","Type":"ContainerDied","Data":"a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4"} Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.383759 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3154810c-312d-4ac4-872f-7becd4fad42c","Type":"ContainerDied","Data":"c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4"} Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.383839 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3154810c-312d-4ac4-872f-7becd4fad42c","Type":"ContainerDied","Data":"fcb0f0afbae2301f5c2042a77b29abf1a31c260214b623f40b35aa4ce79498b7"} Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.383926 4997 scope.go:117] "RemoveContainer" containerID="a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.384192 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.390017 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"cadd4122-9bc4-4e70-97d4-7e062feed5de","Type":"ContainerStarted","Data":"45f674f05294bc21c57daee4316ef766b40f5822bb721afd5acc8925ff46098d"} Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.390511 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.393190 4997 generic.go:334] "Generic (PLEG): container finished" podID="e2dd2984-e4c8-406c-91a0-b41df0b17b8f" containerID="efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685" exitCode=0 Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.393228 4997 generic.go:334] "Generic (PLEG): container finished" podID="e2dd2984-e4c8-406c-91a0-b41df0b17b8f" containerID="3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407" exitCode=143 Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.393276 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e2dd2984-e4c8-406c-91a0-b41df0b17b8f","Type":"ContainerDied","Data":"efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685"} Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.393307 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e2dd2984-e4c8-406c-91a0-b41df0b17b8f","Type":"ContainerDied","Data":"3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407"} Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.393321 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e2dd2984-e4c8-406c-91a0-b41df0b17b8f","Type":"ContainerDied","Data":"85d9ed6f3a21e2b45253121ab9e0eb357db2bad939658f3829a92a9e499dc737"} Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.393384 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.415428 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.415413703 podStartE2EDuration="2.415413703s" podCreationTimestamp="2025-12-05 08:57:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:57:52.407355137 +0000 UTC m=+7372.936262418" watchObservedRunningTime="2025-12-05 08:57:52.415413703 +0000 UTC m=+7372.944320964" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.417836 4997 scope.go:117] "RemoveContainer" containerID="c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.437077 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdfd9\" (UniqueName: \"kubernetes.io/projected/3154810c-312d-4ac4-872f-7becd4fad42c-kube-api-access-sdfd9\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.437114 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3154810c-312d-4ac4-872f-7becd4fad42c-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.437128 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pz4x7\" (UniqueName: \"kubernetes.io/projected/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-kube-api-access-pz4x7\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.437148 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3154810c-312d-4ac4-872f-7becd4fad42c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.437162 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.437173 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2dd2984-e4c8-406c-91a0-b41df0b17b8f-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.460207 4997 scope.go:117] "RemoveContainer" containerID="a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4" Dec 05 08:57:52 crc kubenswrapper[4997]: E1205 08:57:52.460980 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4\": container with ID starting with a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4 not found: ID does not exist" containerID="a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.462719 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4"} err="failed to get container status \"a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4\": rpc error: code = NotFound desc = could not find container \"a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4\": container with ID starting with a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4 not found: ID does not exist" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.462830 4997 scope.go:117] "RemoveContainer" containerID="c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4" Dec 05 08:57:52 crc kubenswrapper[4997]: E1205 08:57:52.465135 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4\": container with ID starting with c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4 not found: ID does not exist" containerID="c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.465205 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4"} err="failed to get container status \"c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4\": rpc error: code = NotFound desc = could not find container \"c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4\": container with ID starting with c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4 not found: ID does not exist" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.465235 4997 scope.go:117] "RemoveContainer" containerID="a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.468974 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4"} err="failed to get container status \"a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4\": rpc error: code = NotFound desc = could not find container \"a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4\": container with ID starting with a039de8517de4e1f96a337628d7d3dbc73009b94383bda92cf74933c489fe9c4 not found: ID does not exist" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.469186 4997 scope.go:117] "RemoveContainer" containerID="c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.470279 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4"} err="failed to get container status \"c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4\": rpc error: code = NotFound desc = could not find container \"c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4\": container with ID starting with c3033e636d1d1ac5907f3ee64a21fbdf9488a8f9f7e1d0b212c799726beaebf4 not found: ID does not exist" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.471480 4997 scope.go:117] "RemoveContainer" containerID="efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.473843 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.490408 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.505557 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.513633 4997 scope.go:117] "RemoveContainer" containerID="3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.519225 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.528789 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 08:57:52 crc kubenswrapper[4997]: E1205 08:57:52.529259 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3" containerName="nova-manage" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.529280 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3" containerName="nova-manage" Dec 05 08:57:52 crc kubenswrapper[4997]: E1205 08:57:52.529296 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3154810c-312d-4ac4-872f-7becd4fad42c" containerName="nova-api-api" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.529304 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3154810c-312d-4ac4-872f-7becd4fad42c" containerName="nova-api-api" Dec 05 08:57:52 crc kubenswrapper[4997]: E1205 08:57:52.529323 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2dd2984-e4c8-406c-91a0-b41df0b17b8f" containerName="nova-metadata-metadata" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.529332 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2dd2984-e4c8-406c-91a0-b41df0b17b8f" containerName="nova-metadata-metadata" Dec 05 08:57:52 crc kubenswrapper[4997]: E1205 08:57:52.529342 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2dd2984-e4c8-406c-91a0-b41df0b17b8f" containerName="nova-metadata-log" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.529352 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2dd2984-e4c8-406c-91a0-b41df0b17b8f" containerName="nova-metadata-log" Dec 05 08:57:52 crc kubenswrapper[4997]: E1205 08:57:52.529372 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3154810c-312d-4ac4-872f-7becd4fad42c" containerName="nova-api-log" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.529379 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3154810c-312d-4ac4-872f-7becd4fad42c" containerName="nova-api-log" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.529640 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="3154810c-312d-4ac4-872f-7becd4fad42c" containerName="nova-api-api" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.529656 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2dd2984-e4c8-406c-91a0-b41df0b17b8f" containerName="nova-metadata-metadata" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.529672 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3" containerName="nova-manage" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.529697 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="3154810c-312d-4ac4-872f-7becd4fad42c" containerName="nova-api-log" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.529707 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2dd2984-e4c8-406c-91a0-b41df0b17b8f" containerName="nova-metadata-log" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.531004 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.533347 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.537706 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ddtzs" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.538223 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ddtzs" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.540068 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.542552 4997 scope.go:117] "RemoveContainer" containerID="efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685" Dec 05 08:57:52 crc kubenswrapper[4997]: E1205 08:57:52.543023 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685\": container with ID starting with efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685 not found: ID does not exist" containerID="efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.543054 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685"} err="failed to get container status \"efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685\": rpc error: code = NotFound desc = could not find container \"efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685\": container with ID starting with efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685 not found: ID does not exist" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.543072 4997 scope.go:117] "RemoveContainer" containerID="3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407" Dec 05 08:57:52 crc kubenswrapper[4997]: E1205 08:57:52.549922 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407\": container with ID starting with 3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407 not found: ID does not exist" containerID="3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.549984 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407"} err="failed to get container status \"3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407\": rpc error: code = NotFound desc = could not find container \"3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407\": container with ID starting with 3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407 not found: ID does not exist" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.550015 4997 scope.go:117] "RemoveContainer" containerID="efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.551216 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685"} err="failed to get container status \"efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685\": rpc error: code = NotFound desc = could not find container \"efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685\": container with ID starting with efc51b09cde222896fd712820488cd9793134622d92b79bf305850bfccbf7685 not found: ID does not exist" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.551316 4997 scope.go:117] "RemoveContainer" containerID="3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.551907 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.552246 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407"} err="failed to get container status \"3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407\": rpc error: code = NotFound desc = could not find container \"3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407\": container with ID starting with 3a764b866cbe27f8df5172f92d526286618a22a5e7f35069fa26657abcfcd407 not found: ID does not exist" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.553938 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.556575 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.565723 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.589848 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ddtzs" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.640599 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2698529d-c31d-43eb-940f-5eee6049968f-config-data\") pod \"nova-metadata-0\" (UID: \"2698529d-c31d-43eb-940f-5eee6049968f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.640706 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2698529d-c31d-43eb-940f-5eee6049968f-logs\") pod \"nova-metadata-0\" (UID: \"2698529d-c31d-43eb-940f-5eee6049968f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.640761 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10c752a8-8c4f-42a4-9497-b4ae222788bc-config-data\") pod \"nova-api-0\" (UID: \"10c752a8-8c4f-42a4-9497-b4ae222788bc\") " pod="openstack/nova-api-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.640781 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vjm8\" (UniqueName: \"kubernetes.io/projected/10c752a8-8c4f-42a4-9497-b4ae222788bc-kube-api-access-5vjm8\") pod \"nova-api-0\" (UID: \"10c752a8-8c4f-42a4-9497-b4ae222788bc\") " pod="openstack/nova-api-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.640807 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2698529d-c31d-43eb-940f-5eee6049968f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2698529d-c31d-43eb-940f-5eee6049968f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.640833 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hft4l\" (UniqueName: \"kubernetes.io/projected/2698529d-c31d-43eb-940f-5eee6049968f-kube-api-access-hft4l\") pod \"nova-metadata-0\" (UID: \"2698529d-c31d-43eb-940f-5eee6049968f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.640877 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10c752a8-8c4f-42a4-9497-b4ae222788bc-logs\") pod \"nova-api-0\" (UID: \"10c752a8-8c4f-42a4-9497-b4ae222788bc\") " pod="openstack/nova-api-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.640934 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c752a8-8c4f-42a4-9497-b4ae222788bc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"10c752a8-8c4f-42a4-9497-b4ae222788bc\") " pod="openstack/nova-api-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.742700 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10c752a8-8c4f-42a4-9497-b4ae222788bc-config-data\") pod \"nova-api-0\" (UID: \"10c752a8-8c4f-42a4-9497-b4ae222788bc\") " pod="openstack/nova-api-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.742960 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vjm8\" (UniqueName: \"kubernetes.io/projected/10c752a8-8c4f-42a4-9497-b4ae222788bc-kube-api-access-5vjm8\") pod \"nova-api-0\" (UID: \"10c752a8-8c4f-42a4-9497-b4ae222788bc\") " pod="openstack/nova-api-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.742985 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2698529d-c31d-43eb-940f-5eee6049968f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2698529d-c31d-43eb-940f-5eee6049968f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.743005 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hft4l\" (UniqueName: \"kubernetes.io/projected/2698529d-c31d-43eb-940f-5eee6049968f-kube-api-access-hft4l\") pod \"nova-metadata-0\" (UID: \"2698529d-c31d-43eb-940f-5eee6049968f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.743049 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10c752a8-8c4f-42a4-9497-b4ae222788bc-logs\") pod \"nova-api-0\" (UID: \"10c752a8-8c4f-42a4-9497-b4ae222788bc\") " pod="openstack/nova-api-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.743079 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c752a8-8c4f-42a4-9497-b4ae222788bc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"10c752a8-8c4f-42a4-9497-b4ae222788bc\") " pod="openstack/nova-api-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.743171 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2698529d-c31d-43eb-940f-5eee6049968f-config-data\") pod \"nova-metadata-0\" (UID: \"2698529d-c31d-43eb-940f-5eee6049968f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.743204 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2698529d-c31d-43eb-940f-5eee6049968f-logs\") pod \"nova-metadata-0\" (UID: \"2698529d-c31d-43eb-940f-5eee6049968f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.743638 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2698529d-c31d-43eb-940f-5eee6049968f-logs\") pod \"nova-metadata-0\" (UID: \"2698529d-c31d-43eb-940f-5eee6049968f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.744582 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10c752a8-8c4f-42a4-9497-b4ae222788bc-logs\") pod \"nova-api-0\" (UID: \"10c752a8-8c4f-42a4-9497-b4ae222788bc\") " pod="openstack/nova-api-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.763752 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10c752a8-8c4f-42a4-9497-b4ae222788bc-config-data\") pod \"nova-api-0\" (UID: \"10c752a8-8c4f-42a4-9497-b4ae222788bc\") " pod="openstack/nova-api-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.764393 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c752a8-8c4f-42a4-9497-b4ae222788bc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"10c752a8-8c4f-42a4-9497-b4ae222788bc\") " pod="openstack/nova-api-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.766241 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2698529d-c31d-43eb-940f-5eee6049968f-config-data\") pod \"nova-metadata-0\" (UID: \"2698529d-c31d-43eb-940f-5eee6049968f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.766554 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2698529d-c31d-43eb-940f-5eee6049968f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2698529d-c31d-43eb-940f-5eee6049968f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.769813 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hft4l\" (UniqueName: \"kubernetes.io/projected/2698529d-c31d-43eb-940f-5eee6049968f-kube-api-access-hft4l\") pod \"nova-metadata-0\" (UID: \"2698529d-c31d-43eb-940f-5eee6049968f\") " pod="openstack/nova-metadata-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.771452 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vjm8\" (UniqueName: \"kubernetes.io/projected/10c752a8-8c4f-42a4-9497-b4ae222788bc-kube-api-access-5vjm8\") pod \"nova-api-0\" (UID: \"10c752a8-8c4f-42a4-9497-b4ae222788bc\") " pod="openstack/nova-api-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.861446 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:57:52 crc kubenswrapper[4997]: I1205 08:57:52.877132 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:57:53 crc kubenswrapper[4997]: I1205 08:57:53.337446 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:57:53 crc kubenswrapper[4997]: W1205 08:57:53.338198 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10c752a8_8c4f_42a4_9497_b4ae222788bc.slice/crio-1e4646703b00cbd3681ab5f508581a7de679e65b9905b81da5a8ba63e0a7d3a1 WatchSource:0}: Error finding container 1e4646703b00cbd3681ab5f508581a7de679e65b9905b81da5a8ba63e0a7d3a1: Status 404 returned error can't find the container with id 1e4646703b00cbd3681ab5f508581a7de679e65b9905b81da5a8ba63e0a7d3a1 Dec 05 08:57:53 crc kubenswrapper[4997]: I1205 08:57:53.404525 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"10c752a8-8c4f-42a4-9497-b4ae222788bc","Type":"ContainerStarted","Data":"1e4646703b00cbd3681ab5f508581a7de679e65b9905b81da5a8ba63e0a7d3a1"} Dec 05 08:57:53 crc kubenswrapper[4997]: I1205 08:57:53.405899 4997 generic.go:334] "Generic (PLEG): container finished" podID="b731c82f-0c89-4897-ab96-bf5fe10adbde" containerID="a65e0964ec99b944af43aadba0c305b732dbe20f3cad10b804dbe225467077f0" exitCode=0 Dec 05 08:57:53 crc kubenswrapper[4997]: I1205 08:57:53.405991 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b731c82f-0c89-4897-ab96-bf5fe10adbde","Type":"ContainerDied","Data":"a65e0964ec99b944af43aadba0c305b732dbe20f3cad10b804dbe225467077f0"} Dec 05 08:57:53 crc kubenswrapper[4997]: I1205 08:57:53.446835 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:57:53 crc kubenswrapper[4997]: I1205 08:57:53.463408 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ddtzs" Dec 05 08:57:53 crc kubenswrapper[4997]: I1205 08:57:53.518771 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ddtzs"] Dec 05 08:57:53 crc kubenswrapper[4997]: I1205 08:57:53.760885 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3154810c-312d-4ac4-872f-7becd4fad42c" path="/var/lib/kubelet/pods/3154810c-312d-4ac4-872f-7becd4fad42c/volumes" Dec 05 08:57:53 crc kubenswrapper[4997]: I1205 08:57:53.761729 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2dd2984-e4c8-406c-91a0-b41df0b17b8f" path="/var/lib/kubelet/pods/e2dd2984-e4c8-406c-91a0-b41df0b17b8f/volumes" Dec 05 08:57:53 crc kubenswrapper[4997]: I1205 08:57:53.884994 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:57:53 crc kubenswrapper[4997]: I1205 08:57:53.898905 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:57:53 crc kubenswrapper[4997]: I1205 08:57:53.903160 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:57:53 crc kubenswrapper[4997]: I1205 08:57:53.964017 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d66fd5885-nb4nr"] Dec 05 08:57:53 crc kubenswrapper[4997]: I1205 08:57:53.964258 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" podUID="d9d96a8f-1144-4680-952a-37f0580bfec0" containerName="dnsmasq-dns" containerID="cri-o://392ec1087f285d4ec9ff0f0f251b0edb60f288b2189931b51a95c7c4e534f891" gracePeriod=10 Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.190713 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.272998 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdmz2\" (UniqueName: \"kubernetes.io/projected/b731c82f-0c89-4897-ab96-bf5fe10adbde-kube-api-access-tdmz2\") pod \"b731c82f-0c89-4897-ab96-bf5fe10adbde\" (UID: \"b731c82f-0c89-4897-ab96-bf5fe10adbde\") " Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.273086 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b731c82f-0c89-4897-ab96-bf5fe10adbde-config-data\") pod \"b731c82f-0c89-4897-ab96-bf5fe10adbde\" (UID: \"b731c82f-0c89-4897-ab96-bf5fe10adbde\") " Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.273166 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b731c82f-0c89-4897-ab96-bf5fe10adbde-combined-ca-bundle\") pod \"b731c82f-0c89-4897-ab96-bf5fe10adbde\" (UID: \"b731c82f-0c89-4897-ab96-bf5fe10adbde\") " Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.281351 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b731c82f-0c89-4897-ab96-bf5fe10adbde-kube-api-access-tdmz2" (OuterVolumeSpecName: "kube-api-access-tdmz2") pod "b731c82f-0c89-4897-ab96-bf5fe10adbde" (UID: "b731c82f-0c89-4897-ab96-bf5fe10adbde"). InnerVolumeSpecName "kube-api-access-tdmz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.328732 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b731c82f-0c89-4897-ab96-bf5fe10adbde-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b731c82f-0c89-4897-ab96-bf5fe10adbde" (UID: "b731c82f-0c89-4897-ab96-bf5fe10adbde"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.328698 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b731c82f-0c89-4897-ab96-bf5fe10adbde-config-data" (OuterVolumeSpecName: "config-data") pod "b731c82f-0c89-4897-ab96-bf5fe10adbde" (UID: "b731c82f-0c89-4897-ab96-bf5fe10adbde"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.375154 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdmz2\" (UniqueName: \"kubernetes.io/projected/b731c82f-0c89-4897-ab96-bf5fe10adbde-kube-api-access-tdmz2\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.375189 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b731c82f-0c89-4897-ab96-bf5fe10adbde-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.375204 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b731c82f-0c89-4897-ab96-bf5fe10adbde-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.375366 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.423001 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2698529d-c31d-43eb-940f-5eee6049968f","Type":"ContainerStarted","Data":"058a98233594c255f126f64174ad011eb820878514c62f60f050944f04f7d5ab"} Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.423040 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2698529d-c31d-43eb-940f-5eee6049968f","Type":"ContainerStarted","Data":"876fcd471b6cc44ddb752682f0fb575b0ef56ef63e5cd351cd4d109e063a22e8"} Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.423051 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2698529d-c31d-43eb-940f-5eee6049968f","Type":"ContainerStarted","Data":"4ceec146f9f58d1b208289caf6b55c0a65a15fb2cc5bd79417a78d52d032ca24"} Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.425580 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b731c82f-0c89-4897-ab96-bf5fe10adbde","Type":"ContainerDied","Data":"4c8cdc3c7f0d42ec4adfa2e042179ebb4d28b5e08b674d0e3311725a3f6ea367"} Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.425636 4997 scope.go:117] "RemoveContainer" containerID="a65e0964ec99b944af43aadba0c305b732dbe20f3cad10b804dbe225467077f0" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.425704 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.429159 4997 generic.go:334] "Generic (PLEG): container finished" podID="d9d96a8f-1144-4680-952a-37f0580bfec0" containerID="392ec1087f285d4ec9ff0f0f251b0edb60f288b2189931b51a95c7c4e534f891" exitCode=0 Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.429223 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" event={"ID":"d9d96a8f-1144-4680-952a-37f0580bfec0","Type":"ContainerDied","Data":"392ec1087f285d4ec9ff0f0f251b0edb60f288b2189931b51a95c7c4e534f891"} Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.429256 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" event={"ID":"d9d96a8f-1144-4680-952a-37f0580bfec0","Type":"ContainerDied","Data":"03f42fff35f42b8a91a40558d6ff22427bf83037b938383342863bcdd2ba74e0"} Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.429328 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d66fd5885-nb4nr" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.438343 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"10c752a8-8c4f-42a4-9497-b4ae222788bc","Type":"ContainerStarted","Data":"dae181f36626243af1e9ea6ee27a6717e88cee749a5953bd62afe49a19a125ef"} Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.438403 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"10c752a8-8c4f-42a4-9497-b4ae222788bc","Type":"ContainerStarted","Data":"4fc9327f8eb3e195e86413bdd836dc2c17547f019bf0ff23334c605fd02a7948"} Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.452327 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.452293046 podStartE2EDuration="2.452293046s" podCreationTimestamp="2025-12-05 08:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:57:54.441093305 +0000 UTC m=+7374.970000566" watchObservedRunningTime="2025-12-05 08:57:54.452293046 +0000 UTC m=+7374.981200307" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.453718 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.458151 4997 scope.go:117] "RemoveContainer" containerID="392ec1087f285d4ec9ff0f0f251b0edb60f288b2189931b51a95c7c4e534f891" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.466970 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.466944041 podStartE2EDuration="2.466944041s" podCreationTimestamp="2025-12-05 08:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:57:54.464983738 +0000 UTC m=+7374.993890999" watchObservedRunningTime="2025-12-05 08:57:54.466944041 +0000 UTC m=+7374.995851302" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.475862 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-ovsdbserver-nb\") pod \"d9d96a8f-1144-4680-952a-37f0580bfec0\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.475994 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzddl\" (UniqueName: \"kubernetes.io/projected/d9d96a8f-1144-4680-952a-37f0580bfec0-kube-api-access-wzddl\") pod \"d9d96a8f-1144-4680-952a-37f0580bfec0\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.476124 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-config\") pod \"d9d96a8f-1144-4680-952a-37f0580bfec0\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.476283 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-ovsdbserver-sb\") pod \"d9d96a8f-1144-4680-952a-37f0580bfec0\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.476314 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-dns-svc\") pod \"d9d96a8f-1144-4680-952a-37f0580bfec0\" (UID: \"d9d96a8f-1144-4680-952a-37f0580bfec0\") " Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.480061 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9d96a8f-1144-4680-952a-37f0580bfec0-kube-api-access-wzddl" (OuterVolumeSpecName: "kube-api-access-wzddl") pod "d9d96a8f-1144-4680-952a-37f0580bfec0" (UID: "d9d96a8f-1144-4680-952a-37f0580bfec0"). InnerVolumeSpecName "kube-api-access-wzddl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.483554 4997 scope.go:117] "RemoveContainer" containerID="d79d83516589ac4159831c7598ca2713d4b370dd4a870b8f97f91f9777d638c3" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.524996 4997 scope.go:117] "RemoveContainer" containerID="392ec1087f285d4ec9ff0f0f251b0edb60f288b2189931b51a95c7c4e534f891" Dec 05 08:57:54 crc kubenswrapper[4997]: E1205 08:57:54.528521 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"392ec1087f285d4ec9ff0f0f251b0edb60f288b2189931b51a95c7c4e534f891\": container with ID starting with 392ec1087f285d4ec9ff0f0f251b0edb60f288b2189931b51a95c7c4e534f891 not found: ID does not exist" containerID="392ec1087f285d4ec9ff0f0f251b0edb60f288b2189931b51a95c7c4e534f891" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.528588 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"392ec1087f285d4ec9ff0f0f251b0edb60f288b2189931b51a95c7c4e534f891"} err="failed to get container status \"392ec1087f285d4ec9ff0f0f251b0edb60f288b2189931b51a95c7c4e534f891\": rpc error: code = NotFound desc = could not find container \"392ec1087f285d4ec9ff0f0f251b0edb60f288b2189931b51a95c7c4e534f891\": container with ID starting with 392ec1087f285d4ec9ff0f0f251b0edb60f288b2189931b51a95c7c4e534f891 not found: ID does not exist" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.528644 4997 scope.go:117] "RemoveContainer" containerID="d79d83516589ac4159831c7598ca2713d4b370dd4a870b8f97f91f9777d638c3" Dec 05 08:57:54 crc kubenswrapper[4997]: E1205 08:57:54.532085 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d79d83516589ac4159831c7598ca2713d4b370dd4a870b8f97f91f9777d638c3\": container with ID starting with d79d83516589ac4159831c7598ca2713d4b370dd4a870b8f97f91f9777d638c3 not found: ID does not exist" containerID="d79d83516589ac4159831c7598ca2713d4b370dd4a870b8f97f91f9777d638c3" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.532150 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d79d83516589ac4159831c7598ca2713d4b370dd4a870b8f97f91f9777d638c3"} err="failed to get container status \"d79d83516589ac4159831c7598ca2713d4b370dd4a870b8f97f91f9777d638c3\": rpc error: code = NotFound desc = could not find container \"d79d83516589ac4159831c7598ca2713d4b370dd4a870b8f97f91f9777d638c3\": container with ID starting with d79d83516589ac4159831c7598ca2713d4b370dd4a870b8f97f91f9777d638c3 not found: ID does not exist" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.575343 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d9d96a8f-1144-4680-952a-37f0580bfec0" (UID: "d9d96a8f-1144-4680-952a-37f0580bfec0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.584042 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzddl\" (UniqueName: \"kubernetes.io/projected/d9d96a8f-1144-4680-952a-37f0580bfec0-kube-api-access-wzddl\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.584084 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.600311 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.603664 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d9d96a8f-1144-4680-952a-37f0580bfec0" (UID: "d9d96a8f-1144-4680-952a-37f0580bfec0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.612194 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.619983 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:57:54 crc kubenswrapper[4997]: E1205 08:57:54.620489 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d96a8f-1144-4680-952a-37f0580bfec0" containerName="dnsmasq-dns" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.620512 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d96a8f-1144-4680-952a-37f0580bfec0" containerName="dnsmasq-dns" Dec 05 08:57:54 crc kubenswrapper[4997]: E1205 08:57:54.620531 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d96a8f-1144-4680-952a-37f0580bfec0" containerName="init" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.620539 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d96a8f-1144-4680-952a-37f0580bfec0" containerName="init" Dec 05 08:57:54 crc kubenswrapper[4997]: E1205 08:57:54.620566 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b731c82f-0c89-4897-ab96-bf5fe10adbde" containerName="nova-scheduler-scheduler" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.620573 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b731c82f-0c89-4897-ab96-bf5fe10adbde" containerName="nova-scheduler-scheduler" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.620998 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9d96a8f-1144-4680-952a-37f0580bfec0" containerName="dnsmasq-dns" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.621026 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="b731c82f-0c89-4897-ab96-bf5fe10adbde" containerName="nova-scheduler-scheduler" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.622152 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.623954 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.648227 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.658162 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d9d96a8f-1144-4680-952a-37f0580bfec0" (UID: "d9d96a8f-1144-4680-952a-37f0580bfec0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.676943 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-config" (OuterVolumeSpecName: "config") pod "d9d96a8f-1144-4680-952a-37f0580bfec0" (UID: "d9d96a8f-1144-4680-952a-37f0580bfec0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.685306 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.685335 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.685344 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9d96a8f-1144-4680-952a-37f0580bfec0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.786298 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/475e068f-94be-4d18-a55d-0e1940748dfb-config-data\") pod \"nova-scheduler-0\" (UID: \"475e068f-94be-4d18-a55d-0e1940748dfb\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.786426 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjk6z\" (UniqueName: \"kubernetes.io/projected/475e068f-94be-4d18-a55d-0e1940748dfb-kube-api-access-xjk6z\") pod \"nova-scheduler-0\" (UID: \"475e068f-94be-4d18-a55d-0e1940748dfb\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.786493 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/475e068f-94be-4d18-a55d-0e1940748dfb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"475e068f-94be-4d18-a55d-0e1940748dfb\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.789900 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d66fd5885-nb4nr"] Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.797126 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d66fd5885-nb4nr"] Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.887757 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/475e068f-94be-4d18-a55d-0e1940748dfb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"475e068f-94be-4d18-a55d-0e1940748dfb\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.887909 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/475e068f-94be-4d18-a55d-0e1940748dfb-config-data\") pod \"nova-scheduler-0\" (UID: \"475e068f-94be-4d18-a55d-0e1940748dfb\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.887940 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjk6z\" (UniqueName: \"kubernetes.io/projected/475e068f-94be-4d18-a55d-0e1940748dfb-kube-api-access-xjk6z\") pod \"nova-scheduler-0\" (UID: \"475e068f-94be-4d18-a55d-0e1940748dfb\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.893180 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/475e068f-94be-4d18-a55d-0e1940748dfb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"475e068f-94be-4d18-a55d-0e1940748dfb\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.893680 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/475e068f-94be-4d18-a55d-0e1940748dfb-config-data\") pod \"nova-scheduler-0\" (UID: \"475e068f-94be-4d18-a55d-0e1940748dfb\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.905765 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjk6z\" (UniqueName: \"kubernetes.io/projected/475e068f-94be-4d18-a55d-0e1940748dfb-kube-api-access-xjk6z\") pod \"nova-scheduler-0\" (UID: \"475e068f-94be-4d18-a55d-0e1940748dfb\") " pod="openstack/nova-scheduler-0" Dec 05 08:57:54 crc kubenswrapper[4997]: I1205 08:57:54.949465 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:57:55 crc kubenswrapper[4997]: I1205 08:57:55.388759 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:57:55 crc kubenswrapper[4997]: I1205 08:57:55.448453 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"475e068f-94be-4d18-a55d-0e1940748dfb","Type":"ContainerStarted","Data":"73bcaa581d3d0e1953a8bc1b281e5b169effa5a7a1070582ac7be71970f1290c"} Dec 05 08:57:55 crc kubenswrapper[4997]: I1205 08:57:55.450911 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ddtzs" podUID="bb204e34-3079-463a-833b-fcd2e5d2dd29" containerName="registry-server" containerID="cri-o://3d4ec069a69e22ac7bdab98e76e5b8d29b174359374d39b6d056c58745188140" gracePeriod=2 Dec 05 08:57:55 crc kubenswrapper[4997]: I1205 08:57:55.761144 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b731c82f-0c89-4897-ab96-bf5fe10adbde" path="/var/lib/kubelet/pods/b731c82f-0c89-4897-ab96-bf5fe10adbde/volumes" Dec 05 08:57:55 crc kubenswrapper[4997]: I1205 08:57:55.762496 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9d96a8f-1144-4680-952a-37f0580bfec0" path="/var/lib/kubelet/pods/d9d96a8f-1144-4680-952a-37f0580bfec0/volumes" Dec 05 08:57:55 crc kubenswrapper[4997]: I1205 08:57:55.907109 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ddtzs" Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.015415 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb204e34-3079-463a-833b-fcd2e5d2dd29-catalog-content\") pod \"bb204e34-3079-463a-833b-fcd2e5d2dd29\" (UID: \"bb204e34-3079-463a-833b-fcd2e5d2dd29\") " Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.015475 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hsng\" (UniqueName: \"kubernetes.io/projected/bb204e34-3079-463a-833b-fcd2e5d2dd29-kube-api-access-4hsng\") pod \"bb204e34-3079-463a-833b-fcd2e5d2dd29\" (UID: \"bb204e34-3079-463a-833b-fcd2e5d2dd29\") " Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.015496 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb204e34-3079-463a-833b-fcd2e5d2dd29-utilities\") pod \"bb204e34-3079-463a-833b-fcd2e5d2dd29\" (UID: \"bb204e34-3079-463a-833b-fcd2e5d2dd29\") " Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.015995 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb204e34-3079-463a-833b-fcd2e5d2dd29-utilities" (OuterVolumeSpecName: "utilities") pod "bb204e34-3079-463a-833b-fcd2e5d2dd29" (UID: "bb204e34-3079-463a-833b-fcd2e5d2dd29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.016449 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb204e34-3079-463a-833b-fcd2e5d2dd29-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.018939 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb204e34-3079-463a-833b-fcd2e5d2dd29-kube-api-access-4hsng" (OuterVolumeSpecName: "kube-api-access-4hsng") pod "bb204e34-3079-463a-833b-fcd2e5d2dd29" (UID: "bb204e34-3079-463a-833b-fcd2e5d2dd29"). InnerVolumeSpecName "kube-api-access-4hsng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.041520 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb204e34-3079-463a-833b-fcd2e5d2dd29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb204e34-3079-463a-833b-fcd2e5d2dd29" (UID: "bb204e34-3079-463a-833b-fcd2e5d2dd29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.118021 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb204e34-3079-463a-833b-fcd2e5d2dd29-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.118111 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hsng\" (UniqueName: \"kubernetes.io/projected/bb204e34-3079-463a-833b-fcd2e5d2dd29-kube-api-access-4hsng\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.465828 4997 generic.go:334] "Generic (PLEG): container finished" podID="bb204e34-3079-463a-833b-fcd2e5d2dd29" containerID="3d4ec069a69e22ac7bdab98e76e5b8d29b174359374d39b6d056c58745188140" exitCode=0 Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.466153 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddtzs" event={"ID":"bb204e34-3079-463a-833b-fcd2e5d2dd29","Type":"ContainerDied","Data":"3d4ec069a69e22ac7bdab98e76e5b8d29b174359374d39b6d056c58745188140"} Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.466181 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddtzs" event={"ID":"bb204e34-3079-463a-833b-fcd2e5d2dd29","Type":"ContainerDied","Data":"b3a80eb07f1f3c03b44fe3c090c6a96ed779d5c5858a0950ac9dbedfaaecd76f"} Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.466201 4997 scope.go:117] "RemoveContainer" containerID="3d4ec069a69e22ac7bdab98e76e5b8d29b174359374d39b6d056c58745188140" Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.466316 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ddtzs" Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.470324 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"475e068f-94be-4d18-a55d-0e1940748dfb","Type":"ContainerStarted","Data":"dca43fc19b1f57dc4e660c95619d128f26c8ff15bbab861870f742c4a77bcfa4"} Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.497100 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.497079492 podStartE2EDuration="2.497079492s" podCreationTimestamp="2025-12-05 08:57:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:57:56.488334376 +0000 UTC m=+7377.017241647" watchObservedRunningTime="2025-12-05 08:57:56.497079492 +0000 UTC m=+7377.025986753" Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.511367 4997 scope.go:117] "RemoveContainer" containerID="5de72f4aa8c1f52947b1965ec66a5e532f6d3c3355fef628f5b9b7aeae7ab210" Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.512182 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ddtzs"] Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.523960 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ddtzs"] Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.541064 4997 scope.go:117] "RemoveContainer" containerID="9e8cf71b7f0898961ed6b3b9f800f532557dbab642ccdb154c3862a8e2e2d2cc" Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.592655 4997 scope.go:117] "RemoveContainer" containerID="3d4ec069a69e22ac7bdab98e76e5b8d29b174359374d39b6d056c58745188140" Dec 05 08:57:56 crc kubenswrapper[4997]: E1205 08:57:56.593163 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d4ec069a69e22ac7bdab98e76e5b8d29b174359374d39b6d056c58745188140\": container with ID starting with 3d4ec069a69e22ac7bdab98e76e5b8d29b174359374d39b6d056c58745188140 not found: ID does not exist" containerID="3d4ec069a69e22ac7bdab98e76e5b8d29b174359374d39b6d056c58745188140" Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.593219 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d4ec069a69e22ac7bdab98e76e5b8d29b174359374d39b6d056c58745188140"} err="failed to get container status \"3d4ec069a69e22ac7bdab98e76e5b8d29b174359374d39b6d056c58745188140\": rpc error: code = NotFound desc = could not find container \"3d4ec069a69e22ac7bdab98e76e5b8d29b174359374d39b6d056c58745188140\": container with ID starting with 3d4ec069a69e22ac7bdab98e76e5b8d29b174359374d39b6d056c58745188140 not found: ID does not exist" Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.593253 4997 scope.go:117] "RemoveContainer" containerID="5de72f4aa8c1f52947b1965ec66a5e532f6d3c3355fef628f5b9b7aeae7ab210" Dec 05 08:57:56 crc kubenswrapper[4997]: E1205 08:57:56.593974 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5de72f4aa8c1f52947b1965ec66a5e532f6d3c3355fef628f5b9b7aeae7ab210\": container with ID starting with 5de72f4aa8c1f52947b1965ec66a5e532f6d3c3355fef628f5b9b7aeae7ab210 not found: ID does not exist" containerID="5de72f4aa8c1f52947b1965ec66a5e532f6d3c3355fef628f5b9b7aeae7ab210" Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.593998 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5de72f4aa8c1f52947b1965ec66a5e532f6d3c3355fef628f5b9b7aeae7ab210"} err="failed to get container status \"5de72f4aa8c1f52947b1965ec66a5e532f6d3c3355fef628f5b9b7aeae7ab210\": rpc error: code = NotFound desc = could not find container \"5de72f4aa8c1f52947b1965ec66a5e532f6d3c3355fef628f5b9b7aeae7ab210\": container with ID starting with 5de72f4aa8c1f52947b1965ec66a5e532f6d3c3355fef628f5b9b7aeae7ab210 not found: ID does not exist" Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.594016 4997 scope.go:117] "RemoveContainer" containerID="9e8cf71b7f0898961ed6b3b9f800f532557dbab642ccdb154c3862a8e2e2d2cc" Dec 05 08:57:56 crc kubenswrapper[4997]: E1205 08:57:56.594296 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e8cf71b7f0898961ed6b3b9f800f532557dbab642ccdb154c3862a8e2e2d2cc\": container with ID starting with 9e8cf71b7f0898961ed6b3b9f800f532557dbab642ccdb154c3862a8e2e2d2cc not found: ID does not exist" containerID="9e8cf71b7f0898961ed6b3b9f800f532557dbab642ccdb154c3862a8e2e2d2cc" Dec 05 08:57:56 crc kubenswrapper[4997]: I1205 08:57:56.594354 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e8cf71b7f0898961ed6b3b9f800f532557dbab642ccdb154c3862a8e2e2d2cc"} err="failed to get container status \"9e8cf71b7f0898961ed6b3b9f800f532557dbab642ccdb154c3862a8e2e2d2cc\": rpc error: code = NotFound desc = could not find container \"9e8cf71b7f0898961ed6b3b9f800f532557dbab642ccdb154c3862a8e2e2d2cc\": container with ID starting with 9e8cf71b7f0898961ed6b3b9f800f532557dbab642ccdb154c3862a8e2e2d2cc not found: ID does not exist" Dec 05 08:57:57 crc kubenswrapper[4997]: I1205 08:57:57.763579 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb204e34-3079-463a-833b-fcd2e5d2dd29" path="/var/lib/kubelet/pods/bb204e34-3079-463a-833b-fcd2e5d2dd29/volumes" Dec 05 08:57:57 crc kubenswrapper[4997]: I1205 08:57:57.877939 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 08:57:57 crc kubenswrapper[4997]: I1205 08:57:57.878007 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 08:57:59 crc kubenswrapper[4997]: I1205 08:57:59.950164 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 08:58:00 crc kubenswrapper[4997]: I1205 08:58:00.828859 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.342931 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-qb474"] Dec 05 08:58:01 crc kubenswrapper[4997]: E1205 08:58:01.343478 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb204e34-3079-463a-833b-fcd2e5d2dd29" containerName="extract-content" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.343500 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb204e34-3079-463a-833b-fcd2e5d2dd29" containerName="extract-content" Dec 05 08:58:01 crc kubenswrapper[4997]: E1205 08:58:01.343527 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb204e34-3079-463a-833b-fcd2e5d2dd29" containerName="registry-server" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.343541 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb204e34-3079-463a-833b-fcd2e5d2dd29" containerName="registry-server" Dec 05 08:58:01 crc kubenswrapper[4997]: E1205 08:58:01.343576 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb204e34-3079-463a-833b-fcd2e5d2dd29" containerName="extract-utilities" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.343592 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb204e34-3079-463a-833b-fcd2e5d2dd29" containerName="extract-utilities" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.343967 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb204e34-3079-463a-833b-fcd2e5d2dd29" containerName="registry-server" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.345413 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-qb474" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.347724 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.354245 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.376856 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-qb474"] Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.529203 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzn94\" (UniqueName: \"kubernetes.io/projected/6b8eda8f-e194-47b6-a429-07fbd87596ae-kube-api-access-lzn94\") pod \"nova-cell1-cell-mapping-qb474\" (UID: \"6b8eda8f-e194-47b6-a429-07fbd87596ae\") " pod="openstack/nova-cell1-cell-mapping-qb474" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.529331 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-qb474\" (UID: \"6b8eda8f-e194-47b6-a429-07fbd87596ae\") " pod="openstack/nova-cell1-cell-mapping-qb474" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.529382 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-scripts\") pod \"nova-cell1-cell-mapping-qb474\" (UID: \"6b8eda8f-e194-47b6-a429-07fbd87596ae\") " pod="openstack/nova-cell1-cell-mapping-qb474" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.529415 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-config-data\") pod \"nova-cell1-cell-mapping-qb474\" (UID: \"6b8eda8f-e194-47b6-a429-07fbd87596ae\") " pod="openstack/nova-cell1-cell-mapping-qb474" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.631060 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-config-data\") pod \"nova-cell1-cell-mapping-qb474\" (UID: \"6b8eda8f-e194-47b6-a429-07fbd87596ae\") " pod="openstack/nova-cell1-cell-mapping-qb474" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.631283 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzn94\" (UniqueName: \"kubernetes.io/projected/6b8eda8f-e194-47b6-a429-07fbd87596ae-kube-api-access-lzn94\") pod \"nova-cell1-cell-mapping-qb474\" (UID: \"6b8eda8f-e194-47b6-a429-07fbd87596ae\") " pod="openstack/nova-cell1-cell-mapping-qb474" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.631396 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-qb474\" (UID: \"6b8eda8f-e194-47b6-a429-07fbd87596ae\") " pod="openstack/nova-cell1-cell-mapping-qb474" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.631437 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-scripts\") pod \"nova-cell1-cell-mapping-qb474\" (UID: \"6b8eda8f-e194-47b6-a429-07fbd87596ae\") " pod="openstack/nova-cell1-cell-mapping-qb474" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.637743 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-scripts\") pod \"nova-cell1-cell-mapping-qb474\" (UID: \"6b8eda8f-e194-47b6-a429-07fbd87596ae\") " pod="openstack/nova-cell1-cell-mapping-qb474" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.641105 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-qb474\" (UID: \"6b8eda8f-e194-47b6-a429-07fbd87596ae\") " pod="openstack/nova-cell1-cell-mapping-qb474" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.641833 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-config-data\") pod \"nova-cell1-cell-mapping-qb474\" (UID: \"6b8eda8f-e194-47b6-a429-07fbd87596ae\") " pod="openstack/nova-cell1-cell-mapping-qb474" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.650374 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzn94\" (UniqueName: \"kubernetes.io/projected/6b8eda8f-e194-47b6-a429-07fbd87596ae-kube-api-access-lzn94\") pod \"nova-cell1-cell-mapping-qb474\" (UID: \"6b8eda8f-e194-47b6-a429-07fbd87596ae\") " pod="openstack/nova-cell1-cell-mapping-qb474" Dec 05 08:58:01 crc kubenswrapper[4997]: I1205 08:58:01.670382 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-qb474" Dec 05 08:58:02 crc kubenswrapper[4997]: I1205 08:58:02.150848 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-qb474"] Dec 05 08:58:02 crc kubenswrapper[4997]: I1205 08:58:02.522688 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-qb474" event={"ID":"6b8eda8f-e194-47b6-a429-07fbd87596ae","Type":"ContainerStarted","Data":"c8bab95a601887e8c2af010a315e78a4efd70781ad4d1cb0a918c0ace2958620"} Dec 05 08:58:02 crc kubenswrapper[4997]: I1205 08:58:02.862446 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 08:58:02 crc kubenswrapper[4997]: I1205 08:58:02.862866 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 08:58:02 crc kubenswrapper[4997]: I1205 08:58:02.878589 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 08:58:02 crc kubenswrapper[4997]: I1205 08:58:02.878657 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 08:58:03 crc kubenswrapper[4997]: I1205 08:58:03.532089 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-qb474" event={"ID":"6b8eda8f-e194-47b6-a429-07fbd87596ae","Type":"ContainerStarted","Data":"0fdc373de62acc63a374a7697504a0e6168c81577a14a035e7b3a0b2c3613768"} Dec 05 08:58:03 crc kubenswrapper[4997]: I1205 08:58:03.559639 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-qb474" podStartSLOduration=2.559611218 podStartE2EDuration="2.559611218s" podCreationTimestamp="2025-12-05 08:58:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:58:03.551725795 +0000 UTC m=+7384.080633066" watchObservedRunningTime="2025-12-05 08:58:03.559611218 +0000 UTC m=+7384.088518479" Dec 05 08:58:04 crc kubenswrapper[4997]: I1205 08:58:04.029306 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="10c752a8-8c4f-42a4-9497-b4ae222788bc" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.82:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:58:04 crc kubenswrapper[4997]: I1205 08:58:04.029756 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="10c752a8-8c4f-42a4-9497-b4ae222788bc" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.82:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:58:04 crc kubenswrapper[4997]: I1205 08:58:04.029839 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2698529d-c31d-43eb-940f-5eee6049968f" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.83:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:58:04 crc kubenswrapper[4997]: I1205 08:58:04.029840 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2698529d-c31d-43eb-940f-5eee6049968f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.83:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:58:04 crc kubenswrapper[4997]: I1205 08:58:04.950026 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 08:58:04 crc kubenswrapper[4997]: I1205 08:58:04.974708 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 08:58:05 crc kubenswrapper[4997]: I1205 08:58:05.577831 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 08:58:07 crc kubenswrapper[4997]: I1205 08:58:07.568659 4997 generic.go:334] "Generic (PLEG): container finished" podID="6b8eda8f-e194-47b6-a429-07fbd87596ae" containerID="0fdc373de62acc63a374a7697504a0e6168c81577a14a035e7b3a0b2c3613768" exitCode=0 Dec 05 08:58:07 crc kubenswrapper[4997]: I1205 08:58:07.568764 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-qb474" event={"ID":"6b8eda8f-e194-47b6-a429-07fbd87596ae","Type":"ContainerDied","Data":"0fdc373de62acc63a374a7697504a0e6168c81577a14a035e7b3a0b2c3613768"} Dec 05 08:58:08 crc kubenswrapper[4997]: I1205 08:58:08.990687 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-qb474" Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.071881 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-combined-ca-bundle\") pod \"6b8eda8f-e194-47b6-a429-07fbd87596ae\" (UID: \"6b8eda8f-e194-47b6-a429-07fbd87596ae\") " Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.072145 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzn94\" (UniqueName: \"kubernetes.io/projected/6b8eda8f-e194-47b6-a429-07fbd87596ae-kube-api-access-lzn94\") pod \"6b8eda8f-e194-47b6-a429-07fbd87596ae\" (UID: \"6b8eda8f-e194-47b6-a429-07fbd87596ae\") " Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.072225 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-config-data\") pod \"6b8eda8f-e194-47b6-a429-07fbd87596ae\" (UID: \"6b8eda8f-e194-47b6-a429-07fbd87596ae\") " Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.072304 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-scripts\") pod \"6b8eda8f-e194-47b6-a429-07fbd87596ae\" (UID: \"6b8eda8f-e194-47b6-a429-07fbd87596ae\") " Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.077818 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-scripts" (OuterVolumeSpecName: "scripts") pod "6b8eda8f-e194-47b6-a429-07fbd87596ae" (UID: "6b8eda8f-e194-47b6-a429-07fbd87596ae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.079071 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b8eda8f-e194-47b6-a429-07fbd87596ae-kube-api-access-lzn94" (OuterVolumeSpecName: "kube-api-access-lzn94") pod "6b8eda8f-e194-47b6-a429-07fbd87596ae" (UID: "6b8eda8f-e194-47b6-a429-07fbd87596ae"). InnerVolumeSpecName "kube-api-access-lzn94". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.102908 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-config-data" (OuterVolumeSpecName: "config-data") pod "6b8eda8f-e194-47b6-a429-07fbd87596ae" (UID: "6b8eda8f-e194-47b6-a429-07fbd87596ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.105639 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b8eda8f-e194-47b6-a429-07fbd87596ae" (UID: "6b8eda8f-e194-47b6-a429-07fbd87596ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.174875 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzn94\" (UniqueName: \"kubernetes.io/projected/6b8eda8f-e194-47b6-a429-07fbd87596ae-kube-api-access-lzn94\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.175114 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.175175 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.175260 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8eda8f-e194-47b6-a429-07fbd87596ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.592446 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-qb474" event={"ID":"6b8eda8f-e194-47b6-a429-07fbd87596ae","Type":"ContainerDied","Data":"c8bab95a601887e8c2af010a315e78a4efd70781ad4d1cb0a918c0ace2958620"} Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.592772 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8bab95a601887e8c2af010a315e78a4efd70781ad4d1cb0a918c0ace2958620" Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.592585 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-qb474" Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.778529 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.778789 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="475e068f-94be-4d18-a55d-0e1940748dfb" containerName="nova-scheduler-scheduler" containerID="cri-o://dca43fc19b1f57dc4e660c95619d128f26c8ff15bbab861870f742c4a77bcfa4" gracePeriod=30 Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.786757 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.787026 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="10c752a8-8c4f-42a4-9497-b4ae222788bc" containerName="nova-api-log" containerID="cri-o://4fc9327f8eb3e195e86413bdd836dc2c17547f019bf0ff23334c605fd02a7948" gracePeriod=30 Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.787570 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="10c752a8-8c4f-42a4-9497-b4ae222788bc" containerName="nova-api-api" containerID="cri-o://dae181f36626243af1e9ea6ee27a6717e88cee749a5953bd62afe49a19a125ef" gracePeriod=30 Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.798018 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.798238 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2698529d-c31d-43eb-940f-5eee6049968f" containerName="nova-metadata-log" containerID="cri-o://876fcd471b6cc44ddb752682f0fb575b0ef56ef63e5cd351cd4d109e063a22e8" gracePeriod=30 Dec 05 08:58:09 crc kubenswrapper[4997]: I1205 08:58:09.798328 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2698529d-c31d-43eb-940f-5eee6049968f" containerName="nova-metadata-metadata" containerID="cri-o://058a98233594c255f126f64174ad011eb820878514c62f60f050944f04f7d5ab" gracePeriod=30 Dec 05 08:58:09 crc kubenswrapper[4997]: E1205 08:58:09.952787 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dca43fc19b1f57dc4e660c95619d128f26c8ff15bbab861870f742c4a77bcfa4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 08:58:09 crc kubenswrapper[4997]: E1205 08:58:09.954783 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dca43fc19b1f57dc4e660c95619d128f26c8ff15bbab861870f742c4a77bcfa4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 08:58:09 crc kubenswrapper[4997]: E1205 08:58:09.956548 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dca43fc19b1f57dc4e660c95619d128f26c8ff15bbab861870f742c4a77bcfa4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 08:58:09 crc kubenswrapper[4997]: E1205 08:58:09.956631 4997 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="475e068f-94be-4d18-a55d-0e1940748dfb" containerName="nova-scheduler-scheduler" Dec 05 08:58:11 crc kubenswrapper[4997]: I1205 08:58:11.612516 4997 generic.go:334] "Generic (PLEG): container finished" podID="2698529d-c31d-43eb-940f-5eee6049968f" containerID="876fcd471b6cc44ddb752682f0fb575b0ef56ef63e5cd351cd4d109e063a22e8" exitCode=143 Dec 05 08:58:11 crc kubenswrapper[4997]: I1205 08:58:11.612591 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2698529d-c31d-43eb-940f-5eee6049968f","Type":"ContainerDied","Data":"876fcd471b6cc44ddb752682f0fb575b0ef56ef63e5cd351cd4d109e063a22e8"} Dec 05 08:58:11 crc kubenswrapper[4997]: I1205 08:58:11.614968 4997 generic.go:334] "Generic (PLEG): container finished" podID="10c752a8-8c4f-42a4-9497-b4ae222788bc" containerID="4fc9327f8eb3e195e86413bdd836dc2c17547f019bf0ff23334c605fd02a7948" exitCode=143 Dec 05 08:58:11 crc kubenswrapper[4997]: I1205 08:58:11.615032 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"10c752a8-8c4f-42a4-9497-b4ae222788bc","Type":"ContainerDied","Data":"4fc9327f8eb3e195e86413bdd836dc2c17547f019bf0ff23334c605fd02a7948"} Dec 05 08:58:12 crc kubenswrapper[4997]: I1205 08:58:12.625228 4997 generic.go:334] "Generic (PLEG): container finished" podID="475e068f-94be-4d18-a55d-0e1940748dfb" containerID="dca43fc19b1f57dc4e660c95619d128f26c8ff15bbab861870f742c4a77bcfa4" exitCode=0 Dec 05 08:58:12 crc kubenswrapper[4997]: I1205 08:58:12.625405 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"475e068f-94be-4d18-a55d-0e1940748dfb","Type":"ContainerDied","Data":"dca43fc19b1f57dc4e660c95619d128f26c8ff15bbab861870f742c4a77bcfa4"} Dec 05 08:58:12 crc kubenswrapper[4997]: I1205 08:58:12.625651 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"475e068f-94be-4d18-a55d-0e1940748dfb","Type":"ContainerDied","Data":"73bcaa581d3d0e1953a8bc1b281e5b169effa5a7a1070582ac7be71970f1290c"} Dec 05 08:58:12 crc kubenswrapper[4997]: I1205 08:58:12.625680 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73bcaa581d3d0e1953a8bc1b281e5b169effa5a7a1070582ac7be71970f1290c" Dec 05 08:58:12 crc kubenswrapper[4997]: I1205 08:58:12.645502 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:58:12 crc kubenswrapper[4997]: I1205 08:58:12.735679 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/475e068f-94be-4d18-a55d-0e1940748dfb-config-data\") pod \"475e068f-94be-4d18-a55d-0e1940748dfb\" (UID: \"475e068f-94be-4d18-a55d-0e1940748dfb\") " Dec 05 08:58:12 crc kubenswrapper[4997]: I1205 08:58:12.736407 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/475e068f-94be-4d18-a55d-0e1940748dfb-combined-ca-bundle\") pod \"475e068f-94be-4d18-a55d-0e1940748dfb\" (UID: \"475e068f-94be-4d18-a55d-0e1940748dfb\") " Dec 05 08:58:12 crc kubenswrapper[4997]: I1205 08:58:12.736578 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjk6z\" (UniqueName: \"kubernetes.io/projected/475e068f-94be-4d18-a55d-0e1940748dfb-kube-api-access-xjk6z\") pod \"475e068f-94be-4d18-a55d-0e1940748dfb\" (UID: \"475e068f-94be-4d18-a55d-0e1940748dfb\") " Dec 05 08:58:12 crc kubenswrapper[4997]: I1205 08:58:12.741989 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/475e068f-94be-4d18-a55d-0e1940748dfb-kube-api-access-xjk6z" (OuterVolumeSpecName: "kube-api-access-xjk6z") pod "475e068f-94be-4d18-a55d-0e1940748dfb" (UID: "475e068f-94be-4d18-a55d-0e1940748dfb"). InnerVolumeSpecName "kube-api-access-xjk6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:58:12 crc kubenswrapper[4997]: I1205 08:58:12.762412 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/475e068f-94be-4d18-a55d-0e1940748dfb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "475e068f-94be-4d18-a55d-0e1940748dfb" (UID: "475e068f-94be-4d18-a55d-0e1940748dfb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:58:12 crc kubenswrapper[4997]: I1205 08:58:12.767283 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/475e068f-94be-4d18-a55d-0e1940748dfb-config-data" (OuterVolumeSpecName: "config-data") pod "475e068f-94be-4d18-a55d-0e1940748dfb" (UID: "475e068f-94be-4d18-a55d-0e1940748dfb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:58:12 crc kubenswrapper[4997]: I1205 08:58:12.839047 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjk6z\" (UniqueName: \"kubernetes.io/projected/475e068f-94be-4d18-a55d-0e1940748dfb-kube-api-access-xjk6z\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:12 crc kubenswrapper[4997]: I1205 08:58:12.839088 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/475e068f-94be-4d18-a55d-0e1940748dfb-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:12 crc kubenswrapper[4997]: I1205 08:58:12.839105 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/475e068f-94be-4d18-a55d-0e1940748dfb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.419079 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.426795 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.447890 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10c752a8-8c4f-42a4-9497-b4ae222788bc-config-data\") pod \"10c752a8-8c4f-42a4-9497-b4ae222788bc\" (UID: \"10c752a8-8c4f-42a4-9497-b4ae222788bc\") " Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.447970 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10c752a8-8c4f-42a4-9497-b4ae222788bc-logs\") pod \"10c752a8-8c4f-42a4-9497-b4ae222788bc\" (UID: \"10c752a8-8c4f-42a4-9497-b4ae222788bc\") " Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.448013 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c752a8-8c4f-42a4-9497-b4ae222788bc-combined-ca-bundle\") pod \"10c752a8-8c4f-42a4-9497-b4ae222788bc\" (UID: \"10c752a8-8c4f-42a4-9497-b4ae222788bc\") " Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.448075 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vjm8\" (UniqueName: \"kubernetes.io/projected/10c752a8-8c4f-42a4-9497-b4ae222788bc-kube-api-access-5vjm8\") pod \"10c752a8-8c4f-42a4-9497-b4ae222788bc\" (UID: \"10c752a8-8c4f-42a4-9497-b4ae222788bc\") " Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.448643 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10c752a8-8c4f-42a4-9497-b4ae222788bc-logs" (OuterVolumeSpecName: "logs") pod "10c752a8-8c4f-42a4-9497-b4ae222788bc" (UID: "10c752a8-8c4f-42a4-9497-b4ae222788bc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.450438 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10c752a8-8c4f-42a4-9497-b4ae222788bc-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.453910 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10c752a8-8c4f-42a4-9497-b4ae222788bc-kube-api-access-5vjm8" (OuterVolumeSpecName: "kube-api-access-5vjm8") pod "10c752a8-8c4f-42a4-9497-b4ae222788bc" (UID: "10c752a8-8c4f-42a4-9497-b4ae222788bc"). InnerVolumeSpecName "kube-api-access-5vjm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.480961 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10c752a8-8c4f-42a4-9497-b4ae222788bc-config-data" (OuterVolumeSpecName: "config-data") pod "10c752a8-8c4f-42a4-9497-b4ae222788bc" (UID: "10c752a8-8c4f-42a4-9497-b4ae222788bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.481057 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10c752a8-8c4f-42a4-9497-b4ae222788bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10c752a8-8c4f-42a4-9497-b4ae222788bc" (UID: "10c752a8-8c4f-42a4-9497-b4ae222788bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.551458 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2698529d-c31d-43eb-940f-5eee6049968f-config-data\") pod \"2698529d-c31d-43eb-940f-5eee6049968f\" (UID: \"2698529d-c31d-43eb-940f-5eee6049968f\") " Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.551555 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hft4l\" (UniqueName: \"kubernetes.io/projected/2698529d-c31d-43eb-940f-5eee6049968f-kube-api-access-hft4l\") pod \"2698529d-c31d-43eb-940f-5eee6049968f\" (UID: \"2698529d-c31d-43eb-940f-5eee6049968f\") " Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.551672 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2698529d-c31d-43eb-940f-5eee6049968f-logs\") pod \"2698529d-c31d-43eb-940f-5eee6049968f\" (UID: \"2698529d-c31d-43eb-940f-5eee6049968f\") " Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.551701 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2698529d-c31d-43eb-940f-5eee6049968f-combined-ca-bundle\") pod \"2698529d-c31d-43eb-940f-5eee6049968f\" (UID: \"2698529d-c31d-43eb-940f-5eee6049968f\") " Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.552097 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10c752a8-8c4f-42a4-9497-b4ae222788bc-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.552120 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c752a8-8c4f-42a4-9497-b4ae222788bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.552131 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vjm8\" (UniqueName: \"kubernetes.io/projected/10c752a8-8c4f-42a4-9497-b4ae222788bc-kube-api-access-5vjm8\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.552481 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2698529d-c31d-43eb-940f-5eee6049968f-logs" (OuterVolumeSpecName: "logs") pod "2698529d-c31d-43eb-940f-5eee6049968f" (UID: "2698529d-c31d-43eb-940f-5eee6049968f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.554928 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2698529d-c31d-43eb-940f-5eee6049968f-kube-api-access-hft4l" (OuterVolumeSpecName: "kube-api-access-hft4l") pod "2698529d-c31d-43eb-940f-5eee6049968f" (UID: "2698529d-c31d-43eb-940f-5eee6049968f"). InnerVolumeSpecName "kube-api-access-hft4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.571833 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2698529d-c31d-43eb-940f-5eee6049968f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2698529d-c31d-43eb-940f-5eee6049968f" (UID: "2698529d-c31d-43eb-940f-5eee6049968f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.572321 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2698529d-c31d-43eb-940f-5eee6049968f-config-data" (OuterVolumeSpecName: "config-data") pod "2698529d-c31d-43eb-940f-5eee6049968f" (UID: "2698529d-c31d-43eb-940f-5eee6049968f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.634269 4997 generic.go:334] "Generic (PLEG): container finished" podID="10c752a8-8c4f-42a4-9497-b4ae222788bc" containerID="dae181f36626243af1e9ea6ee27a6717e88cee749a5953bd62afe49a19a125ef" exitCode=0 Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.634350 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"10c752a8-8c4f-42a4-9497-b4ae222788bc","Type":"ContainerDied","Data":"dae181f36626243af1e9ea6ee27a6717e88cee749a5953bd62afe49a19a125ef"} Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.634383 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"10c752a8-8c4f-42a4-9497-b4ae222788bc","Type":"ContainerDied","Data":"1e4646703b00cbd3681ab5f508581a7de679e65b9905b81da5a8ba63e0a7d3a1"} Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.634403 4997 scope.go:117] "RemoveContainer" containerID="dae181f36626243af1e9ea6ee27a6717e88cee749a5953bd62afe49a19a125ef" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.634353 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.637720 4997 generic.go:334] "Generic (PLEG): container finished" podID="2698529d-c31d-43eb-940f-5eee6049968f" containerID="058a98233594c255f126f64174ad011eb820878514c62f60f050944f04f7d5ab" exitCode=0 Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.637828 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.637830 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.637825 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2698529d-c31d-43eb-940f-5eee6049968f","Type":"ContainerDied","Data":"058a98233594c255f126f64174ad011eb820878514c62f60f050944f04f7d5ab"} Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.637929 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2698529d-c31d-43eb-940f-5eee6049968f","Type":"ContainerDied","Data":"4ceec146f9f58d1b208289caf6b55c0a65a15fb2cc5bd79417a78d52d032ca24"} Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.653867 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2698529d-c31d-43eb-940f-5eee6049968f-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.653897 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2698529d-c31d-43eb-940f-5eee6049968f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.653950 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2698529d-c31d-43eb-940f-5eee6049968f-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.653962 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hft4l\" (UniqueName: \"kubernetes.io/projected/2698529d-c31d-43eb-940f-5eee6049968f-kube-api-access-hft4l\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.666851 4997 scope.go:117] "RemoveContainer" containerID="4fc9327f8eb3e195e86413bdd836dc2c17547f019bf0ff23334c605fd02a7948" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.679313 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.703357 4997 scope.go:117] "RemoveContainer" containerID="dae181f36626243af1e9ea6ee27a6717e88cee749a5953bd62afe49a19a125ef" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.706710 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:58:13 crc kubenswrapper[4997]: E1205 08:58:13.706872 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dae181f36626243af1e9ea6ee27a6717e88cee749a5953bd62afe49a19a125ef\": container with ID starting with dae181f36626243af1e9ea6ee27a6717e88cee749a5953bd62afe49a19a125ef not found: ID does not exist" containerID="dae181f36626243af1e9ea6ee27a6717e88cee749a5953bd62afe49a19a125ef" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.706983 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dae181f36626243af1e9ea6ee27a6717e88cee749a5953bd62afe49a19a125ef"} err="failed to get container status \"dae181f36626243af1e9ea6ee27a6717e88cee749a5953bd62afe49a19a125ef\": rpc error: code = NotFound desc = could not find container \"dae181f36626243af1e9ea6ee27a6717e88cee749a5953bd62afe49a19a125ef\": container with ID starting with dae181f36626243af1e9ea6ee27a6717e88cee749a5953bd62afe49a19a125ef not found: ID does not exist" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.707063 4997 scope.go:117] "RemoveContainer" containerID="4fc9327f8eb3e195e86413bdd836dc2c17547f019bf0ff23334c605fd02a7948" Dec 05 08:58:13 crc kubenswrapper[4997]: E1205 08:58:13.710885 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fc9327f8eb3e195e86413bdd836dc2c17547f019bf0ff23334c605fd02a7948\": container with ID starting with 4fc9327f8eb3e195e86413bdd836dc2c17547f019bf0ff23334c605fd02a7948 not found: ID does not exist" containerID="4fc9327f8eb3e195e86413bdd836dc2c17547f019bf0ff23334c605fd02a7948" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.711019 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fc9327f8eb3e195e86413bdd836dc2c17547f019bf0ff23334c605fd02a7948"} err="failed to get container status \"4fc9327f8eb3e195e86413bdd836dc2c17547f019bf0ff23334c605fd02a7948\": rpc error: code = NotFound desc = could not find container \"4fc9327f8eb3e195e86413bdd836dc2c17547f019bf0ff23334c605fd02a7948\": container with ID starting with 4fc9327f8eb3e195e86413bdd836dc2c17547f019bf0ff23334c605fd02a7948 not found: ID does not exist" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.711116 4997 scope.go:117] "RemoveContainer" containerID="058a98233594c255f126f64174ad011eb820878514c62f60f050944f04f7d5ab" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.720380 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.727910 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.734823 4997 scope.go:117] "RemoveContainer" containerID="876fcd471b6cc44ddb752682f0fb575b0ef56ef63e5cd351cd4d109e063a22e8" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.740525 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 08:58:13 crc kubenswrapper[4997]: E1205 08:58:13.740997 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b8eda8f-e194-47b6-a429-07fbd87596ae" containerName="nova-manage" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.741014 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b8eda8f-e194-47b6-a429-07fbd87596ae" containerName="nova-manage" Dec 05 08:58:13 crc kubenswrapper[4997]: E1205 08:58:13.741033 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10c752a8-8c4f-42a4-9497-b4ae222788bc" containerName="nova-api-log" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.741039 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="10c752a8-8c4f-42a4-9497-b4ae222788bc" containerName="nova-api-log" Dec 05 08:58:13 crc kubenswrapper[4997]: E1205 08:58:13.741051 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2698529d-c31d-43eb-940f-5eee6049968f" containerName="nova-metadata-log" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.741057 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2698529d-c31d-43eb-940f-5eee6049968f" containerName="nova-metadata-log" Dec 05 08:58:13 crc kubenswrapper[4997]: E1205 08:58:13.741080 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10c752a8-8c4f-42a4-9497-b4ae222788bc" containerName="nova-api-api" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.741086 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="10c752a8-8c4f-42a4-9497-b4ae222788bc" containerName="nova-api-api" Dec 05 08:58:13 crc kubenswrapper[4997]: E1205 08:58:13.741097 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="475e068f-94be-4d18-a55d-0e1940748dfb" containerName="nova-scheduler-scheduler" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.741103 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="475e068f-94be-4d18-a55d-0e1940748dfb" containerName="nova-scheduler-scheduler" Dec 05 08:58:13 crc kubenswrapper[4997]: E1205 08:58:13.741121 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2698529d-c31d-43eb-940f-5eee6049968f" containerName="nova-metadata-metadata" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.741127 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2698529d-c31d-43eb-940f-5eee6049968f" containerName="nova-metadata-metadata" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.741306 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b8eda8f-e194-47b6-a429-07fbd87596ae" containerName="nova-manage" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.741325 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="10c752a8-8c4f-42a4-9497-b4ae222788bc" containerName="nova-api-log" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.741335 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2698529d-c31d-43eb-940f-5eee6049968f" containerName="nova-metadata-metadata" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.741345 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2698529d-c31d-43eb-940f-5eee6049968f" containerName="nova-metadata-log" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.741355 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="10c752a8-8c4f-42a4-9497-b4ae222788bc" containerName="nova-api-api" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.741367 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="475e068f-94be-4d18-a55d-0e1940748dfb" containerName="nova-scheduler-scheduler" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.742585 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.754196 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.754372 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.778576 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10c752a8-8c4f-42a4-9497-b4ae222788bc" path="/var/lib/kubelet/pods/10c752a8-8c4f-42a4-9497-b4ae222788bc/volumes" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.779222 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2698529d-c31d-43eb-940f-5eee6049968f" path="/var/lib/kubelet/pods/2698529d-c31d-43eb-940f-5eee6049968f/volumes" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.779825 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.779855 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.781273 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.783501 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.784139 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.793141 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.802825 4997 scope.go:117] "RemoveContainer" containerID="058a98233594c255f126f64174ad011eb820878514c62f60f050944f04f7d5ab" Dec 05 08:58:13 crc kubenswrapper[4997]: E1205 08:58:13.803289 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"058a98233594c255f126f64174ad011eb820878514c62f60f050944f04f7d5ab\": container with ID starting with 058a98233594c255f126f64174ad011eb820878514c62f60f050944f04f7d5ab not found: ID does not exist" containerID="058a98233594c255f126f64174ad011eb820878514c62f60f050944f04f7d5ab" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.803316 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"058a98233594c255f126f64174ad011eb820878514c62f60f050944f04f7d5ab"} err="failed to get container status \"058a98233594c255f126f64174ad011eb820878514c62f60f050944f04f7d5ab\": rpc error: code = NotFound desc = could not find container \"058a98233594c255f126f64174ad011eb820878514c62f60f050944f04f7d5ab\": container with ID starting with 058a98233594c255f126f64174ad011eb820878514c62f60f050944f04f7d5ab not found: ID does not exist" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.803356 4997 scope.go:117] "RemoveContainer" containerID="876fcd471b6cc44ddb752682f0fb575b0ef56ef63e5cd351cd4d109e063a22e8" Dec 05 08:58:13 crc kubenswrapper[4997]: E1205 08:58:13.804061 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"876fcd471b6cc44ddb752682f0fb575b0ef56ef63e5cd351cd4d109e063a22e8\": container with ID starting with 876fcd471b6cc44ddb752682f0fb575b0ef56ef63e5cd351cd4d109e063a22e8 not found: ID does not exist" containerID="876fcd471b6cc44ddb752682f0fb575b0ef56ef63e5cd351cd4d109e063a22e8" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.804105 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"876fcd471b6cc44ddb752682f0fb575b0ef56ef63e5cd351cd4d109e063a22e8"} err="failed to get container status \"876fcd471b6cc44ddb752682f0fb575b0ef56ef63e5cd351cd4d109e063a22e8\": rpc error: code = NotFound desc = could not find container \"876fcd471b6cc44ddb752682f0fb575b0ef56ef63e5cd351cd4d109e063a22e8\": container with ID starting with 876fcd471b6cc44ddb752682f0fb575b0ef56ef63e5cd351cd4d109e063a22e8 not found: ID does not exist" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.806711 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.808074 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.809538 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.821509 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.859359 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-config-data\") pod \"nova-metadata-0\" (UID: \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\") " pod="openstack/nova-metadata-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.859446 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c366489-0d4c-48e1-99be-5b94881cd468-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c366489-0d4c-48e1-99be-5b94881cd468\") " pod="openstack/nova-api-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.859487 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-logs\") pod \"nova-metadata-0\" (UID: \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\") " pod="openstack/nova-metadata-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.859545 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\") " pod="openstack/nova-metadata-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.859685 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700689bf-8f80-457b-8daf-3dfde8871759-config-data\") pod \"nova-scheduler-0\" (UID: \"700689bf-8f80-457b-8daf-3dfde8871759\") " pod="openstack/nova-scheduler-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.859718 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c366489-0d4c-48e1-99be-5b94881cd468-config-data\") pod \"nova-api-0\" (UID: \"8c366489-0d4c-48e1-99be-5b94881cd468\") " pod="openstack/nova-api-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.859767 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54j9h\" (UniqueName: \"kubernetes.io/projected/8c366489-0d4c-48e1-99be-5b94881cd468-kube-api-access-54j9h\") pod \"nova-api-0\" (UID: \"8c366489-0d4c-48e1-99be-5b94881cd468\") " pod="openstack/nova-api-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.859793 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700689bf-8f80-457b-8daf-3dfde8871759-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"700689bf-8f80-457b-8daf-3dfde8871759\") " pod="openstack/nova-scheduler-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.859847 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfvwt\" (UniqueName: \"kubernetes.io/projected/700689bf-8f80-457b-8daf-3dfde8871759-kube-api-access-pfvwt\") pod \"nova-scheduler-0\" (UID: \"700689bf-8f80-457b-8daf-3dfde8871759\") " pod="openstack/nova-scheduler-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.859874 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c366489-0d4c-48e1-99be-5b94881cd468-logs\") pod \"nova-api-0\" (UID: \"8c366489-0d4c-48e1-99be-5b94881cd468\") " pod="openstack/nova-api-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.859893 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc5pb\" (UniqueName: \"kubernetes.io/projected/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-kube-api-access-dc5pb\") pod \"nova-metadata-0\" (UID: \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\") " pod="openstack/nova-metadata-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.961804 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54j9h\" (UniqueName: \"kubernetes.io/projected/8c366489-0d4c-48e1-99be-5b94881cd468-kube-api-access-54j9h\") pod \"nova-api-0\" (UID: \"8c366489-0d4c-48e1-99be-5b94881cd468\") " pod="openstack/nova-api-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.961865 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700689bf-8f80-457b-8daf-3dfde8871759-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"700689bf-8f80-457b-8daf-3dfde8871759\") " pod="openstack/nova-scheduler-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.961907 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfvwt\" (UniqueName: \"kubernetes.io/projected/700689bf-8f80-457b-8daf-3dfde8871759-kube-api-access-pfvwt\") pod \"nova-scheduler-0\" (UID: \"700689bf-8f80-457b-8daf-3dfde8871759\") " pod="openstack/nova-scheduler-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.961936 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c366489-0d4c-48e1-99be-5b94881cd468-logs\") pod \"nova-api-0\" (UID: \"8c366489-0d4c-48e1-99be-5b94881cd468\") " pod="openstack/nova-api-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.961961 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc5pb\" (UniqueName: \"kubernetes.io/projected/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-kube-api-access-dc5pb\") pod \"nova-metadata-0\" (UID: \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\") " pod="openstack/nova-metadata-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.962004 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-config-data\") pod \"nova-metadata-0\" (UID: \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\") " pod="openstack/nova-metadata-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.962033 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c366489-0d4c-48e1-99be-5b94881cd468-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c366489-0d4c-48e1-99be-5b94881cd468\") " pod="openstack/nova-api-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.962063 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-logs\") pod \"nova-metadata-0\" (UID: \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\") " pod="openstack/nova-metadata-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.962103 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\") " pod="openstack/nova-metadata-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.962175 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700689bf-8f80-457b-8daf-3dfde8871759-config-data\") pod \"nova-scheduler-0\" (UID: \"700689bf-8f80-457b-8daf-3dfde8871759\") " pod="openstack/nova-scheduler-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.962196 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c366489-0d4c-48e1-99be-5b94881cd468-config-data\") pod \"nova-api-0\" (UID: \"8c366489-0d4c-48e1-99be-5b94881cd468\") " pod="openstack/nova-api-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.962835 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c366489-0d4c-48e1-99be-5b94881cd468-logs\") pod \"nova-api-0\" (UID: \"8c366489-0d4c-48e1-99be-5b94881cd468\") " pod="openstack/nova-api-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.963013 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-logs\") pod \"nova-metadata-0\" (UID: \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\") " pod="openstack/nova-metadata-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.966307 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c366489-0d4c-48e1-99be-5b94881cd468-config-data\") pod \"nova-api-0\" (UID: \"8c366489-0d4c-48e1-99be-5b94881cd468\") " pod="openstack/nova-api-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.966706 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700689bf-8f80-457b-8daf-3dfde8871759-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"700689bf-8f80-457b-8daf-3dfde8871759\") " pod="openstack/nova-scheduler-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.967241 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c366489-0d4c-48e1-99be-5b94881cd468-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c366489-0d4c-48e1-99be-5b94881cd468\") " pod="openstack/nova-api-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.967668 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-config-data\") pod \"nova-metadata-0\" (UID: \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\") " pod="openstack/nova-metadata-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.967677 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\") " pod="openstack/nova-metadata-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.970600 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700689bf-8f80-457b-8daf-3dfde8871759-config-data\") pod \"nova-scheduler-0\" (UID: \"700689bf-8f80-457b-8daf-3dfde8871759\") " pod="openstack/nova-scheduler-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.978630 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc5pb\" (UniqueName: \"kubernetes.io/projected/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-kube-api-access-dc5pb\") pod \"nova-metadata-0\" (UID: \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\") " pod="openstack/nova-metadata-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.978884 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54j9h\" (UniqueName: \"kubernetes.io/projected/8c366489-0d4c-48e1-99be-5b94881cd468-kube-api-access-54j9h\") pod \"nova-api-0\" (UID: \"8c366489-0d4c-48e1-99be-5b94881cd468\") " pod="openstack/nova-api-0" Dec 05 08:58:13 crc kubenswrapper[4997]: I1205 08:58:13.980418 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfvwt\" (UniqueName: \"kubernetes.io/projected/700689bf-8f80-457b-8daf-3dfde8871759-kube-api-access-pfvwt\") pod \"nova-scheduler-0\" (UID: \"700689bf-8f80-457b-8daf-3dfde8871759\") " pod="openstack/nova-scheduler-0" Dec 05 08:58:14 crc kubenswrapper[4997]: I1205 08:58:14.106561 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:58:14 crc kubenswrapper[4997]: I1205 08:58:14.131342 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:58:14 crc kubenswrapper[4997]: I1205 08:58:14.138342 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:58:14 crc kubenswrapper[4997]: I1205 08:58:14.575564 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:58:14 crc kubenswrapper[4997]: W1205 08:58:14.578926 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ad83fe5_dc32_4e76_a5ef_0bf1b68c1733.slice/crio-8bb7feaf6080b19cc8277f3e0d29967465aed7ba35e985dd9698d9247a255f39 WatchSource:0}: Error finding container 8bb7feaf6080b19cc8277f3e0d29967465aed7ba35e985dd9698d9247a255f39: Status 404 returned error can't find the container with id 8bb7feaf6080b19cc8277f3e0d29967465aed7ba35e985dd9698d9247a255f39 Dec 05 08:58:14 crc kubenswrapper[4997]: I1205 08:58:14.588519 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:58:14 crc kubenswrapper[4997]: I1205 08:58:14.654677 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:58:14 crc kubenswrapper[4997]: I1205 08:58:14.656562 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733","Type":"ContainerStarted","Data":"8bb7feaf6080b19cc8277f3e0d29967465aed7ba35e985dd9698d9247a255f39"} Dec 05 08:58:14 crc kubenswrapper[4997]: I1205 08:58:14.660036 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c366489-0d4c-48e1-99be-5b94881cd468","Type":"ContainerStarted","Data":"d148f4b9a2cd1ef771d7c390ca13219cdbaa19bfe9234154137b206d17af8d26"} Dec 05 08:58:14 crc kubenswrapper[4997]: W1205 08:58:14.668059 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod700689bf_8f80_457b_8daf_3dfde8871759.slice/crio-7763d3fb8bfb455d9ade86a711668b89034949666a376525c1f336fafdafbbb3 WatchSource:0}: Error finding container 7763d3fb8bfb455d9ade86a711668b89034949666a376525c1f336fafdafbbb3: Status 404 returned error can't find the container with id 7763d3fb8bfb455d9ade86a711668b89034949666a376525c1f336fafdafbbb3 Dec 05 08:58:15 crc kubenswrapper[4997]: I1205 08:58:15.683991 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c366489-0d4c-48e1-99be-5b94881cd468","Type":"ContainerStarted","Data":"576ff2a343d43ad6e34c58d476cc43aab9729c86badcf63506fdae3033d099be"} Dec 05 08:58:15 crc kubenswrapper[4997]: I1205 08:58:15.684538 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c366489-0d4c-48e1-99be-5b94881cd468","Type":"ContainerStarted","Data":"a621dd45aff6ca71d2c373c426d8b890b899a53bd037c5203583e71b917ba244"} Dec 05 08:58:15 crc kubenswrapper[4997]: I1205 08:58:15.686548 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733","Type":"ContainerStarted","Data":"ed33e13d3a1086153e0da6846482e061e4d7c00955733b7c2c0ac7b1a7662040"} Dec 05 08:58:15 crc kubenswrapper[4997]: I1205 08:58:15.686585 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733","Type":"ContainerStarted","Data":"315590eb9eb2ef36418367ff8e38e1d476a35163132ddf8c6324fa1e31586887"} Dec 05 08:58:15 crc kubenswrapper[4997]: I1205 08:58:15.688403 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"700689bf-8f80-457b-8daf-3dfde8871759","Type":"ContainerStarted","Data":"5e1515944c4d5662ce252478df8e925f34222dd1e61fb08a7663defebb4255f3"} Dec 05 08:58:15 crc kubenswrapper[4997]: I1205 08:58:15.688437 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"700689bf-8f80-457b-8daf-3dfde8871759","Type":"ContainerStarted","Data":"7763d3fb8bfb455d9ade86a711668b89034949666a376525c1f336fafdafbbb3"} Dec 05 08:58:15 crc kubenswrapper[4997]: I1205 08:58:15.718570 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.718548129 podStartE2EDuration="2.718548129s" podCreationTimestamp="2025-12-05 08:58:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:58:15.70558996 +0000 UTC m=+7396.234497241" watchObservedRunningTime="2025-12-05 08:58:15.718548129 +0000 UTC m=+7396.247455390" Dec 05 08:58:15 crc kubenswrapper[4997]: I1205 08:58:15.733851 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.73382974 podStartE2EDuration="2.73382974s" podCreationTimestamp="2025-12-05 08:58:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:58:15.727697994 +0000 UTC m=+7396.256605275" watchObservedRunningTime="2025-12-05 08:58:15.73382974 +0000 UTC m=+7396.262737001" Dec 05 08:58:15 crc kubenswrapper[4997]: I1205 08:58:15.756606 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.7565847310000002 podStartE2EDuration="2.756584731s" podCreationTimestamp="2025-12-05 08:58:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:58:15.746563872 +0000 UTC m=+7396.275471143" watchObservedRunningTime="2025-12-05 08:58:15.756584731 +0000 UTC m=+7396.285492012" Dec 05 08:58:15 crc kubenswrapper[4997]: I1205 08:58:15.759670 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="475e068f-94be-4d18-a55d-0e1940748dfb" path="/var/lib/kubelet/pods/475e068f-94be-4d18-a55d-0e1940748dfb/volumes" Dec 05 08:58:19 crc kubenswrapper[4997]: I1205 08:58:19.132652 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 08:58:19 crc kubenswrapper[4997]: I1205 08:58:19.132969 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 08:58:19 crc kubenswrapper[4997]: I1205 08:58:19.139444 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 08:58:19 crc kubenswrapper[4997]: I1205 08:58:19.770315 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:58:19 crc kubenswrapper[4997]: I1205 08:58:19.770409 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:58:24 crc kubenswrapper[4997]: I1205 08:58:24.107261 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 08:58:24 crc kubenswrapper[4997]: I1205 08:58:24.107912 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 08:58:24 crc kubenswrapper[4997]: I1205 08:58:24.132320 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 08:58:24 crc kubenswrapper[4997]: I1205 08:58:24.132368 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 08:58:24 crc kubenswrapper[4997]: I1205 08:58:24.139160 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 08:58:24 crc kubenswrapper[4997]: I1205 08:58:24.183727 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 08:58:24 crc kubenswrapper[4997]: I1205 08:58:24.817891 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 08:58:25 crc kubenswrapper[4997]: I1205 08:58:25.270872 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.87:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:58:25 crc kubenswrapper[4997]: I1205 08:58:25.271155 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8c366489-0d4c-48e1-99be-5b94881cd468" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.86:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:58:25 crc kubenswrapper[4997]: I1205 08:58:25.271187 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8c366489-0d4c-48e1-99be-5b94881cd468" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.86:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:58:25 crc kubenswrapper[4997]: I1205 08:58:25.271526 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.87:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.112210 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.112993 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.114031 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.114187 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.117001 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.117353 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.136438 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.137511 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.138427 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.298276 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-694f8cd549-6hns9"] Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.306656 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.335415 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-694f8cd549-6hns9"] Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.357947 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-config\") pod \"dnsmasq-dns-694f8cd549-6hns9\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.358044 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-dns-svc\") pod \"dnsmasq-dns-694f8cd549-6hns9\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.358110 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrhsw\" (UniqueName: \"kubernetes.io/projected/7a0cf64f-5798-4682-8fde-9998c6df1e96-kube-api-access-hrhsw\") pod \"dnsmasq-dns-694f8cd549-6hns9\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.358145 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-ovsdbserver-sb\") pod \"dnsmasq-dns-694f8cd549-6hns9\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.358196 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-ovsdbserver-nb\") pod \"dnsmasq-dns-694f8cd549-6hns9\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.460527 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-ovsdbserver-nb\") pod \"dnsmasq-dns-694f8cd549-6hns9\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.460863 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-config\") pod \"dnsmasq-dns-694f8cd549-6hns9\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.460945 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-dns-svc\") pod \"dnsmasq-dns-694f8cd549-6hns9\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.461005 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrhsw\" (UniqueName: \"kubernetes.io/projected/7a0cf64f-5798-4682-8fde-9998c6df1e96-kube-api-access-hrhsw\") pod \"dnsmasq-dns-694f8cd549-6hns9\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.461033 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-ovsdbserver-sb\") pod \"dnsmasq-dns-694f8cd549-6hns9\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.461872 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-ovsdbserver-sb\") pod \"dnsmasq-dns-694f8cd549-6hns9\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.462407 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-ovsdbserver-nb\") pod \"dnsmasq-dns-694f8cd549-6hns9\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.463015 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-config\") pod \"dnsmasq-dns-694f8cd549-6hns9\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.463426 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-dns-svc\") pod \"dnsmasq-dns-694f8cd549-6hns9\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.488035 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrhsw\" (UniqueName: \"kubernetes.io/projected/7a0cf64f-5798-4682-8fde-9998c6df1e96-kube-api-access-hrhsw\") pod \"dnsmasq-dns-694f8cd549-6hns9\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.646554 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:34 crc kubenswrapper[4997]: I1205 08:58:34.910012 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 08:58:35 crc kubenswrapper[4997]: I1205 08:58:35.281451 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-694f8cd549-6hns9"] Dec 05 08:58:35 crc kubenswrapper[4997]: E1205 08:58:35.805022 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a0cf64f_5798_4682_8fde_9998c6df1e96.slice/crio-conmon-5aaf53ba0b9812f2a9b38ba9a9b887bcd626f2505649d98788f866fbbe7ed9a5.scope\": RecentStats: unable to find data in memory cache]" Dec 05 08:58:35 crc kubenswrapper[4997]: I1205 08:58:35.897668 4997 generic.go:334] "Generic (PLEG): container finished" podID="7a0cf64f-5798-4682-8fde-9998c6df1e96" containerID="5aaf53ba0b9812f2a9b38ba9a9b887bcd626f2505649d98788f866fbbe7ed9a5" exitCode=0 Dec 05 08:58:35 crc kubenswrapper[4997]: I1205 08:58:35.897874 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-694f8cd549-6hns9" event={"ID":"7a0cf64f-5798-4682-8fde-9998c6df1e96","Type":"ContainerDied","Data":"5aaf53ba0b9812f2a9b38ba9a9b887bcd626f2505649d98788f866fbbe7ed9a5"} Dec 05 08:58:35 crc kubenswrapper[4997]: I1205 08:58:35.899123 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-694f8cd549-6hns9" event={"ID":"7a0cf64f-5798-4682-8fde-9998c6df1e96","Type":"ContainerStarted","Data":"8db8dd447b25eb3ff0df4efdf53d0168fa6af4b5588292be8b164fd60b3779fb"} Dec 05 08:58:36 crc kubenswrapper[4997]: I1205 08:58:36.907405 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-694f8cd549-6hns9" event={"ID":"7a0cf64f-5798-4682-8fde-9998c6df1e96","Type":"ContainerStarted","Data":"07329078d555356e3e2a6ca4b0f9e265bcc6dd5ae02938df058c4eedba1c3d9c"} Dec 05 08:58:36 crc kubenswrapper[4997]: I1205 08:58:36.938409 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-694f8cd549-6hns9" podStartSLOduration=2.938390753 podStartE2EDuration="2.938390753s" podCreationTimestamp="2025-12-05 08:58:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:58:36.933238585 +0000 UTC m=+7417.462145846" watchObservedRunningTime="2025-12-05 08:58:36.938390753 +0000 UTC m=+7417.467298034" Dec 05 08:58:37 crc kubenswrapper[4997]: I1205 08:58:37.915857 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:42 crc kubenswrapper[4997]: I1205 08:58:42.096026 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vrbbh"] Dec 05 08:58:42 crc kubenswrapper[4997]: I1205 08:58:42.099941 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrbbh" Dec 05 08:58:42 crc kubenswrapper[4997]: I1205 08:58:42.115024 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vrbbh"] Dec 05 08:58:42 crc kubenswrapper[4997]: I1205 08:58:42.204399 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80e12f51-1310-4be7-a5e9-0f0204dd3e16-catalog-content\") pod \"community-operators-vrbbh\" (UID: \"80e12f51-1310-4be7-a5e9-0f0204dd3e16\") " pod="openshift-marketplace/community-operators-vrbbh" Dec 05 08:58:42 crc kubenswrapper[4997]: I1205 08:58:42.204460 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80e12f51-1310-4be7-a5e9-0f0204dd3e16-utilities\") pod \"community-operators-vrbbh\" (UID: \"80e12f51-1310-4be7-a5e9-0f0204dd3e16\") " pod="openshift-marketplace/community-operators-vrbbh" Dec 05 08:58:42 crc kubenswrapper[4997]: I1205 08:58:42.204633 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjj2h\" (UniqueName: \"kubernetes.io/projected/80e12f51-1310-4be7-a5e9-0f0204dd3e16-kube-api-access-fjj2h\") pod \"community-operators-vrbbh\" (UID: \"80e12f51-1310-4be7-a5e9-0f0204dd3e16\") " pod="openshift-marketplace/community-operators-vrbbh" Dec 05 08:58:42 crc kubenswrapper[4997]: I1205 08:58:42.306524 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjj2h\" (UniqueName: \"kubernetes.io/projected/80e12f51-1310-4be7-a5e9-0f0204dd3e16-kube-api-access-fjj2h\") pod \"community-operators-vrbbh\" (UID: \"80e12f51-1310-4be7-a5e9-0f0204dd3e16\") " pod="openshift-marketplace/community-operators-vrbbh" Dec 05 08:58:42 crc kubenswrapper[4997]: I1205 08:58:42.306655 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80e12f51-1310-4be7-a5e9-0f0204dd3e16-catalog-content\") pod \"community-operators-vrbbh\" (UID: \"80e12f51-1310-4be7-a5e9-0f0204dd3e16\") " pod="openshift-marketplace/community-operators-vrbbh" Dec 05 08:58:42 crc kubenswrapper[4997]: I1205 08:58:42.306676 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80e12f51-1310-4be7-a5e9-0f0204dd3e16-utilities\") pod \"community-operators-vrbbh\" (UID: \"80e12f51-1310-4be7-a5e9-0f0204dd3e16\") " pod="openshift-marketplace/community-operators-vrbbh" Dec 05 08:58:42 crc kubenswrapper[4997]: I1205 08:58:42.307221 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80e12f51-1310-4be7-a5e9-0f0204dd3e16-utilities\") pod \"community-operators-vrbbh\" (UID: \"80e12f51-1310-4be7-a5e9-0f0204dd3e16\") " pod="openshift-marketplace/community-operators-vrbbh" Dec 05 08:58:42 crc kubenswrapper[4997]: I1205 08:58:42.307237 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80e12f51-1310-4be7-a5e9-0f0204dd3e16-catalog-content\") pod \"community-operators-vrbbh\" (UID: \"80e12f51-1310-4be7-a5e9-0f0204dd3e16\") " pod="openshift-marketplace/community-operators-vrbbh" Dec 05 08:58:42 crc kubenswrapper[4997]: I1205 08:58:42.325813 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjj2h\" (UniqueName: \"kubernetes.io/projected/80e12f51-1310-4be7-a5e9-0f0204dd3e16-kube-api-access-fjj2h\") pod \"community-operators-vrbbh\" (UID: \"80e12f51-1310-4be7-a5e9-0f0204dd3e16\") " pod="openshift-marketplace/community-operators-vrbbh" Dec 05 08:58:42 crc kubenswrapper[4997]: I1205 08:58:42.433341 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrbbh" Dec 05 08:58:42 crc kubenswrapper[4997]: I1205 08:58:42.996919 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vrbbh"] Dec 05 08:58:43 crc kubenswrapper[4997]: I1205 08:58:43.977543 4997 generic.go:334] "Generic (PLEG): container finished" podID="80e12f51-1310-4be7-a5e9-0f0204dd3e16" containerID="bd87dfcdc23c29eeabe70f1365a950a69196b5b03280e635ab80bf4211dc7f7e" exitCode=0 Dec 05 08:58:43 crc kubenswrapper[4997]: I1205 08:58:43.977655 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrbbh" event={"ID":"80e12f51-1310-4be7-a5e9-0f0204dd3e16","Type":"ContainerDied","Data":"bd87dfcdc23c29eeabe70f1365a950a69196b5b03280e635ab80bf4211dc7f7e"} Dec 05 08:58:43 crc kubenswrapper[4997]: I1205 08:58:43.977970 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrbbh" event={"ID":"80e12f51-1310-4be7-a5e9-0f0204dd3e16","Type":"ContainerStarted","Data":"97d19b9fd523ba8ba4416f833a2b456660428a0cddaed4e0bd407295be361236"} Dec 05 08:58:43 crc kubenswrapper[4997]: I1205 08:58:43.979720 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 08:58:44 crc kubenswrapper[4997]: I1205 08:58:44.648831 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:58:44 crc kubenswrapper[4997]: I1205 08:58:44.723326 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cd6b446c-ncp6v"] Dec 05 08:58:44 crc kubenswrapper[4997]: I1205 08:58:44.724738 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" podUID="d66c0d05-f098-459a-9398-cf7b1ae4edf1" containerName="dnsmasq-dns" containerID="cri-o://afe2705efd72071cb13623dcd699f8cd8b042f47f2111908cff7f90a10632d8f" gracePeriod=10 Dec 05 08:58:44 crc kubenswrapper[4997]: I1205 08:58:44.995303 4997 generic.go:334] "Generic (PLEG): container finished" podID="d66c0d05-f098-459a-9398-cf7b1ae4edf1" containerID="afe2705efd72071cb13623dcd699f8cd8b042f47f2111908cff7f90a10632d8f" exitCode=0 Dec 05 08:58:44 crc kubenswrapper[4997]: I1205 08:58:44.995425 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" event={"ID":"d66c0d05-f098-459a-9398-cf7b1ae4edf1","Type":"ContainerDied","Data":"afe2705efd72071cb13623dcd699f8cd8b042f47f2111908cff7f90a10632d8f"} Dec 05 08:58:44 crc kubenswrapper[4997]: I1205 08:58:44.998176 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrbbh" event={"ID":"80e12f51-1310-4be7-a5e9-0f0204dd3e16","Type":"ContainerStarted","Data":"71b0d675a1714a9952113b71215fa56c095b31b5733f7899d24e7901aca2932f"} Dec 05 08:58:45 crc kubenswrapper[4997]: I1205 08:58:45.174328 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:58:45 crc kubenswrapper[4997]: I1205 08:58:45.268386 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-ovsdbserver-nb\") pod \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " Dec 05 08:58:45 crc kubenswrapper[4997]: I1205 08:58:45.268726 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-ovsdbserver-sb\") pod \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " Dec 05 08:58:45 crc kubenswrapper[4997]: I1205 08:58:45.268862 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-dns-svc\") pod \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " Dec 05 08:58:45 crc kubenswrapper[4997]: I1205 08:58:45.268942 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bz668\" (UniqueName: \"kubernetes.io/projected/d66c0d05-f098-459a-9398-cf7b1ae4edf1-kube-api-access-bz668\") pod \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " Dec 05 08:58:45 crc kubenswrapper[4997]: I1205 08:58:45.269130 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-config\") pod \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\" (UID: \"d66c0d05-f098-459a-9398-cf7b1ae4edf1\") " Dec 05 08:58:45 crc kubenswrapper[4997]: I1205 08:58:45.317603 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d66c0d05-f098-459a-9398-cf7b1ae4edf1" (UID: "d66c0d05-f098-459a-9398-cf7b1ae4edf1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:58:45 crc kubenswrapper[4997]: I1205 08:58:45.333136 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d66c0d05-f098-459a-9398-cf7b1ae4edf1" (UID: "d66c0d05-f098-459a-9398-cf7b1ae4edf1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:58:45 crc kubenswrapper[4997]: I1205 08:58:45.334565 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-config" (OuterVolumeSpecName: "config") pod "d66c0d05-f098-459a-9398-cf7b1ae4edf1" (UID: "d66c0d05-f098-459a-9398-cf7b1ae4edf1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:58:45 crc kubenswrapper[4997]: I1205 08:58:45.347325 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d66c0d05-f098-459a-9398-cf7b1ae4edf1-kube-api-access-bz668" (OuterVolumeSpecName: "kube-api-access-bz668") pod "d66c0d05-f098-459a-9398-cf7b1ae4edf1" (UID: "d66c0d05-f098-459a-9398-cf7b1ae4edf1"). InnerVolumeSpecName "kube-api-access-bz668". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:58:45 crc kubenswrapper[4997]: I1205 08:58:45.347698 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d66c0d05-f098-459a-9398-cf7b1ae4edf1" (UID: "d66c0d05-f098-459a-9398-cf7b1ae4edf1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:58:45 crc kubenswrapper[4997]: I1205 08:58:45.371668 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:45 crc kubenswrapper[4997]: I1205 08:58:45.371706 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:45 crc kubenswrapper[4997]: I1205 08:58:45.371719 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:45 crc kubenswrapper[4997]: I1205 08:58:45.371756 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d66c0d05-f098-459a-9398-cf7b1ae4edf1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:45 crc kubenswrapper[4997]: I1205 08:58:45.371774 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bz668\" (UniqueName: \"kubernetes.io/projected/d66c0d05-f098-459a-9398-cf7b1ae4edf1-kube-api-access-bz668\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:46 crc kubenswrapper[4997]: I1205 08:58:46.012462 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" event={"ID":"d66c0d05-f098-459a-9398-cf7b1ae4edf1","Type":"ContainerDied","Data":"ea78f35038212f2000fb6b17cdb8e025a9e91643e1e936e1c7e49fe2147783ac"} Dec 05 08:58:46 crc kubenswrapper[4997]: I1205 08:58:46.013726 4997 scope.go:117] "RemoveContainer" containerID="afe2705efd72071cb13623dcd699f8cd8b042f47f2111908cff7f90a10632d8f" Dec 05 08:58:46 crc kubenswrapper[4997]: I1205 08:58:46.012717 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cd6b446c-ncp6v" Dec 05 08:58:46 crc kubenswrapper[4997]: I1205 08:58:46.014971 4997 generic.go:334] "Generic (PLEG): container finished" podID="80e12f51-1310-4be7-a5e9-0f0204dd3e16" containerID="71b0d675a1714a9952113b71215fa56c095b31b5733f7899d24e7901aca2932f" exitCode=0 Dec 05 08:58:46 crc kubenswrapper[4997]: I1205 08:58:46.015077 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrbbh" event={"ID":"80e12f51-1310-4be7-a5e9-0f0204dd3e16","Type":"ContainerDied","Data":"71b0d675a1714a9952113b71215fa56c095b31b5733f7899d24e7901aca2932f"} Dec 05 08:58:46 crc kubenswrapper[4997]: I1205 08:58:46.035159 4997 scope.go:117] "RemoveContainer" containerID="393d8d1969cce96ea0817d1a04cb970a551bb7554378965786e81c0c048d659b" Dec 05 08:58:46 crc kubenswrapper[4997]: I1205 08:58:46.065544 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cd6b446c-ncp6v"] Dec 05 08:58:46 crc kubenswrapper[4997]: I1205 08:58:46.075430 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cd6b446c-ncp6v"] Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.026343 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrbbh" event={"ID":"80e12f51-1310-4be7-a5e9-0f0204dd3e16","Type":"ContainerStarted","Data":"3eec9998a89dbd866cefc2b4ad4a886fd7ad8f7106ff19f51ae988332a5e28d1"} Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.050649 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vrbbh" podStartSLOduration=2.594362321 podStartE2EDuration="5.050632647s" podCreationTimestamp="2025-12-05 08:58:42 +0000 UTC" firstStartedPulling="2025-12-05 08:58:43.97939614 +0000 UTC m=+7424.508303401" lastFinishedPulling="2025-12-05 08:58:46.435666476 +0000 UTC m=+7426.964573727" observedRunningTime="2025-12-05 08:58:47.04331484 +0000 UTC m=+7427.572222111" watchObservedRunningTime="2025-12-05 08:58:47.050632647 +0000 UTC m=+7427.579539928" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.150494 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-wczzn"] Dec 05 08:58:47 crc kubenswrapper[4997]: E1205 08:58:47.151258 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d66c0d05-f098-459a-9398-cf7b1ae4edf1" containerName="dnsmasq-dns" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.151391 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d66c0d05-f098-459a-9398-cf7b1ae4edf1" containerName="dnsmasq-dns" Dec 05 08:58:47 crc kubenswrapper[4997]: E1205 08:58:47.151501 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d66c0d05-f098-459a-9398-cf7b1ae4edf1" containerName="init" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.151581 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d66c0d05-f098-459a-9398-cf7b1ae4edf1" containerName="init" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.151924 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="d66c0d05-f098-459a-9398-cf7b1ae4edf1" containerName="dnsmasq-dns" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.153393 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wczzn" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.164093 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wczzn"] Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.250991 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-0f52-account-create-update-glgr7"] Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.252147 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0f52-account-create-update-glgr7" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.255110 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.266090 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-0f52-account-create-update-glgr7"] Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.314151 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc7d296f-ae2b-4178-abdb-135f3e239e72-operator-scripts\") pod \"cinder-db-create-wczzn\" (UID: \"fc7d296f-ae2b-4178-abdb-135f3e239e72\") " pod="openstack/cinder-db-create-wczzn" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.314478 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtwp8\" (UniqueName: \"kubernetes.io/projected/fc7d296f-ae2b-4178-abdb-135f3e239e72-kube-api-access-rtwp8\") pod \"cinder-db-create-wczzn\" (UID: \"fc7d296f-ae2b-4178-abdb-135f3e239e72\") " pod="openstack/cinder-db-create-wczzn" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.416082 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97m8m\" (UniqueName: \"kubernetes.io/projected/db566dd6-c449-499f-9bf3-457a8657e5da-kube-api-access-97m8m\") pod \"cinder-0f52-account-create-update-glgr7\" (UID: \"db566dd6-c449-499f-9bf3-457a8657e5da\") " pod="openstack/cinder-0f52-account-create-update-glgr7" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.416133 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtwp8\" (UniqueName: \"kubernetes.io/projected/fc7d296f-ae2b-4178-abdb-135f3e239e72-kube-api-access-rtwp8\") pod \"cinder-db-create-wczzn\" (UID: \"fc7d296f-ae2b-4178-abdb-135f3e239e72\") " pod="openstack/cinder-db-create-wczzn" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.416188 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc7d296f-ae2b-4178-abdb-135f3e239e72-operator-scripts\") pod \"cinder-db-create-wczzn\" (UID: \"fc7d296f-ae2b-4178-abdb-135f3e239e72\") " pod="openstack/cinder-db-create-wczzn" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.416274 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db566dd6-c449-499f-9bf3-457a8657e5da-operator-scripts\") pod \"cinder-0f52-account-create-update-glgr7\" (UID: \"db566dd6-c449-499f-9bf3-457a8657e5da\") " pod="openstack/cinder-0f52-account-create-update-glgr7" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.416841 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc7d296f-ae2b-4178-abdb-135f3e239e72-operator-scripts\") pod \"cinder-db-create-wczzn\" (UID: \"fc7d296f-ae2b-4178-abdb-135f3e239e72\") " pod="openstack/cinder-db-create-wczzn" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.448140 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtwp8\" (UniqueName: \"kubernetes.io/projected/fc7d296f-ae2b-4178-abdb-135f3e239e72-kube-api-access-rtwp8\") pod \"cinder-db-create-wczzn\" (UID: \"fc7d296f-ae2b-4178-abdb-135f3e239e72\") " pod="openstack/cinder-db-create-wczzn" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.474648 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wczzn" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.520590 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97m8m\" (UniqueName: \"kubernetes.io/projected/db566dd6-c449-499f-9bf3-457a8657e5da-kube-api-access-97m8m\") pod \"cinder-0f52-account-create-update-glgr7\" (UID: \"db566dd6-c449-499f-9bf3-457a8657e5da\") " pod="openstack/cinder-0f52-account-create-update-glgr7" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.520976 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db566dd6-c449-499f-9bf3-457a8657e5da-operator-scripts\") pod \"cinder-0f52-account-create-update-glgr7\" (UID: \"db566dd6-c449-499f-9bf3-457a8657e5da\") " pod="openstack/cinder-0f52-account-create-update-glgr7" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.521987 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db566dd6-c449-499f-9bf3-457a8657e5da-operator-scripts\") pod \"cinder-0f52-account-create-update-glgr7\" (UID: \"db566dd6-c449-499f-9bf3-457a8657e5da\") " pod="openstack/cinder-0f52-account-create-update-glgr7" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.538057 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97m8m\" (UniqueName: \"kubernetes.io/projected/db566dd6-c449-499f-9bf3-457a8657e5da-kube-api-access-97m8m\") pod \"cinder-0f52-account-create-update-glgr7\" (UID: \"db566dd6-c449-499f-9bf3-457a8657e5da\") " pod="openstack/cinder-0f52-account-create-update-glgr7" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.567819 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0f52-account-create-update-glgr7" Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.781749 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d66c0d05-f098-459a-9398-cf7b1ae4edf1" path="/var/lib/kubelet/pods/d66c0d05-f098-459a-9398-cf7b1ae4edf1/volumes" Dec 05 08:58:47 crc kubenswrapper[4997]: W1205 08:58:47.838919 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc7d296f_ae2b_4178_abdb_135f3e239e72.slice/crio-fb86d80ef059979dfb99af15c0d0ba680008c48c76cfcf901192896e345c95af WatchSource:0}: Error finding container fb86d80ef059979dfb99af15c0d0ba680008c48c76cfcf901192896e345c95af: Status 404 returned error can't find the container with id fb86d80ef059979dfb99af15c0d0ba680008c48c76cfcf901192896e345c95af Dec 05 08:58:47 crc kubenswrapper[4997]: I1205 08:58:47.842090 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wczzn"] Dec 05 08:58:48 crc kubenswrapper[4997]: I1205 08:58:48.043537 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wczzn" event={"ID":"fc7d296f-ae2b-4178-abdb-135f3e239e72","Type":"ContainerStarted","Data":"76f07e30bd6fec0c5ecea375ce071e42b3293940713fe10e1d1753e1585ba432"} Dec 05 08:58:48 crc kubenswrapper[4997]: I1205 08:58:48.043931 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wczzn" event={"ID":"fc7d296f-ae2b-4178-abdb-135f3e239e72","Type":"ContainerStarted","Data":"fb86d80ef059979dfb99af15c0d0ba680008c48c76cfcf901192896e345c95af"} Dec 05 08:58:48 crc kubenswrapper[4997]: I1205 08:58:48.068470 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-wczzn" podStartSLOduration=1.068449466 podStartE2EDuration="1.068449466s" podCreationTimestamp="2025-12-05 08:58:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:58:48.061232383 +0000 UTC m=+7428.590139654" watchObservedRunningTime="2025-12-05 08:58:48.068449466 +0000 UTC m=+7428.597356727" Dec 05 08:58:48 crc kubenswrapper[4997]: W1205 08:58:48.146185 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb566dd6_c449_499f_9bf3_457a8657e5da.slice/crio-43fae88e57773719f5573d3e7deaac3b2a039f19710b361d4dbe096a79a4f356 WatchSource:0}: Error finding container 43fae88e57773719f5573d3e7deaac3b2a039f19710b361d4dbe096a79a4f356: Status 404 returned error can't find the container with id 43fae88e57773719f5573d3e7deaac3b2a039f19710b361d4dbe096a79a4f356 Dec 05 08:58:48 crc kubenswrapper[4997]: I1205 08:58:48.150581 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-0f52-account-create-update-glgr7"] Dec 05 08:58:49 crc kubenswrapper[4997]: I1205 08:58:49.057529 4997 generic.go:334] "Generic (PLEG): container finished" podID="fc7d296f-ae2b-4178-abdb-135f3e239e72" containerID="76f07e30bd6fec0c5ecea375ce071e42b3293940713fe10e1d1753e1585ba432" exitCode=0 Dec 05 08:58:49 crc kubenswrapper[4997]: I1205 08:58:49.057638 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wczzn" event={"ID":"fc7d296f-ae2b-4178-abdb-135f3e239e72","Type":"ContainerDied","Data":"76f07e30bd6fec0c5ecea375ce071e42b3293940713fe10e1d1753e1585ba432"} Dec 05 08:58:49 crc kubenswrapper[4997]: I1205 08:58:49.061076 4997 generic.go:334] "Generic (PLEG): container finished" podID="db566dd6-c449-499f-9bf3-457a8657e5da" containerID="f322072eb980292b4c462cf9b6fe234392255c65125ec6d0c4ea92760528fbe2" exitCode=0 Dec 05 08:58:49 crc kubenswrapper[4997]: I1205 08:58:49.061124 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0f52-account-create-update-glgr7" event={"ID":"db566dd6-c449-499f-9bf3-457a8657e5da","Type":"ContainerDied","Data":"f322072eb980292b4c462cf9b6fe234392255c65125ec6d0c4ea92760528fbe2"} Dec 05 08:58:49 crc kubenswrapper[4997]: I1205 08:58:49.061153 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0f52-account-create-update-glgr7" event={"ID":"db566dd6-c449-499f-9bf3-457a8657e5da","Type":"ContainerStarted","Data":"43fae88e57773719f5573d3e7deaac3b2a039f19710b361d4dbe096a79a4f356"} Dec 05 08:58:49 crc kubenswrapper[4997]: I1205 08:58:49.770658 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:58:49 crc kubenswrapper[4997]: I1205 08:58:49.770772 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:58:49 crc kubenswrapper[4997]: I1205 08:58:49.770809 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 08:58:49 crc kubenswrapper[4997]: I1205 08:58:49.772387 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:58:49 crc kubenswrapper[4997]: I1205 08:58:49.772467 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" gracePeriod=600 Dec 05 08:58:49 crc kubenswrapper[4997]: E1205 08:58:49.901497 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.072649 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" exitCode=0 Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.072706 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b"} Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.072776 4997 scope.go:117] "RemoveContainer" containerID="c7247cb7e5929fb7303801e82d11af6c5303b8ce2c4cf98c220b4780c5e2c9f4" Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.073403 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 08:58:50 crc kubenswrapper[4997]: E1205 08:58:50.073705 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.513397 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wczzn" Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.520191 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0f52-account-create-update-glgr7" Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.714840 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97m8m\" (UniqueName: \"kubernetes.io/projected/db566dd6-c449-499f-9bf3-457a8657e5da-kube-api-access-97m8m\") pod \"db566dd6-c449-499f-9bf3-457a8657e5da\" (UID: \"db566dd6-c449-499f-9bf3-457a8657e5da\") " Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.715121 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc7d296f-ae2b-4178-abdb-135f3e239e72-operator-scripts\") pod \"fc7d296f-ae2b-4178-abdb-135f3e239e72\" (UID: \"fc7d296f-ae2b-4178-abdb-135f3e239e72\") " Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.715944 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtwp8\" (UniqueName: \"kubernetes.io/projected/fc7d296f-ae2b-4178-abdb-135f3e239e72-kube-api-access-rtwp8\") pod \"fc7d296f-ae2b-4178-abdb-135f3e239e72\" (UID: \"fc7d296f-ae2b-4178-abdb-135f3e239e72\") " Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.716002 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db566dd6-c449-499f-9bf3-457a8657e5da-operator-scripts\") pod \"db566dd6-c449-499f-9bf3-457a8657e5da\" (UID: \"db566dd6-c449-499f-9bf3-457a8657e5da\") " Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.716073 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc7d296f-ae2b-4178-abdb-135f3e239e72-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fc7d296f-ae2b-4178-abdb-135f3e239e72" (UID: "fc7d296f-ae2b-4178-abdb-135f3e239e72"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.716528 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db566dd6-c449-499f-9bf3-457a8657e5da-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "db566dd6-c449-499f-9bf3-457a8657e5da" (UID: "db566dd6-c449-499f-9bf3-457a8657e5da"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.716717 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc7d296f-ae2b-4178-abdb-135f3e239e72-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.716736 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db566dd6-c449-499f-9bf3-457a8657e5da-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.720020 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db566dd6-c449-499f-9bf3-457a8657e5da-kube-api-access-97m8m" (OuterVolumeSpecName: "kube-api-access-97m8m") pod "db566dd6-c449-499f-9bf3-457a8657e5da" (UID: "db566dd6-c449-499f-9bf3-457a8657e5da"). InnerVolumeSpecName "kube-api-access-97m8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.722325 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc7d296f-ae2b-4178-abdb-135f3e239e72-kube-api-access-rtwp8" (OuterVolumeSpecName: "kube-api-access-rtwp8") pod "fc7d296f-ae2b-4178-abdb-135f3e239e72" (UID: "fc7d296f-ae2b-4178-abdb-135f3e239e72"). InnerVolumeSpecName "kube-api-access-rtwp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.817791 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtwp8\" (UniqueName: \"kubernetes.io/projected/fc7d296f-ae2b-4178-abdb-135f3e239e72-kube-api-access-rtwp8\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:50 crc kubenswrapper[4997]: I1205 08:58:50.817827 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97m8m\" (UniqueName: \"kubernetes.io/projected/db566dd6-c449-499f-9bf3-457a8657e5da-kube-api-access-97m8m\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:51 crc kubenswrapper[4997]: I1205 08:58:51.088660 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0f52-account-create-update-glgr7" Dec 05 08:58:51 crc kubenswrapper[4997]: I1205 08:58:51.088660 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0f52-account-create-update-glgr7" event={"ID":"db566dd6-c449-499f-9bf3-457a8657e5da","Type":"ContainerDied","Data":"43fae88e57773719f5573d3e7deaac3b2a039f19710b361d4dbe096a79a4f356"} Dec 05 08:58:51 crc kubenswrapper[4997]: I1205 08:58:51.088789 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43fae88e57773719f5573d3e7deaac3b2a039f19710b361d4dbe096a79a4f356" Dec 05 08:58:51 crc kubenswrapper[4997]: I1205 08:58:51.091010 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wczzn" event={"ID":"fc7d296f-ae2b-4178-abdb-135f3e239e72","Type":"ContainerDied","Data":"fb86d80ef059979dfb99af15c0d0ba680008c48c76cfcf901192896e345c95af"} Dec 05 08:58:51 crc kubenswrapper[4997]: I1205 08:58:51.091055 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb86d80ef059979dfb99af15c0d0ba680008c48c76cfcf901192896e345c95af" Dec 05 08:58:51 crc kubenswrapper[4997]: I1205 08:58:51.091078 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wczzn" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.416571 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-8vfdh"] Dec 05 08:58:52 crc kubenswrapper[4997]: E1205 08:58:52.418155 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc7d296f-ae2b-4178-abdb-135f3e239e72" containerName="mariadb-database-create" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.418240 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc7d296f-ae2b-4178-abdb-135f3e239e72" containerName="mariadb-database-create" Dec 05 08:58:52 crc kubenswrapper[4997]: E1205 08:58:52.418323 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db566dd6-c449-499f-9bf3-457a8657e5da" containerName="mariadb-account-create-update" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.418406 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="db566dd6-c449-499f-9bf3-457a8657e5da" containerName="mariadb-account-create-update" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.418641 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc7d296f-ae2b-4178-abdb-135f3e239e72" containerName="mariadb-database-create" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.418723 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="db566dd6-c449-499f-9bf3-457a8657e5da" containerName="mariadb-account-create-update" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.419573 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.423278 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.423515 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-65jbp" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.423524 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.427797 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8vfdh"] Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.433687 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vrbbh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.433739 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vrbbh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.479892 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vrbbh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.554291 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-combined-ca-bundle\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.554544 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bdc0b8fc-c988-410a-9b97-d80cb48754a2-etc-machine-id\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.554743 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-db-sync-config-data\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.554861 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wzhp\" (UniqueName: \"kubernetes.io/projected/bdc0b8fc-c988-410a-9b97-d80cb48754a2-kube-api-access-9wzhp\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.554919 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-scripts\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.554941 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-config-data\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.657192 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-combined-ca-bundle\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.657289 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bdc0b8fc-c988-410a-9b97-d80cb48754a2-etc-machine-id\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.657331 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-db-sync-config-data\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.657373 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wzhp\" (UniqueName: \"kubernetes.io/projected/bdc0b8fc-c988-410a-9b97-d80cb48754a2-kube-api-access-9wzhp\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.657397 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-config-data\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.657413 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-scripts\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.659722 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bdc0b8fc-c988-410a-9b97-d80cb48754a2-etc-machine-id\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.662461 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-db-sync-config-data\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.663689 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-config-data\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.666462 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-combined-ca-bundle\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.667092 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-scripts\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.675035 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wzhp\" (UniqueName: \"kubernetes.io/projected/bdc0b8fc-c988-410a-9b97-d80cb48754a2-kube-api-access-9wzhp\") pod \"cinder-db-sync-8vfdh\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:52 crc kubenswrapper[4997]: I1205 08:58:52.742672 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:58:53 crc kubenswrapper[4997]: I1205 08:58:53.155523 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vrbbh" Dec 05 08:58:53 crc kubenswrapper[4997]: I1205 08:58:53.203983 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vrbbh"] Dec 05 08:58:53 crc kubenswrapper[4997]: I1205 08:58:53.317371 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8vfdh"] Dec 05 08:58:54 crc kubenswrapper[4997]: I1205 08:58:54.125250 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8vfdh" event={"ID":"bdc0b8fc-c988-410a-9b97-d80cb48754a2","Type":"ContainerStarted","Data":"d702bcda42a8e6c2cabb3236c2f9d078169cb5b4c868a7665f230e502315cccd"} Dec 05 08:58:55 crc kubenswrapper[4997]: I1205 08:58:55.138917 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vrbbh" podUID="80e12f51-1310-4be7-a5e9-0f0204dd3e16" containerName="registry-server" containerID="cri-o://3eec9998a89dbd866cefc2b4ad4a886fd7ad8f7106ff19f51ae988332a5e28d1" gracePeriod=2 Dec 05 08:58:55 crc kubenswrapper[4997]: I1205 08:58:55.647824 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrbbh" Dec 05 08:58:55 crc kubenswrapper[4997]: I1205 08:58:55.820688 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80e12f51-1310-4be7-a5e9-0f0204dd3e16-catalog-content\") pod \"80e12f51-1310-4be7-a5e9-0f0204dd3e16\" (UID: \"80e12f51-1310-4be7-a5e9-0f0204dd3e16\") " Dec 05 08:58:55 crc kubenswrapper[4997]: I1205 08:58:55.820830 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjj2h\" (UniqueName: \"kubernetes.io/projected/80e12f51-1310-4be7-a5e9-0f0204dd3e16-kube-api-access-fjj2h\") pod \"80e12f51-1310-4be7-a5e9-0f0204dd3e16\" (UID: \"80e12f51-1310-4be7-a5e9-0f0204dd3e16\") " Dec 05 08:58:55 crc kubenswrapper[4997]: I1205 08:58:55.821197 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80e12f51-1310-4be7-a5e9-0f0204dd3e16-utilities\") pod \"80e12f51-1310-4be7-a5e9-0f0204dd3e16\" (UID: \"80e12f51-1310-4be7-a5e9-0f0204dd3e16\") " Dec 05 08:58:55 crc kubenswrapper[4997]: I1205 08:58:55.822004 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80e12f51-1310-4be7-a5e9-0f0204dd3e16-utilities" (OuterVolumeSpecName: "utilities") pod "80e12f51-1310-4be7-a5e9-0f0204dd3e16" (UID: "80e12f51-1310-4be7-a5e9-0f0204dd3e16"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:58:55 crc kubenswrapper[4997]: I1205 08:58:55.822135 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80e12f51-1310-4be7-a5e9-0f0204dd3e16-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:55 crc kubenswrapper[4997]: I1205 08:58:55.826073 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80e12f51-1310-4be7-a5e9-0f0204dd3e16-kube-api-access-fjj2h" (OuterVolumeSpecName: "kube-api-access-fjj2h") pod "80e12f51-1310-4be7-a5e9-0f0204dd3e16" (UID: "80e12f51-1310-4be7-a5e9-0f0204dd3e16"). InnerVolumeSpecName "kube-api-access-fjj2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:58:55 crc kubenswrapper[4997]: I1205 08:58:55.870138 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80e12f51-1310-4be7-a5e9-0f0204dd3e16-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80e12f51-1310-4be7-a5e9-0f0204dd3e16" (UID: "80e12f51-1310-4be7-a5e9-0f0204dd3e16"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:58:55 crc kubenswrapper[4997]: I1205 08:58:55.923151 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80e12f51-1310-4be7-a5e9-0f0204dd3e16-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:55 crc kubenswrapper[4997]: I1205 08:58:55.923188 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjj2h\" (UniqueName: \"kubernetes.io/projected/80e12f51-1310-4be7-a5e9-0f0204dd3e16-kube-api-access-fjj2h\") on node \"crc\" DevicePath \"\"" Dec 05 08:58:56 crc kubenswrapper[4997]: I1205 08:58:56.148883 4997 generic.go:334] "Generic (PLEG): container finished" podID="80e12f51-1310-4be7-a5e9-0f0204dd3e16" containerID="3eec9998a89dbd866cefc2b4ad4a886fd7ad8f7106ff19f51ae988332a5e28d1" exitCode=0 Dec 05 08:58:56 crc kubenswrapper[4997]: I1205 08:58:56.148930 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrbbh" event={"ID":"80e12f51-1310-4be7-a5e9-0f0204dd3e16","Type":"ContainerDied","Data":"3eec9998a89dbd866cefc2b4ad4a886fd7ad8f7106ff19f51ae988332a5e28d1"} Dec 05 08:58:56 crc kubenswrapper[4997]: I1205 08:58:56.148967 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrbbh" event={"ID":"80e12f51-1310-4be7-a5e9-0f0204dd3e16","Type":"ContainerDied","Data":"97d19b9fd523ba8ba4416f833a2b456660428a0cddaed4e0bd407295be361236"} Dec 05 08:58:56 crc kubenswrapper[4997]: I1205 08:58:56.148985 4997 scope.go:117] "RemoveContainer" containerID="3eec9998a89dbd866cefc2b4ad4a886fd7ad8f7106ff19f51ae988332a5e28d1" Dec 05 08:58:56 crc kubenswrapper[4997]: I1205 08:58:56.148933 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrbbh" Dec 05 08:58:56 crc kubenswrapper[4997]: I1205 08:58:56.178261 4997 scope.go:117] "RemoveContainer" containerID="71b0d675a1714a9952113b71215fa56c095b31b5733f7899d24e7901aca2932f" Dec 05 08:58:56 crc kubenswrapper[4997]: I1205 08:58:56.192055 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vrbbh"] Dec 05 08:58:56 crc kubenswrapper[4997]: I1205 08:58:56.200510 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vrbbh"] Dec 05 08:58:56 crc kubenswrapper[4997]: I1205 08:58:56.224433 4997 scope.go:117] "RemoveContainer" containerID="bd87dfcdc23c29eeabe70f1365a950a69196b5b03280e635ab80bf4211dc7f7e" Dec 05 08:58:56 crc kubenswrapper[4997]: I1205 08:58:56.245728 4997 scope.go:117] "RemoveContainer" containerID="3eec9998a89dbd866cefc2b4ad4a886fd7ad8f7106ff19f51ae988332a5e28d1" Dec 05 08:58:56 crc kubenswrapper[4997]: E1205 08:58:56.248136 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eec9998a89dbd866cefc2b4ad4a886fd7ad8f7106ff19f51ae988332a5e28d1\": container with ID starting with 3eec9998a89dbd866cefc2b4ad4a886fd7ad8f7106ff19f51ae988332a5e28d1 not found: ID does not exist" containerID="3eec9998a89dbd866cefc2b4ad4a886fd7ad8f7106ff19f51ae988332a5e28d1" Dec 05 08:58:56 crc kubenswrapper[4997]: I1205 08:58:56.248185 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eec9998a89dbd866cefc2b4ad4a886fd7ad8f7106ff19f51ae988332a5e28d1"} err="failed to get container status \"3eec9998a89dbd866cefc2b4ad4a886fd7ad8f7106ff19f51ae988332a5e28d1\": rpc error: code = NotFound desc = could not find container \"3eec9998a89dbd866cefc2b4ad4a886fd7ad8f7106ff19f51ae988332a5e28d1\": container with ID starting with 3eec9998a89dbd866cefc2b4ad4a886fd7ad8f7106ff19f51ae988332a5e28d1 not found: ID does not exist" Dec 05 08:58:56 crc kubenswrapper[4997]: I1205 08:58:56.248218 4997 scope.go:117] "RemoveContainer" containerID="71b0d675a1714a9952113b71215fa56c095b31b5733f7899d24e7901aca2932f" Dec 05 08:58:56 crc kubenswrapper[4997]: E1205 08:58:56.248634 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71b0d675a1714a9952113b71215fa56c095b31b5733f7899d24e7901aca2932f\": container with ID starting with 71b0d675a1714a9952113b71215fa56c095b31b5733f7899d24e7901aca2932f not found: ID does not exist" containerID="71b0d675a1714a9952113b71215fa56c095b31b5733f7899d24e7901aca2932f" Dec 05 08:58:56 crc kubenswrapper[4997]: I1205 08:58:56.248675 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71b0d675a1714a9952113b71215fa56c095b31b5733f7899d24e7901aca2932f"} err="failed to get container status \"71b0d675a1714a9952113b71215fa56c095b31b5733f7899d24e7901aca2932f\": rpc error: code = NotFound desc = could not find container \"71b0d675a1714a9952113b71215fa56c095b31b5733f7899d24e7901aca2932f\": container with ID starting with 71b0d675a1714a9952113b71215fa56c095b31b5733f7899d24e7901aca2932f not found: ID does not exist" Dec 05 08:58:56 crc kubenswrapper[4997]: I1205 08:58:56.248693 4997 scope.go:117] "RemoveContainer" containerID="bd87dfcdc23c29eeabe70f1365a950a69196b5b03280e635ab80bf4211dc7f7e" Dec 05 08:58:56 crc kubenswrapper[4997]: E1205 08:58:56.249106 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd87dfcdc23c29eeabe70f1365a950a69196b5b03280e635ab80bf4211dc7f7e\": container with ID starting with bd87dfcdc23c29eeabe70f1365a950a69196b5b03280e635ab80bf4211dc7f7e not found: ID does not exist" containerID="bd87dfcdc23c29eeabe70f1365a950a69196b5b03280e635ab80bf4211dc7f7e" Dec 05 08:58:56 crc kubenswrapper[4997]: I1205 08:58:56.249162 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd87dfcdc23c29eeabe70f1365a950a69196b5b03280e635ab80bf4211dc7f7e"} err="failed to get container status \"bd87dfcdc23c29eeabe70f1365a950a69196b5b03280e635ab80bf4211dc7f7e\": rpc error: code = NotFound desc = could not find container \"bd87dfcdc23c29eeabe70f1365a950a69196b5b03280e635ab80bf4211dc7f7e\": container with ID starting with bd87dfcdc23c29eeabe70f1365a950a69196b5b03280e635ab80bf4211dc7f7e not found: ID does not exist" Dec 05 08:58:56 crc kubenswrapper[4997]: E1205 08:58:56.346314 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80e12f51_1310_4be7_a5e9_0f0204dd3e16.slice/crio-97d19b9fd523ba8ba4416f833a2b456660428a0cddaed4e0bd407295be361236\": RecentStats: unable to find data in memory cache]" Dec 05 08:58:57 crc kubenswrapper[4997]: I1205 08:58:57.759512 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80e12f51-1310-4be7-a5e9-0f0204dd3e16" path="/var/lib/kubelet/pods/80e12f51-1310-4be7-a5e9-0f0204dd3e16/volumes" Dec 05 08:59:04 crc kubenswrapper[4997]: I1205 08:59:04.749879 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 08:59:04 crc kubenswrapper[4997]: E1205 08:59:04.750474 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:59:12 crc kubenswrapper[4997]: I1205 08:59:12.310512 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8vfdh" event={"ID":"bdc0b8fc-c988-410a-9b97-d80cb48754a2","Type":"ContainerStarted","Data":"2f2cb21e9120ff0d3ccb1e72be0184d31016379bca916dd03c6f31891722c88b"} Dec 05 08:59:12 crc kubenswrapper[4997]: I1205 08:59:12.327019 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-8vfdh" podStartSLOduration=2.24830626 podStartE2EDuration="20.326982016s" podCreationTimestamp="2025-12-05 08:58:52 +0000 UTC" firstStartedPulling="2025-12-05 08:58:53.31990494 +0000 UTC m=+7433.848812201" lastFinishedPulling="2025-12-05 08:59:11.398580696 +0000 UTC m=+7451.927487957" observedRunningTime="2025-12-05 08:59:12.32489231 +0000 UTC m=+7452.853799581" watchObservedRunningTime="2025-12-05 08:59:12.326982016 +0000 UTC m=+7452.855889277" Dec 05 08:59:14 crc kubenswrapper[4997]: I1205 08:59:14.331853 4997 generic.go:334] "Generic (PLEG): container finished" podID="bdc0b8fc-c988-410a-9b97-d80cb48754a2" containerID="2f2cb21e9120ff0d3ccb1e72be0184d31016379bca916dd03c6f31891722c88b" exitCode=0 Dec 05 08:59:14 crc kubenswrapper[4997]: I1205 08:59:14.331976 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8vfdh" event={"ID":"bdc0b8fc-c988-410a-9b97-d80cb48754a2","Type":"ContainerDied","Data":"2f2cb21e9120ff0d3ccb1e72be0184d31016379bca916dd03c6f31891722c88b"} Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.670054 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.708851 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-config-data\") pod \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.708902 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wzhp\" (UniqueName: \"kubernetes.io/projected/bdc0b8fc-c988-410a-9b97-d80cb48754a2-kube-api-access-9wzhp\") pod \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.708946 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-combined-ca-bundle\") pod \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.709007 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-db-sync-config-data\") pod \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.709037 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-scripts\") pod \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.709056 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bdc0b8fc-c988-410a-9b97-d80cb48754a2-etc-machine-id\") pod \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\" (UID: \"bdc0b8fc-c988-410a-9b97-d80cb48754a2\") " Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.709509 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bdc0b8fc-c988-410a-9b97-d80cb48754a2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bdc0b8fc-c988-410a-9b97-d80cb48754a2" (UID: "bdc0b8fc-c988-410a-9b97-d80cb48754a2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.715658 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "bdc0b8fc-c988-410a-9b97-d80cb48754a2" (UID: "bdc0b8fc-c988-410a-9b97-d80cb48754a2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.715730 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-scripts" (OuterVolumeSpecName: "scripts") pod "bdc0b8fc-c988-410a-9b97-d80cb48754a2" (UID: "bdc0b8fc-c988-410a-9b97-d80cb48754a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.715731 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdc0b8fc-c988-410a-9b97-d80cb48754a2-kube-api-access-9wzhp" (OuterVolumeSpecName: "kube-api-access-9wzhp") pod "bdc0b8fc-c988-410a-9b97-d80cb48754a2" (UID: "bdc0b8fc-c988-410a-9b97-d80cb48754a2"). InnerVolumeSpecName "kube-api-access-9wzhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.742669 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bdc0b8fc-c988-410a-9b97-d80cb48754a2" (UID: "bdc0b8fc-c988-410a-9b97-d80cb48754a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.749551 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 08:59:15 crc kubenswrapper[4997]: E1205 08:59:15.749893 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.758307 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-config-data" (OuterVolumeSpecName: "config-data") pod "bdc0b8fc-c988-410a-9b97-d80cb48754a2" (UID: "bdc0b8fc-c988-410a-9b97-d80cb48754a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.811431 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.812629 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wzhp\" (UniqueName: \"kubernetes.io/projected/bdc0b8fc-c988-410a-9b97-d80cb48754a2-kube-api-access-9wzhp\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.812650 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.812660 4997 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.812669 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdc0b8fc-c988-410a-9b97-d80cb48754a2-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:15 crc kubenswrapper[4997]: I1205 08:59:15.812677 4997 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bdc0b8fc-c988-410a-9b97-d80cb48754a2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.350310 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8vfdh" event={"ID":"bdc0b8fc-c988-410a-9b97-d80cb48754a2","Type":"ContainerDied","Data":"d702bcda42a8e6c2cabb3236c2f9d078169cb5b4c868a7665f230e502315cccd"} Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.350382 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8vfdh" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.350389 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d702bcda42a8e6c2cabb3236c2f9d078169cb5b4c868a7665f230e502315cccd" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.641903 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d69fcbd89-g6w6k"] Dec 05 08:59:16 crc kubenswrapper[4997]: E1205 08:59:16.644687 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80e12f51-1310-4be7-a5e9-0f0204dd3e16" containerName="registry-server" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.644725 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="80e12f51-1310-4be7-a5e9-0f0204dd3e16" containerName="registry-server" Dec 05 08:59:16 crc kubenswrapper[4997]: E1205 08:59:16.644745 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80e12f51-1310-4be7-a5e9-0f0204dd3e16" containerName="extract-utilities" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.644754 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="80e12f51-1310-4be7-a5e9-0f0204dd3e16" containerName="extract-utilities" Dec 05 08:59:16 crc kubenswrapper[4997]: E1205 08:59:16.644786 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80e12f51-1310-4be7-a5e9-0f0204dd3e16" containerName="extract-content" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.644795 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="80e12f51-1310-4be7-a5e9-0f0204dd3e16" containerName="extract-content" Dec 05 08:59:16 crc kubenswrapper[4997]: E1205 08:59:16.644815 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdc0b8fc-c988-410a-9b97-d80cb48754a2" containerName="cinder-db-sync" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.644823 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdc0b8fc-c988-410a-9b97-d80cb48754a2" containerName="cinder-db-sync" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.645086 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdc0b8fc-c988-410a-9b97-d80cb48754a2" containerName="cinder-db-sync" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.645101 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="80e12f51-1310-4be7-a5e9-0f0204dd3e16" containerName="registry-server" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.646805 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.672029 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d69fcbd89-g6w6k"] Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.731549 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-ovsdbserver-sb\") pod \"dnsmasq-dns-d69fcbd89-g6w6k\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.731591 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-config\") pod \"dnsmasq-dns-d69fcbd89-g6w6k\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.731822 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7t4j\" (UniqueName: \"kubernetes.io/projected/127f17b4-1636-4e11-932b-a15383b8132d-kube-api-access-h7t4j\") pod \"dnsmasq-dns-d69fcbd89-g6w6k\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.731956 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-dns-svc\") pod \"dnsmasq-dns-d69fcbd89-g6w6k\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.732026 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-ovsdbserver-nb\") pod \"dnsmasq-dns-d69fcbd89-g6w6k\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.807158 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.808766 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.815051 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.815218 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.815256 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.815274 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-65jbp" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.823560 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.836483 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rhbw\" (UniqueName: \"kubernetes.io/projected/af352d23-3b54-41ce-b382-73611152bf30-kube-api-access-6rhbw\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.836810 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7t4j\" (UniqueName: \"kubernetes.io/projected/127f17b4-1636-4e11-932b-a15383b8132d-kube-api-access-h7t4j\") pod \"dnsmasq-dns-d69fcbd89-g6w6k\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.836873 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-scripts\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.837219 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-config-data\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.837307 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-dns-svc\") pod \"dnsmasq-dns-d69fcbd89-g6w6k\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.837426 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af352d23-3b54-41ce-b382-73611152bf30-logs\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.837504 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-ovsdbserver-nb\") pod \"dnsmasq-dns-d69fcbd89-g6w6k\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.837591 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.837662 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-config-data-custom\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.837807 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-ovsdbserver-sb\") pod \"dnsmasq-dns-d69fcbd89-g6w6k\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.837848 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-config\") pod \"dnsmasq-dns-d69fcbd89-g6w6k\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.837925 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af352d23-3b54-41ce-b382-73611152bf30-etc-machine-id\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.840357 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-ovsdbserver-nb\") pod \"dnsmasq-dns-d69fcbd89-g6w6k\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.840394 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-ovsdbserver-sb\") pod \"dnsmasq-dns-d69fcbd89-g6w6k\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.840663 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-dns-svc\") pod \"dnsmasq-dns-d69fcbd89-g6w6k\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.841065 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-config\") pod \"dnsmasq-dns-d69fcbd89-g6w6k\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.854644 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7t4j\" (UniqueName: \"kubernetes.io/projected/127f17b4-1636-4e11-932b-a15383b8132d-kube-api-access-h7t4j\") pod \"dnsmasq-dns-d69fcbd89-g6w6k\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.939349 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.939680 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-config-data-custom\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.939733 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af352d23-3b54-41ce-b382-73611152bf30-etc-machine-id\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.939781 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rhbw\" (UniqueName: \"kubernetes.io/projected/af352d23-3b54-41ce-b382-73611152bf30-kube-api-access-6rhbw\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.939849 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-scripts\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.940225 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-config-data\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.940663 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af352d23-3b54-41ce-b382-73611152bf30-logs\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.939894 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af352d23-3b54-41ce-b382-73611152bf30-etc-machine-id\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.941001 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af352d23-3b54-41ce-b382-73611152bf30-logs\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.943734 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-scripts\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.944410 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-config-data\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.948045 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-config-data-custom\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.957243 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rhbw\" (UniqueName: \"kubernetes.io/projected/af352d23-3b54-41ce-b382-73611152bf30-kube-api-access-6rhbw\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.960107 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " pod="openstack/cinder-api-0" Dec 05 08:59:16 crc kubenswrapper[4997]: I1205 08:59:16.984740 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:17 crc kubenswrapper[4997]: I1205 08:59:17.125027 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 08:59:17 crc kubenswrapper[4997]: I1205 08:59:17.285432 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d69fcbd89-g6w6k"] Dec 05 08:59:17 crc kubenswrapper[4997]: I1205 08:59:17.387265 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" event={"ID":"127f17b4-1636-4e11-932b-a15383b8132d","Type":"ContainerStarted","Data":"c930ef5a7239f2f29d8f16ba58def8958d612c8e860a71eb0e7c5cf6bedeac34"} Dec 05 08:59:17 crc kubenswrapper[4997]: I1205 08:59:17.393629 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:59:17 crc kubenswrapper[4997]: W1205 08:59:17.400822 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf352d23_3b54_41ce_b382_73611152bf30.slice/crio-da6506909722af71f54899090e6cba69a4ff8a1c87d8f1c1c616bf51ee990af5 WatchSource:0}: Error finding container da6506909722af71f54899090e6cba69a4ff8a1c87d8f1c1c616bf51ee990af5: Status 404 returned error can't find the container with id da6506909722af71f54899090e6cba69a4ff8a1c87d8f1c1c616bf51ee990af5 Dec 05 08:59:18 crc kubenswrapper[4997]: I1205 08:59:18.450206 4997 generic.go:334] "Generic (PLEG): container finished" podID="127f17b4-1636-4e11-932b-a15383b8132d" containerID="584351b11a83e68616611079acabefdbefdeb7b783697ecd7d8b26b8d5cc0c7a" exitCode=0 Dec 05 08:59:18 crc kubenswrapper[4997]: I1205 08:59:18.450438 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" event={"ID":"127f17b4-1636-4e11-932b-a15383b8132d","Type":"ContainerDied","Data":"584351b11a83e68616611079acabefdbefdeb7b783697ecd7d8b26b8d5cc0c7a"} Dec 05 08:59:18 crc kubenswrapper[4997]: I1205 08:59:18.453987 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"af352d23-3b54-41ce-b382-73611152bf30","Type":"ContainerStarted","Data":"cbe02905034056ee64697ed4e7969882b6b240d233968b88185095818be54cf0"} Dec 05 08:59:18 crc kubenswrapper[4997]: I1205 08:59:18.454044 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"af352d23-3b54-41ce-b382-73611152bf30","Type":"ContainerStarted","Data":"da6506909722af71f54899090e6cba69a4ff8a1c87d8f1c1c616bf51ee990af5"} Dec 05 08:59:19 crc kubenswrapper[4997]: I1205 08:59:19.466948 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" event={"ID":"127f17b4-1636-4e11-932b-a15383b8132d","Type":"ContainerStarted","Data":"44ccca432d52c528cdc51a7e3b204d1fa8b317d36bc087e12bdf15ab4057b21d"} Dec 05 08:59:19 crc kubenswrapper[4997]: I1205 08:59:19.468732 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:19 crc kubenswrapper[4997]: I1205 08:59:19.471174 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"af352d23-3b54-41ce-b382-73611152bf30","Type":"ContainerStarted","Data":"7e61220eb02cd00838d606d86785fd88e5260eb8bd3d4f005cc85dfe57840c64"} Dec 05 08:59:19 crc kubenswrapper[4997]: I1205 08:59:19.471395 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 05 08:59:19 crc kubenswrapper[4997]: I1205 08:59:19.489978 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" podStartSLOduration=3.48996287 podStartE2EDuration="3.48996287s" podCreationTimestamp="2025-12-05 08:59:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:59:19.489400235 +0000 UTC m=+7460.018307496" watchObservedRunningTime="2025-12-05 08:59:19.48996287 +0000 UTC m=+7460.018870121" Dec 05 08:59:19 crc kubenswrapper[4997]: I1205 08:59:19.514057 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.514034919 podStartE2EDuration="3.514034919s" podCreationTimestamp="2025-12-05 08:59:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:59:19.511849211 +0000 UTC m=+7460.040756512" watchObservedRunningTime="2025-12-05 08:59:19.514034919 +0000 UTC m=+7460.042942180" Dec 05 08:59:26 crc kubenswrapper[4997]: I1205 08:59:26.749692 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 08:59:26 crc kubenswrapper[4997]: E1205 08:59:26.750797 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:59:26 crc kubenswrapper[4997]: I1205 08:59:26.985812 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.087483 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-694f8cd549-6hns9"] Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.087792 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-694f8cd549-6hns9" podUID="7a0cf64f-5798-4682-8fde-9998c6df1e96" containerName="dnsmasq-dns" containerID="cri-o://07329078d555356e3e2a6ca4b0f9e265bcc6dd5ae02938df058c4eedba1c3d9c" gracePeriod=10 Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.632588 4997 generic.go:334] "Generic (PLEG): container finished" podID="7a0cf64f-5798-4682-8fde-9998c6df1e96" containerID="07329078d555356e3e2a6ca4b0f9e265bcc6dd5ae02938df058c4eedba1c3d9c" exitCode=0 Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.634582 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-694f8cd549-6hns9" event={"ID":"7a0cf64f-5798-4682-8fde-9998c6df1e96","Type":"ContainerDied","Data":"07329078d555356e3e2a6ca4b0f9e265bcc6dd5ae02938df058c4eedba1c3d9c"} Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.702734 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.852180 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-config\") pod \"7a0cf64f-5798-4682-8fde-9998c6df1e96\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.852254 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-dns-svc\") pod \"7a0cf64f-5798-4682-8fde-9998c6df1e96\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.852277 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-ovsdbserver-nb\") pod \"7a0cf64f-5798-4682-8fde-9998c6df1e96\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.852367 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrhsw\" (UniqueName: \"kubernetes.io/projected/7a0cf64f-5798-4682-8fde-9998c6df1e96-kube-api-access-hrhsw\") pod \"7a0cf64f-5798-4682-8fde-9998c6df1e96\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.852477 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-ovsdbserver-sb\") pod \"7a0cf64f-5798-4682-8fde-9998c6df1e96\" (UID: \"7a0cf64f-5798-4682-8fde-9998c6df1e96\") " Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.858810 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a0cf64f-5798-4682-8fde-9998c6df1e96-kube-api-access-hrhsw" (OuterVolumeSpecName: "kube-api-access-hrhsw") pod "7a0cf64f-5798-4682-8fde-9998c6df1e96" (UID: "7a0cf64f-5798-4682-8fde-9998c6df1e96"). InnerVolumeSpecName "kube-api-access-hrhsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.924990 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7a0cf64f-5798-4682-8fde-9998c6df1e96" (UID: "7a0cf64f-5798-4682-8fde-9998c6df1e96"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.936835 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7a0cf64f-5798-4682-8fde-9998c6df1e96" (UID: "7a0cf64f-5798-4682-8fde-9998c6df1e96"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.953944 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.953978 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrhsw\" (UniqueName: \"kubernetes.io/projected/7a0cf64f-5798-4682-8fde-9998c6df1e96-kube-api-access-hrhsw\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.953989 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.956040 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7a0cf64f-5798-4682-8fde-9998c6df1e96" (UID: "7a0cf64f-5798-4682-8fde-9998c6df1e96"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:59:27 crc kubenswrapper[4997]: I1205 08:59:27.966247 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-config" (OuterVolumeSpecName: "config") pod "7a0cf64f-5798-4682-8fde-9998c6df1e96" (UID: "7a0cf64f-5798-4682-8fde-9998c6df1e96"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.055335 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.055380 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a0cf64f-5798-4682-8fde-9998c6df1e96-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.645534 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-694f8cd549-6hns9" event={"ID":"7a0cf64f-5798-4682-8fde-9998c6df1e96","Type":"ContainerDied","Data":"8db8dd447b25eb3ff0df4efdf53d0168fa6af4b5588292be8b164fd60b3779fb"} Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.645590 4997 scope.go:117] "RemoveContainer" containerID="07329078d555356e3e2a6ca4b0f9e265bcc6dd5ae02938df058c4eedba1c3d9c" Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.645632 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-694f8cd549-6hns9" Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.678231 4997 scope.go:117] "RemoveContainer" containerID="5aaf53ba0b9812f2a9b38ba9a9b887bcd626f2505649d98788f866fbbe7ed9a5" Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.690057 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-694f8cd549-6hns9"] Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.712690 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-694f8cd549-6hns9"] Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.723673 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.723886 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="187e8113-1011-4622-bd2b-a582a9450a6f" containerName="nova-cell0-conductor-conductor" containerID="cri-o://0c4171ada7a2365ff05d143c62c6b5d2ff029f4808c2fc08d95ee1888d1b7557" gracePeriod=30 Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.745100 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.745749 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="58e1c50e-f246-4d92-ae8c-857da4eb1c10" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://0125569bd24894dec6abb10d0d93c20cd62c994116636351ae75d86287e14989" gracePeriod=30 Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.760850 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.761127 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" containerName="nova-metadata-log" containerID="cri-o://315590eb9eb2ef36418367ff8e38e1d476a35163132ddf8c6324fa1e31586887" gracePeriod=30 Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.761730 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" containerName="nova-metadata-metadata" containerID="cri-o://ed33e13d3a1086153e0da6846482e061e4d7c00955733b7c2c0ac7b1a7662040" gracePeriod=30 Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.779714 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.780008 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8c366489-0d4c-48e1-99be-5b94881cd468" containerName="nova-api-log" containerID="cri-o://a621dd45aff6ca71d2c373c426d8b890b899a53bd037c5203583e71b917ba244" gracePeriod=30 Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.780533 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8c366489-0d4c-48e1-99be-5b94881cd468" containerName="nova-api-api" containerID="cri-o://576ff2a343d43ad6e34c58d476cc43aab9729c86badcf63506fdae3033d099be" gracePeriod=30 Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.786896 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.787094 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="700689bf-8f80-457b-8daf-3dfde8871759" containerName="nova-scheduler-scheduler" containerID="cri-o://5e1515944c4d5662ce252478df8e925f34222dd1e61fb08a7663defebb4255f3" gracePeriod=30 Dec 05 08:59:28 crc kubenswrapper[4997]: I1205 08:59:28.884302 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="58e1c50e-f246-4d92-ae8c-857da4eb1c10" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.78:6080/vnc_lite.html\": dial tcp 10.217.1.78:6080: connect: connection refused" Dec 05 08:59:29 crc kubenswrapper[4997]: E1205 08:59:29.142394 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5e1515944c4d5662ce252478df8e925f34222dd1e61fb08a7663defebb4255f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 08:59:29 crc kubenswrapper[4997]: E1205 08:59:29.143804 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5e1515944c4d5662ce252478df8e925f34222dd1e61fb08a7663defebb4255f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 08:59:29 crc kubenswrapper[4997]: E1205 08:59:29.146330 4997 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5e1515944c4d5662ce252478df8e925f34222dd1e61fb08a7663defebb4255f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 08:59:29 crc kubenswrapper[4997]: E1205 08:59:29.146384 4997 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="700689bf-8f80-457b-8daf-3dfde8871759" containerName="nova-scheduler-scheduler" Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.449866 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.688171 4997 generic.go:334] "Generic (PLEG): container finished" podID="58e1c50e-f246-4d92-ae8c-857da4eb1c10" containerID="0125569bd24894dec6abb10d0d93c20cd62c994116636351ae75d86287e14989" exitCode=0 Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.688640 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"58e1c50e-f246-4d92-ae8c-857da4eb1c10","Type":"ContainerDied","Data":"0125569bd24894dec6abb10d0d93c20cd62c994116636351ae75d86287e14989"} Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.731101 4997 generic.go:334] "Generic (PLEG): container finished" podID="8c366489-0d4c-48e1-99be-5b94881cd468" containerID="a621dd45aff6ca71d2c373c426d8b890b899a53bd037c5203583e71b917ba244" exitCode=143 Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.731241 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c366489-0d4c-48e1-99be-5b94881cd468","Type":"ContainerDied","Data":"a621dd45aff6ca71d2c373c426d8b890b899a53bd037c5203583e71b917ba244"} Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.747604 4997 generic.go:334] "Generic (PLEG): container finished" podID="5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" containerID="315590eb9eb2ef36418367ff8e38e1d476a35163132ddf8c6324fa1e31586887" exitCode=143 Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.747701 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733","Type":"ContainerDied","Data":"315590eb9eb2ef36418367ff8e38e1d476a35163132ddf8c6324fa1e31586887"} Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.781911 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a0cf64f-5798-4682-8fde-9998c6df1e96" path="/var/lib/kubelet/pods/7a0cf64f-5798-4682-8fde-9998c6df1e96/volumes" Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.848742 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.885228 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zfr7\" (UniqueName: \"kubernetes.io/projected/58e1c50e-f246-4d92-ae8c-857da4eb1c10-kube-api-access-6zfr7\") pod \"58e1c50e-f246-4d92-ae8c-857da4eb1c10\" (UID: \"58e1c50e-f246-4d92-ae8c-857da4eb1c10\") " Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.885349 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58e1c50e-f246-4d92-ae8c-857da4eb1c10-config-data\") pod \"58e1c50e-f246-4d92-ae8c-857da4eb1c10\" (UID: \"58e1c50e-f246-4d92-ae8c-857da4eb1c10\") " Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.885388 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e1c50e-f246-4d92-ae8c-857da4eb1c10-combined-ca-bundle\") pod \"58e1c50e-f246-4d92-ae8c-857da4eb1c10\" (UID: \"58e1c50e-f246-4d92-ae8c-857da4eb1c10\") " Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.898839 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58e1c50e-f246-4d92-ae8c-857da4eb1c10-kube-api-access-6zfr7" (OuterVolumeSpecName: "kube-api-access-6zfr7") pod "58e1c50e-f246-4d92-ae8c-857da4eb1c10" (UID: "58e1c50e-f246-4d92-ae8c-857da4eb1c10"). InnerVolumeSpecName "kube-api-access-6zfr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.916501 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58e1c50e-f246-4d92-ae8c-857da4eb1c10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58e1c50e-f246-4d92-ae8c-857da4eb1c10" (UID: "58e1c50e-f246-4d92-ae8c-857da4eb1c10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.919453 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58e1c50e-f246-4d92-ae8c-857da4eb1c10-config-data" (OuterVolumeSpecName: "config-data") pod "58e1c50e-f246-4d92-ae8c-857da4eb1c10" (UID: "58e1c50e-f246-4d92-ae8c-857da4eb1c10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.987636 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58e1c50e-f246-4d92-ae8c-857da4eb1c10-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.987666 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e1c50e-f246-4d92-ae8c-857da4eb1c10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:29 crc kubenswrapper[4997]: I1205 08:59:29.987676 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zfr7\" (UniqueName: \"kubernetes.io/projected/58e1c50e-f246-4d92-ae8c-857da4eb1c10-kube-api-access-6zfr7\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:30 crc kubenswrapper[4997]: I1205 08:59:30.757667 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"58e1c50e-f246-4d92-ae8c-857da4eb1c10","Type":"ContainerDied","Data":"a5b80aa9aabc8d3d5cca45a8140c21cdb883930555095ebe0daf78b36f64b324"} Dec 05 08:59:30 crc kubenswrapper[4997]: I1205 08:59:30.757928 4997 scope.go:117] "RemoveContainer" containerID="0125569bd24894dec6abb10d0d93c20cd62c994116636351ae75d86287e14989" Dec 05 08:59:30 crc kubenswrapper[4997]: I1205 08:59:30.758023 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:59:30 crc kubenswrapper[4997]: I1205 08:59:30.790171 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:59:30 crc kubenswrapper[4997]: I1205 08:59:30.806195 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:59:30 crc kubenswrapper[4997]: I1205 08:59:30.828048 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:59:30 crc kubenswrapper[4997]: E1205 08:59:30.828849 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a0cf64f-5798-4682-8fde-9998c6df1e96" containerName="dnsmasq-dns" Dec 05 08:59:30 crc kubenswrapper[4997]: I1205 08:59:30.828880 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a0cf64f-5798-4682-8fde-9998c6df1e96" containerName="dnsmasq-dns" Dec 05 08:59:30 crc kubenswrapper[4997]: E1205 08:59:30.828892 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58e1c50e-f246-4d92-ae8c-857da4eb1c10" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 08:59:30 crc kubenswrapper[4997]: I1205 08:59:30.828898 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="58e1c50e-f246-4d92-ae8c-857da4eb1c10" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 08:59:30 crc kubenswrapper[4997]: E1205 08:59:30.828934 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a0cf64f-5798-4682-8fde-9998c6df1e96" containerName="init" Dec 05 08:59:30 crc kubenswrapper[4997]: I1205 08:59:30.828941 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a0cf64f-5798-4682-8fde-9998c6df1e96" containerName="init" Dec 05 08:59:30 crc kubenswrapper[4997]: I1205 08:59:30.829106 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="58e1c50e-f246-4d92-ae8c-857da4eb1c10" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 08:59:30 crc kubenswrapper[4997]: I1205 08:59:30.829121 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a0cf64f-5798-4682-8fde-9998c6df1e96" containerName="dnsmasq-dns" Dec 05 08:59:30 crc kubenswrapper[4997]: I1205 08:59:30.829738 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:59:30 crc kubenswrapper[4997]: I1205 08:59:30.832050 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 05 08:59:30 crc kubenswrapper[4997]: I1205 08:59:30.850950 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:59:30 crc kubenswrapper[4997]: I1205 08:59:30.902654 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d30eaf1-eb47-46c2-b87a-18fe74b003e9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4d30eaf1-eb47-46c2-b87a-18fe74b003e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:59:30 crc kubenswrapper[4997]: I1205 08:59:30.903400 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rszmj\" (UniqueName: \"kubernetes.io/projected/4d30eaf1-eb47-46c2-b87a-18fe74b003e9-kube-api-access-rszmj\") pod \"nova-cell1-novncproxy-0\" (UID: \"4d30eaf1-eb47-46c2-b87a-18fe74b003e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:59:30 crc kubenswrapper[4997]: I1205 08:59:30.903685 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d30eaf1-eb47-46c2-b87a-18fe74b003e9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4d30eaf1-eb47-46c2-b87a-18fe74b003e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:59:31 crc kubenswrapper[4997]: I1205 08:59:31.005161 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d30eaf1-eb47-46c2-b87a-18fe74b003e9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4d30eaf1-eb47-46c2-b87a-18fe74b003e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:59:31 crc kubenswrapper[4997]: I1205 08:59:31.005275 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d30eaf1-eb47-46c2-b87a-18fe74b003e9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4d30eaf1-eb47-46c2-b87a-18fe74b003e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:59:31 crc kubenswrapper[4997]: I1205 08:59:31.005308 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rszmj\" (UniqueName: \"kubernetes.io/projected/4d30eaf1-eb47-46c2-b87a-18fe74b003e9-kube-api-access-rszmj\") pod \"nova-cell1-novncproxy-0\" (UID: \"4d30eaf1-eb47-46c2-b87a-18fe74b003e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:59:31 crc kubenswrapper[4997]: I1205 08:59:31.019456 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d30eaf1-eb47-46c2-b87a-18fe74b003e9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4d30eaf1-eb47-46c2-b87a-18fe74b003e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:59:31 crc kubenswrapper[4997]: I1205 08:59:31.033798 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rszmj\" (UniqueName: \"kubernetes.io/projected/4d30eaf1-eb47-46c2-b87a-18fe74b003e9-kube-api-access-rszmj\") pod \"nova-cell1-novncproxy-0\" (UID: \"4d30eaf1-eb47-46c2-b87a-18fe74b003e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:59:31 crc kubenswrapper[4997]: I1205 08:59:31.034134 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d30eaf1-eb47-46c2-b87a-18fe74b003e9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4d30eaf1-eb47-46c2-b87a-18fe74b003e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:59:31 crc kubenswrapper[4997]: I1205 08:59:31.153491 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:59:31 crc kubenswrapper[4997]: I1205 08:59:31.625817 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:59:31 crc kubenswrapper[4997]: I1205 08:59:31.772441 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58e1c50e-f246-4d92-ae8c-857da4eb1c10" path="/var/lib/kubelet/pods/58e1c50e-f246-4d92-ae8c-857da4eb1c10/volumes" Dec 05 08:59:31 crc kubenswrapper[4997]: I1205 08:59:31.775210 4997 generic.go:334] "Generic (PLEG): container finished" podID="187e8113-1011-4622-bd2b-a582a9450a6f" containerID="0c4171ada7a2365ff05d143c62c6b5d2ff029f4808c2fc08d95ee1888d1b7557" exitCode=0 Dec 05 08:59:31 crc kubenswrapper[4997]: I1205 08:59:31.775259 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"187e8113-1011-4622-bd2b-a582a9450a6f","Type":"ContainerDied","Data":"0c4171ada7a2365ff05d143c62c6b5d2ff029f4808c2fc08d95ee1888d1b7557"} Dec 05 08:59:31 crc kubenswrapper[4997]: I1205 08:59:31.777002 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4d30eaf1-eb47-46c2-b87a-18fe74b003e9","Type":"ContainerStarted","Data":"f30556bee388a7d1073fb82c98e0f01270b97b390e35a0afca1eb21a2d327346"} Dec 05 08:59:31 crc kubenswrapper[4997]: I1205 08:59:31.972172 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.023360 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/187e8113-1011-4622-bd2b-a582a9450a6f-config-data\") pod \"187e8113-1011-4622-bd2b-a582a9450a6f\" (UID: \"187e8113-1011-4622-bd2b-a582a9450a6f\") " Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.023457 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/187e8113-1011-4622-bd2b-a582a9450a6f-combined-ca-bundle\") pod \"187e8113-1011-4622-bd2b-a582a9450a6f\" (UID: \"187e8113-1011-4622-bd2b-a582a9450a6f\") " Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.023501 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8l25k\" (UniqueName: \"kubernetes.io/projected/187e8113-1011-4622-bd2b-a582a9450a6f-kube-api-access-8l25k\") pod \"187e8113-1011-4622-bd2b-a582a9450a6f\" (UID: \"187e8113-1011-4622-bd2b-a582a9450a6f\") " Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.030652 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/187e8113-1011-4622-bd2b-a582a9450a6f-kube-api-access-8l25k" (OuterVolumeSpecName: "kube-api-access-8l25k") pod "187e8113-1011-4622-bd2b-a582a9450a6f" (UID: "187e8113-1011-4622-bd2b-a582a9450a6f"). InnerVolumeSpecName "kube-api-access-8l25k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.039386 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.039588 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="cadd4122-9bc4-4e70-97d4-7e062feed5de" containerName="nova-cell1-conductor-conductor" containerID="cri-o://45f674f05294bc21c57daee4316ef766b40f5822bb721afd5acc8925ff46098d" gracePeriod=30 Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.052805 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/187e8113-1011-4622-bd2b-a582a9450a6f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "187e8113-1011-4622-bd2b-a582a9450a6f" (UID: "187e8113-1011-4622-bd2b-a582a9450a6f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.072086 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/187e8113-1011-4622-bd2b-a582a9450a6f-config-data" (OuterVolumeSpecName: "config-data") pod "187e8113-1011-4622-bd2b-a582a9450a6f" (UID: "187e8113-1011-4622-bd2b-a582a9450a6f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.127893 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/187e8113-1011-4622-bd2b-a582a9450a6f-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.127929 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/187e8113-1011-4622-bd2b-a582a9450a6f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.127944 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8l25k\" (UniqueName: \"kubernetes.io/projected/187e8113-1011-4622-bd2b-a582a9450a6f-kube-api-access-8l25k\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.178017 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.87:8775/\": read tcp 10.217.0.2:49310->10.217.1.87:8775: read: connection reset by peer" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.178069 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.87:8775/\": read tcp 10.217.0.2:49318->10.217.1.87:8775: read: connection reset by peer" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.456793 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.533232 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c366489-0d4c-48e1-99be-5b94881cd468-logs\") pod \"8c366489-0d4c-48e1-99be-5b94881cd468\" (UID: \"8c366489-0d4c-48e1-99be-5b94881cd468\") " Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.533563 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c366489-0d4c-48e1-99be-5b94881cd468-combined-ca-bundle\") pod \"8c366489-0d4c-48e1-99be-5b94881cd468\" (UID: \"8c366489-0d4c-48e1-99be-5b94881cd468\") " Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.533768 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c366489-0d4c-48e1-99be-5b94881cd468-config-data\") pod \"8c366489-0d4c-48e1-99be-5b94881cd468\" (UID: \"8c366489-0d4c-48e1-99be-5b94881cd468\") " Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.534020 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54j9h\" (UniqueName: \"kubernetes.io/projected/8c366489-0d4c-48e1-99be-5b94881cd468-kube-api-access-54j9h\") pod \"8c366489-0d4c-48e1-99be-5b94881cd468\" (UID: \"8c366489-0d4c-48e1-99be-5b94881cd468\") " Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.542490 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c366489-0d4c-48e1-99be-5b94881cd468-logs" (OuterVolumeSpecName: "logs") pod "8c366489-0d4c-48e1-99be-5b94881cd468" (UID: "8c366489-0d4c-48e1-99be-5b94881cd468"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.547625 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c366489-0d4c-48e1-99be-5b94881cd468-kube-api-access-54j9h" (OuterVolumeSpecName: "kube-api-access-54j9h") pod "8c366489-0d4c-48e1-99be-5b94881cd468" (UID: "8c366489-0d4c-48e1-99be-5b94881cd468"). InnerVolumeSpecName "kube-api-access-54j9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.566248 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c366489-0d4c-48e1-99be-5b94881cd468-config-data" (OuterVolumeSpecName: "config-data") pod "8c366489-0d4c-48e1-99be-5b94881cd468" (UID: "8c366489-0d4c-48e1-99be-5b94881cd468"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.605115 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c366489-0d4c-48e1-99be-5b94881cd468-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c366489-0d4c-48e1-99be-5b94881cd468" (UID: "8c366489-0d4c-48e1-99be-5b94881cd468"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.639607 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c366489-0d4c-48e1-99be-5b94881cd468-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.639708 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c366489-0d4c-48e1-99be-5b94881cd468-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.639725 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c366489-0d4c-48e1-99be-5b94881cd468-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.639737 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54j9h\" (UniqueName: \"kubernetes.io/projected/8c366489-0d4c-48e1-99be-5b94881cd468-kube-api-access-54j9h\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.678967 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.741460 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-logs\") pod \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\" (UID: \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\") " Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.741941 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc5pb\" (UniqueName: \"kubernetes.io/projected/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-kube-api-access-dc5pb\") pod \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\" (UID: \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\") " Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.742020 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-config-data\") pod \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\" (UID: \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\") " Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.742060 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-combined-ca-bundle\") pod \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\" (UID: \"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733\") " Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.742585 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-logs" (OuterVolumeSpecName: "logs") pod "5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" (UID: "5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.745829 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-kube-api-access-dc5pb" (OuterVolumeSpecName: "kube-api-access-dc5pb") pod "5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" (UID: "5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733"). InnerVolumeSpecName "kube-api-access-dc5pb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.778588 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" (UID: "5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.784521 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-config-data" (OuterVolumeSpecName: "config-data") pod "5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" (UID: "5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.787658 4997 generic.go:334] "Generic (PLEG): container finished" podID="8c366489-0d4c-48e1-99be-5b94881cd468" containerID="576ff2a343d43ad6e34c58d476cc43aab9729c86badcf63506fdae3033d099be" exitCode=0 Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.787779 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c366489-0d4c-48e1-99be-5b94881cd468","Type":"ContainerDied","Data":"576ff2a343d43ad6e34c58d476cc43aab9729c86badcf63506fdae3033d099be"} Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.787818 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c366489-0d4c-48e1-99be-5b94881cd468","Type":"ContainerDied","Data":"d148f4b9a2cd1ef771d7c390ca13219cdbaa19bfe9234154137b206d17af8d26"} Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.787839 4997 scope.go:117] "RemoveContainer" containerID="576ff2a343d43ad6e34c58d476cc43aab9729c86badcf63506fdae3033d099be" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.787963 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.793104 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"187e8113-1011-4622-bd2b-a582a9450a6f","Type":"ContainerDied","Data":"4f8a15ececd1a65272d44ea356426700d9be18b5b1fc57b73a61d3252cd00a69"} Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.793202 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.802492 4997 generic.go:334] "Generic (PLEG): container finished" podID="5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" containerID="ed33e13d3a1086153e0da6846482e061e4d7c00955733b7c2c0ac7b1a7662040" exitCode=0 Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.802753 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733","Type":"ContainerDied","Data":"ed33e13d3a1086153e0da6846482e061e4d7c00955733b7c2c0ac7b1a7662040"} Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.802922 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.803290 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733","Type":"ContainerDied","Data":"8bb7feaf6080b19cc8277f3e0d29967465aed7ba35e985dd9698d9247a255f39"} Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.805865 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4d30eaf1-eb47-46c2-b87a-18fe74b003e9","Type":"ContainerStarted","Data":"76a4cb2d326ab45ee9e0cbc72c2319cce50db19b89e1673bbf94b5c0eef87c41"} Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.812676 4997 scope.go:117] "RemoveContainer" containerID="a621dd45aff6ca71d2c373c426d8b890b899a53bd037c5203583e71b917ba244" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.832965 4997 scope.go:117] "RemoveContainer" containerID="576ff2a343d43ad6e34c58d476cc43aab9729c86badcf63506fdae3033d099be" Dec 05 08:59:32 crc kubenswrapper[4997]: E1205 08:59:32.833383 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"576ff2a343d43ad6e34c58d476cc43aab9729c86badcf63506fdae3033d099be\": container with ID starting with 576ff2a343d43ad6e34c58d476cc43aab9729c86badcf63506fdae3033d099be not found: ID does not exist" containerID="576ff2a343d43ad6e34c58d476cc43aab9729c86badcf63506fdae3033d099be" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.833417 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"576ff2a343d43ad6e34c58d476cc43aab9729c86badcf63506fdae3033d099be"} err="failed to get container status \"576ff2a343d43ad6e34c58d476cc43aab9729c86badcf63506fdae3033d099be\": rpc error: code = NotFound desc = could not find container \"576ff2a343d43ad6e34c58d476cc43aab9729c86badcf63506fdae3033d099be\": container with ID starting with 576ff2a343d43ad6e34c58d476cc43aab9729c86badcf63506fdae3033d099be not found: ID does not exist" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.833442 4997 scope.go:117] "RemoveContainer" containerID="a621dd45aff6ca71d2c373c426d8b890b899a53bd037c5203583e71b917ba244" Dec 05 08:59:32 crc kubenswrapper[4997]: E1205 08:59:32.833738 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a621dd45aff6ca71d2c373c426d8b890b899a53bd037c5203583e71b917ba244\": container with ID starting with a621dd45aff6ca71d2c373c426d8b890b899a53bd037c5203583e71b917ba244 not found: ID does not exist" containerID="a621dd45aff6ca71d2c373c426d8b890b899a53bd037c5203583e71b917ba244" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.833754 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a621dd45aff6ca71d2c373c426d8b890b899a53bd037c5203583e71b917ba244"} err="failed to get container status \"a621dd45aff6ca71d2c373c426d8b890b899a53bd037c5203583e71b917ba244\": rpc error: code = NotFound desc = could not find container \"a621dd45aff6ca71d2c373c426d8b890b899a53bd037c5203583e71b917ba244\": container with ID starting with a621dd45aff6ca71d2c373c426d8b890b899a53bd037c5203583e71b917ba244 not found: ID does not exist" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.833767 4997 scope.go:117] "RemoveContainer" containerID="0c4171ada7a2365ff05d143c62c6b5d2ff029f4808c2fc08d95ee1888d1b7557" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.842759 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.842733211 podStartE2EDuration="2.842733211s" podCreationTimestamp="2025-12-05 08:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:59:32.829683818 +0000 UTC m=+7473.358591079" watchObservedRunningTime="2025-12-05 08:59:32.842733211 +0000 UTC m=+7473.371640472" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.848915 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc5pb\" (UniqueName: \"kubernetes.io/projected/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-kube-api-access-dc5pb\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.848947 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.848958 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.848969 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.867310 4997 scope.go:117] "RemoveContainer" containerID="ed33e13d3a1086153e0da6846482e061e4d7c00955733b7c2c0ac7b1a7662040" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.880075 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.896833 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.919960 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 08:59:32 crc kubenswrapper[4997]: E1205 08:59:32.920413 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c366489-0d4c-48e1-99be-5b94881cd468" containerName="nova-api-api" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.920428 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c366489-0d4c-48e1-99be-5b94881cd468" containerName="nova-api-api" Dec 05 08:59:32 crc kubenswrapper[4997]: E1205 08:59:32.920453 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" containerName="nova-metadata-log" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.920460 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" containerName="nova-metadata-log" Dec 05 08:59:32 crc kubenswrapper[4997]: E1205 08:59:32.920478 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="187e8113-1011-4622-bd2b-a582a9450a6f" containerName="nova-cell0-conductor-conductor" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.920486 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="187e8113-1011-4622-bd2b-a582a9450a6f" containerName="nova-cell0-conductor-conductor" Dec 05 08:59:32 crc kubenswrapper[4997]: E1205 08:59:32.920503 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c366489-0d4c-48e1-99be-5b94881cd468" containerName="nova-api-log" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.920510 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c366489-0d4c-48e1-99be-5b94881cd468" containerName="nova-api-log" Dec 05 08:59:32 crc kubenswrapper[4997]: E1205 08:59:32.920525 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" containerName="nova-metadata-metadata" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.920533 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" containerName="nova-metadata-metadata" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.920798 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" containerName="nova-metadata-log" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.920821 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" containerName="nova-metadata-metadata" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.920832 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="187e8113-1011-4622-bd2b-a582a9450a6f" containerName="nova-cell0-conductor-conductor" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.920848 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c366489-0d4c-48e1-99be-5b94881cd468" containerName="nova-api-log" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.920860 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c366489-0d4c-48e1-99be-5b94881cd468" containerName="nova-api-api" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.922056 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.925643 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.939423 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.954250 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3cbe3c8-d087-461e-a663-c0e8229640b4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f3cbe3c8-d087-461e-a663-c0e8229640b4\") " pod="openstack/nova-api-0" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.954321 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpgsr\" (UniqueName: \"kubernetes.io/projected/f3cbe3c8-d087-461e-a663-c0e8229640b4-kube-api-access-fpgsr\") pod \"nova-api-0\" (UID: \"f3cbe3c8-d087-461e-a663-c0e8229640b4\") " pod="openstack/nova-api-0" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.954441 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3cbe3c8-d087-461e-a663-c0e8229640b4-config-data\") pod \"nova-api-0\" (UID: \"f3cbe3c8-d087-461e-a663-c0e8229640b4\") " pod="openstack/nova-api-0" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.954552 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3cbe3c8-d087-461e-a663-c0e8229640b4-logs\") pod \"nova-api-0\" (UID: \"f3cbe3c8-d087-461e-a663-c0e8229640b4\") " pod="openstack/nova-api-0" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.958525 4997 scope.go:117] "RemoveContainer" containerID="315590eb9eb2ef36418367ff8e38e1d476a35163132ddf8c6324fa1e31586887" Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.961536 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.973744 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.983934 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.991657 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:59:32 crc kubenswrapper[4997]: I1205 08:59:32.998861 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.000371 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.002836 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.007449 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.008959 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.011660 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.021308 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.026249 4997 scope.go:117] "RemoveContainer" containerID="ed33e13d3a1086153e0da6846482e061e4d7c00955733b7c2c0ac7b1a7662040" Dec 05 08:59:33 crc kubenswrapper[4997]: E1205 08:59:33.026740 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed33e13d3a1086153e0da6846482e061e4d7c00955733b7c2c0ac7b1a7662040\": container with ID starting with ed33e13d3a1086153e0da6846482e061e4d7c00955733b7c2c0ac7b1a7662040 not found: ID does not exist" containerID="ed33e13d3a1086153e0da6846482e061e4d7c00955733b7c2c0ac7b1a7662040" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.026799 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed33e13d3a1086153e0da6846482e061e4d7c00955733b7c2c0ac7b1a7662040"} err="failed to get container status \"ed33e13d3a1086153e0da6846482e061e4d7c00955733b7c2c0ac7b1a7662040\": rpc error: code = NotFound desc = could not find container \"ed33e13d3a1086153e0da6846482e061e4d7c00955733b7c2c0ac7b1a7662040\": container with ID starting with ed33e13d3a1086153e0da6846482e061e4d7c00955733b7c2c0ac7b1a7662040 not found: ID does not exist" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.026827 4997 scope.go:117] "RemoveContainer" containerID="315590eb9eb2ef36418367ff8e38e1d476a35163132ddf8c6324fa1e31586887" Dec 05 08:59:33 crc kubenswrapper[4997]: E1205 08:59:33.027210 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"315590eb9eb2ef36418367ff8e38e1d476a35163132ddf8c6324fa1e31586887\": container with ID starting with 315590eb9eb2ef36418367ff8e38e1d476a35163132ddf8c6324fa1e31586887 not found: ID does not exist" containerID="315590eb9eb2ef36418367ff8e38e1d476a35163132ddf8c6324fa1e31586887" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.027242 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"315590eb9eb2ef36418367ff8e38e1d476a35163132ddf8c6324fa1e31586887"} err="failed to get container status \"315590eb9eb2ef36418367ff8e38e1d476a35163132ddf8c6324fa1e31586887\": rpc error: code = NotFound desc = could not find container \"315590eb9eb2ef36418367ff8e38e1d476a35163132ddf8c6324fa1e31586887\": container with ID starting with 315590eb9eb2ef36418367ff8e38e1d476a35163132ddf8c6324fa1e31586887 not found: ID does not exist" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.029942 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.056141 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"04d5f9e7-d90b-4887-af25-cbf4e8a16db0\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.056204 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htfwz\" (UniqueName: \"kubernetes.io/projected/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-kube-api-access-htfwz\") pod \"nova-cell0-conductor-0\" (UID: \"04d5f9e7-d90b-4887-af25-cbf4e8a16db0\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.056246 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79d909f6-46e3-401d-aafa-3d649b632284-logs\") pod \"nova-metadata-0\" (UID: \"79d909f6-46e3-401d-aafa-3d649b632284\") " pod="openstack/nova-metadata-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.056280 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3cbe3c8-d087-461e-a663-c0e8229640b4-config-data\") pod \"nova-api-0\" (UID: \"f3cbe3c8-d087-461e-a663-c0e8229640b4\") " pod="openstack/nova-api-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.056399 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3cbe3c8-d087-461e-a663-c0e8229640b4-logs\") pod \"nova-api-0\" (UID: \"f3cbe3c8-d087-461e-a663-c0e8229640b4\") " pod="openstack/nova-api-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.056479 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d909f6-46e3-401d-aafa-3d649b632284-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"79d909f6-46e3-401d-aafa-3d649b632284\") " pod="openstack/nova-metadata-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.056521 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79d909f6-46e3-401d-aafa-3d649b632284-config-data\") pod \"nova-metadata-0\" (UID: \"79d909f6-46e3-401d-aafa-3d649b632284\") " pod="openstack/nova-metadata-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.056688 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3cbe3c8-d087-461e-a663-c0e8229640b4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f3cbe3c8-d087-461e-a663-c0e8229640b4\") " pod="openstack/nova-api-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.056774 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpgsr\" (UniqueName: \"kubernetes.io/projected/f3cbe3c8-d087-461e-a663-c0e8229640b4-kube-api-access-fpgsr\") pod \"nova-api-0\" (UID: \"f3cbe3c8-d087-461e-a663-c0e8229640b4\") " pod="openstack/nova-api-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.056820 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"04d5f9e7-d90b-4887-af25-cbf4e8a16db0\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.056894 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scq8m\" (UniqueName: \"kubernetes.io/projected/79d909f6-46e3-401d-aafa-3d649b632284-kube-api-access-scq8m\") pod \"nova-metadata-0\" (UID: \"79d909f6-46e3-401d-aafa-3d649b632284\") " pod="openstack/nova-metadata-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.057050 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3cbe3c8-d087-461e-a663-c0e8229640b4-logs\") pod \"nova-api-0\" (UID: \"f3cbe3c8-d087-461e-a663-c0e8229640b4\") " pod="openstack/nova-api-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.062431 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3cbe3c8-d087-461e-a663-c0e8229640b4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f3cbe3c8-d087-461e-a663-c0e8229640b4\") " pod="openstack/nova-api-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.082532 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3cbe3c8-d087-461e-a663-c0e8229640b4-config-data\") pod \"nova-api-0\" (UID: \"f3cbe3c8-d087-461e-a663-c0e8229640b4\") " pod="openstack/nova-api-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.103121 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpgsr\" (UniqueName: \"kubernetes.io/projected/f3cbe3c8-d087-461e-a663-c0e8229640b4-kube-api-access-fpgsr\") pod \"nova-api-0\" (UID: \"f3cbe3c8-d087-461e-a663-c0e8229640b4\") " pod="openstack/nova-api-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.159716 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scq8m\" (UniqueName: \"kubernetes.io/projected/79d909f6-46e3-401d-aafa-3d649b632284-kube-api-access-scq8m\") pod \"nova-metadata-0\" (UID: \"79d909f6-46e3-401d-aafa-3d649b632284\") " pod="openstack/nova-metadata-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.159774 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"04d5f9e7-d90b-4887-af25-cbf4e8a16db0\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.159794 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htfwz\" (UniqueName: \"kubernetes.io/projected/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-kube-api-access-htfwz\") pod \"nova-cell0-conductor-0\" (UID: \"04d5f9e7-d90b-4887-af25-cbf4e8a16db0\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.159818 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79d909f6-46e3-401d-aafa-3d649b632284-logs\") pod \"nova-metadata-0\" (UID: \"79d909f6-46e3-401d-aafa-3d649b632284\") " pod="openstack/nova-metadata-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.160108 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d909f6-46e3-401d-aafa-3d649b632284-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"79d909f6-46e3-401d-aafa-3d649b632284\") " pod="openstack/nova-metadata-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.160628 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79d909f6-46e3-401d-aafa-3d649b632284-logs\") pod \"nova-metadata-0\" (UID: \"79d909f6-46e3-401d-aafa-3d649b632284\") " pod="openstack/nova-metadata-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.160672 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79d909f6-46e3-401d-aafa-3d649b632284-config-data\") pod \"nova-metadata-0\" (UID: \"79d909f6-46e3-401d-aafa-3d649b632284\") " pod="openstack/nova-metadata-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.160758 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"04d5f9e7-d90b-4887-af25-cbf4e8a16db0\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.166125 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d909f6-46e3-401d-aafa-3d649b632284-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"79d909f6-46e3-401d-aafa-3d649b632284\") " pod="openstack/nova-metadata-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.166762 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"04d5f9e7-d90b-4887-af25-cbf4e8a16db0\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.173467 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"04d5f9e7-d90b-4887-af25-cbf4e8a16db0\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.180368 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79d909f6-46e3-401d-aafa-3d649b632284-config-data\") pod \"nova-metadata-0\" (UID: \"79d909f6-46e3-401d-aafa-3d649b632284\") " pod="openstack/nova-metadata-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.194088 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scq8m\" (UniqueName: \"kubernetes.io/projected/79d909f6-46e3-401d-aafa-3d649b632284-kube-api-access-scq8m\") pod \"nova-metadata-0\" (UID: \"79d909f6-46e3-401d-aafa-3d649b632284\") " pod="openstack/nova-metadata-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.194248 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htfwz\" (UniqueName: \"kubernetes.io/projected/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-kube-api-access-htfwz\") pod \"nova-cell0-conductor-0\" (UID: \"04d5f9e7-d90b-4887-af25-cbf4e8a16db0\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.288797 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.332865 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.365312 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.418345 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.591366 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfvwt\" (UniqueName: \"kubernetes.io/projected/700689bf-8f80-457b-8daf-3dfde8871759-kube-api-access-pfvwt\") pod \"700689bf-8f80-457b-8daf-3dfde8871759\" (UID: \"700689bf-8f80-457b-8daf-3dfde8871759\") " Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.591603 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700689bf-8f80-457b-8daf-3dfde8871759-config-data\") pod \"700689bf-8f80-457b-8daf-3dfde8871759\" (UID: \"700689bf-8f80-457b-8daf-3dfde8871759\") " Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.591696 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700689bf-8f80-457b-8daf-3dfde8871759-combined-ca-bundle\") pod \"700689bf-8f80-457b-8daf-3dfde8871759\" (UID: \"700689bf-8f80-457b-8daf-3dfde8871759\") " Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.599222 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/700689bf-8f80-457b-8daf-3dfde8871759-kube-api-access-pfvwt" (OuterVolumeSpecName: "kube-api-access-pfvwt") pod "700689bf-8f80-457b-8daf-3dfde8871759" (UID: "700689bf-8f80-457b-8daf-3dfde8871759"). InnerVolumeSpecName "kube-api-access-pfvwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.643181 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/700689bf-8f80-457b-8daf-3dfde8871759-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "700689bf-8f80-457b-8daf-3dfde8871759" (UID: "700689bf-8f80-457b-8daf-3dfde8871759"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.671436 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/700689bf-8f80-457b-8daf-3dfde8871759-config-data" (OuterVolumeSpecName: "config-data") pod "700689bf-8f80-457b-8daf-3dfde8871759" (UID: "700689bf-8f80-457b-8daf-3dfde8871759"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.694416 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700689bf-8f80-457b-8daf-3dfde8871759-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.694466 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700689bf-8f80-457b-8daf-3dfde8871759-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.694481 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfvwt\" (UniqueName: \"kubernetes.io/projected/700689bf-8f80-457b-8daf-3dfde8871759-kube-api-access-pfvwt\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.763076 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="187e8113-1011-4622-bd2b-a582a9450a6f" path="/var/lib/kubelet/pods/187e8113-1011-4622-bd2b-a582a9450a6f/volumes" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.763875 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733" path="/var/lib/kubelet/pods/5ad83fe5-dc32-4e76-a5ef-0bf1b68c1733/volumes" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.764496 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c366489-0d4c-48e1-99be-5b94881cd468" path="/var/lib/kubelet/pods/8c366489-0d4c-48e1-99be-5b94881cd468/volumes" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.815528 4997 generic.go:334] "Generic (PLEG): container finished" podID="700689bf-8f80-457b-8daf-3dfde8871759" containerID="5e1515944c4d5662ce252478df8e925f34222dd1e61fb08a7663defebb4255f3" exitCode=0 Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.815705 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.816429 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"700689bf-8f80-457b-8daf-3dfde8871759","Type":"ContainerDied","Data":"5e1515944c4d5662ce252478df8e925f34222dd1e61fb08a7663defebb4255f3"} Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.816554 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"700689bf-8f80-457b-8daf-3dfde8871759","Type":"ContainerDied","Data":"7763d3fb8bfb455d9ade86a711668b89034949666a376525c1f336fafdafbbb3"} Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.816577 4997 scope.go:117] "RemoveContainer" containerID="5e1515944c4d5662ce252478df8e925f34222dd1e61fb08a7663defebb4255f3" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.847908 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.858029 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.858388 4997 scope.go:117] "RemoveContainer" containerID="5e1515944c4d5662ce252478df8e925f34222dd1e61fb08a7663defebb4255f3" Dec 05 08:59:33 crc kubenswrapper[4997]: E1205 08:59:33.859001 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e1515944c4d5662ce252478df8e925f34222dd1e61fb08a7663defebb4255f3\": container with ID starting with 5e1515944c4d5662ce252478df8e925f34222dd1e61fb08a7663defebb4255f3 not found: ID does not exist" containerID="5e1515944c4d5662ce252478df8e925f34222dd1e61fb08a7663defebb4255f3" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.859032 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e1515944c4d5662ce252478df8e925f34222dd1e61fb08a7663defebb4255f3"} err="failed to get container status \"5e1515944c4d5662ce252478df8e925f34222dd1e61fb08a7663defebb4255f3\": rpc error: code = NotFound desc = could not find container \"5e1515944c4d5662ce252478df8e925f34222dd1e61fb08a7663defebb4255f3\": container with ID starting with 5e1515944c4d5662ce252478df8e925f34222dd1e61fb08a7663defebb4255f3 not found: ID does not exist" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.872286 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:59:33 crc kubenswrapper[4997]: E1205 08:59:33.872817 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="700689bf-8f80-457b-8daf-3dfde8871759" containerName="nova-scheduler-scheduler" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.872840 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="700689bf-8f80-457b-8daf-3dfde8871759" containerName="nova-scheduler-scheduler" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.873041 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="700689bf-8f80-457b-8daf-3dfde8871759" containerName="nova-scheduler-scheduler" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.873845 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.877204 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.889196 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.898787 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c444\" (UniqueName: \"kubernetes.io/projected/51c4332f-d8d3-479f-af91-3314d7111adc-kube-api-access-9c444\") pod \"nova-scheduler-0\" (UID: \"51c4332f-d8d3-479f-af91-3314d7111adc\") " pod="openstack/nova-scheduler-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.898830 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c4332f-d8d3-479f-af91-3314d7111adc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"51c4332f-d8d3-479f-af91-3314d7111adc\") " pod="openstack/nova-scheduler-0" Dec 05 08:59:33 crc kubenswrapper[4997]: I1205 08:59:33.898896 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51c4332f-d8d3-479f-af91-3314d7111adc-config-data\") pod \"nova-scheduler-0\" (UID: \"51c4332f-d8d3-479f-af91-3314d7111adc\") " pod="openstack/nova-scheduler-0" Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.001396 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c444\" (UniqueName: \"kubernetes.io/projected/51c4332f-d8d3-479f-af91-3314d7111adc-kube-api-access-9c444\") pod \"nova-scheduler-0\" (UID: \"51c4332f-d8d3-479f-af91-3314d7111adc\") " pod="openstack/nova-scheduler-0" Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.001477 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c4332f-d8d3-479f-af91-3314d7111adc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"51c4332f-d8d3-479f-af91-3314d7111adc\") " pod="openstack/nova-scheduler-0" Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.001570 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51c4332f-d8d3-479f-af91-3314d7111adc-config-data\") pod \"nova-scheduler-0\" (UID: \"51c4332f-d8d3-479f-af91-3314d7111adc\") " pod="openstack/nova-scheduler-0" Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.005941 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c4332f-d8d3-479f-af91-3314d7111adc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"51c4332f-d8d3-479f-af91-3314d7111adc\") " pod="openstack/nova-scheduler-0" Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.006280 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51c4332f-d8d3-479f-af91-3314d7111adc-config-data\") pod \"nova-scheduler-0\" (UID: \"51c4332f-d8d3-479f-af91-3314d7111adc\") " pod="openstack/nova-scheduler-0" Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.022572 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c444\" (UniqueName: \"kubernetes.io/projected/51c4332f-d8d3-479f-af91-3314d7111adc-kube-api-access-9c444\") pod \"nova-scheduler-0\" (UID: \"51c4332f-d8d3-479f-af91-3314d7111adc\") " pod="openstack/nova-scheduler-0" Dec 05 08:59:34 crc kubenswrapper[4997]: W1205 08:59:34.036071 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3cbe3c8_d087_461e_a663_c0e8229640b4.slice/crio-be32345d9465c8f0a5383ad4eda6f3ab78331e94ba462968c628d5004f5db8a7 WatchSource:0}: Error finding container be32345d9465c8f0a5383ad4eda6f3ab78331e94ba462968c628d5004f5db8a7: Status 404 returned error can't find the container with id be32345d9465c8f0a5383ad4eda6f3ab78331e94ba462968c628d5004f5db8a7 Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.039286 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:59:34 crc kubenswrapper[4997]: W1205 08:59:34.068850 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04d5f9e7_d90b_4887_af25_cbf4e8a16db0.slice/crio-8ce17ff382b14b6033ec249aa5b67540f64a3d66ea1c8a38bd6349d39baeb0e6 WatchSource:0}: Error finding container 8ce17ff382b14b6033ec249aa5b67540f64a3d66ea1c8a38bd6349d39baeb0e6: Status 404 returned error can't find the container with id 8ce17ff382b14b6033ec249aa5b67540f64a3d66ea1c8a38bd6349d39baeb0e6 Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.070722 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.081828 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.199826 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.672145 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:59:34 crc kubenswrapper[4997]: W1205 08:59:34.677780 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51c4332f_d8d3_479f_af91_3314d7111adc.slice/crio-2e3f7ef91a62475de6a1012179c9e72682747be2dfff6a7e66125482e438c969 WatchSource:0}: Error finding container 2e3f7ef91a62475de6a1012179c9e72682747be2dfff6a7e66125482e438c969: Status 404 returned error can't find the container with id 2e3f7ef91a62475de6a1012179c9e72682747be2dfff6a7e66125482e438c969 Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.842440 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"51c4332f-d8d3-479f-af91-3314d7111adc","Type":"ContainerStarted","Data":"2e3f7ef91a62475de6a1012179c9e72682747be2dfff6a7e66125482e438c969"} Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.849359 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79d909f6-46e3-401d-aafa-3d649b632284","Type":"ContainerStarted","Data":"747c13e8d23d170bcf696dca900cf69ae22620556ef8a97a6ad5111684d57f17"} Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.849407 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79d909f6-46e3-401d-aafa-3d649b632284","Type":"ContainerStarted","Data":"a78ece514e46e4b92e3bb36eb7f2ecd2ff6c5bd515c06adbe710b954d25f52f6"} Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.849417 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79d909f6-46e3-401d-aafa-3d649b632284","Type":"ContainerStarted","Data":"ac239a24fcdb01900d6f89c3f62788c62aa37b6dcf8c49e582522c41aa67db97"} Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.857876 4997 generic.go:334] "Generic (PLEG): container finished" podID="cadd4122-9bc4-4e70-97d4-7e062feed5de" containerID="45f674f05294bc21c57daee4316ef766b40f5822bb721afd5acc8925ff46098d" exitCode=0 Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.857956 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"cadd4122-9bc4-4e70-97d4-7e062feed5de","Type":"ContainerDied","Data":"45f674f05294bc21c57daee4316ef766b40f5822bb721afd5acc8925ff46098d"} Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.861802 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"04d5f9e7-d90b-4887-af25-cbf4e8a16db0","Type":"ContainerStarted","Data":"28b07c5c22ab9328db79dbaaf5998a6025761c2994ace11bb7e357dd95249f1a"} Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.861832 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"04d5f9e7-d90b-4887-af25-cbf4e8a16db0","Type":"ContainerStarted","Data":"8ce17ff382b14b6033ec249aa5b67540f64a3d66ea1c8a38bd6349d39baeb0e6"} Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.862600 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.866115 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f3cbe3c8-d087-461e-a663-c0e8229640b4","Type":"ContainerStarted","Data":"6deab0d56789b3905d671ebbc986574fd95edd3cc69a8e63b485e79625592e1d"} Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.866151 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f3cbe3c8-d087-461e-a663-c0e8229640b4","Type":"ContainerStarted","Data":"944cbd4353d191849652e26d5cd7806761de0e71e59a72861dcce5e9f3b7a7ff"} Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.866161 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f3cbe3c8-d087-461e-a663-c0e8229640b4","Type":"ContainerStarted","Data":"be32345d9465c8f0a5383ad4eda6f3ab78331e94ba462968c628d5004f5db8a7"} Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.894834 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.894813328 podStartE2EDuration="2.894813328s" podCreationTimestamp="2025-12-05 08:59:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:59:34.877925962 +0000 UTC m=+7475.406833223" watchObservedRunningTime="2025-12-05 08:59:34.894813328 +0000 UTC m=+7475.423720589" Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.898235 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.89822201 podStartE2EDuration="2.89822201s" podCreationTimestamp="2025-12-05 08:59:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:59:34.891501089 +0000 UTC m=+7475.420408360" watchObservedRunningTime="2025-12-05 08:59:34.89822201 +0000 UTC m=+7475.427129291" Dec 05 08:59:34 crc kubenswrapper[4997]: I1205 08:59:34.924754 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.924734055 podStartE2EDuration="2.924734055s" podCreationTimestamp="2025-12-05 08:59:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:59:34.91528619 +0000 UTC m=+7475.444193481" watchObservedRunningTime="2025-12-05 08:59:34.924734055 +0000 UTC m=+7475.453641316" Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.117202 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.123091 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadd4122-9bc4-4e70-97d4-7e062feed5de-config-data\") pod \"cadd4122-9bc4-4e70-97d4-7e062feed5de\" (UID: \"cadd4122-9bc4-4e70-97d4-7e062feed5de\") " Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.123147 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadd4122-9bc4-4e70-97d4-7e062feed5de-combined-ca-bundle\") pod \"cadd4122-9bc4-4e70-97d4-7e062feed5de\" (UID: \"cadd4122-9bc4-4e70-97d4-7e062feed5de\") " Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.123178 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zj6zn\" (UniqueName: \"kubernetes.io/projected/cadd4122-9bc4-4e70-97d4-7e062feed5de-kube-api-access-zj6zn\") pod \"cadd4122-9bc4-4e70-97d4-7e062feed5de\" (UID: \"cadd4122-9bc4-4e70-97d4-7e062feed5de\") " Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.132051 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cadd4122-9bc4-4e70-97d4-7e062feed5de-kube-api-access-zj6zn" (OuterVolumeSpecName: "kube-api-access-zj6zn") pod "cadd4122-9bc4-4e70-97d4-7e062feed5de" (UID: "cadd4122-9bc4-4e70-97d4-7e062feed5de"). InnerVolumeSpecName "kube-api-access-zj6zn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.153571 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cadd4122-9bc4-4e70-97d4-7e062feed5de-config-data" (OuterVolumeSpecName: "config-data") pod "cadd4122-9bc4-4e70-97d4-7e062feed5de" (UID: "cadd4122-9bc4-4e70-97d4-7e062feed5de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.159937 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cadd4122-9bc4-4e70-97d4-7e062feed5de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cadd4122-9bc4-4e70-97d4-7e062feed5de" (UID: "cadd4122-9bc4-4e70-97d4-7e062feed5de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.224376 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadd4122-9bc4-4e70-97d4-7e062feed5de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.224409 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zj6zn\" (UniqueName: \"kubernetes.io/projected/cadd4122-9bc4-4e70-97d4-7e062feed5de-kube-api-access-zj6zn\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.224421 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadd4122-9bc4-4e70-97d4-7e062feed5de-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.762400 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="700689bf-8f80-457b-8daf-3dfde8871759" path="/var/lib/kubelet/pods/700689bf-8f80-457b-8daf-3dfde8871759/volumes" Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.876592 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"51c4332f-d8d3-479f-af91-3314d7111adc","Type":"ContainerStarted","Data":"35db5a225a2271f9e2cace9eeead88d9507db1e4658527d66fe51579558496fb"} Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.878336 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"cadd4122-9bc4-4e70-97d4-7e062feed5de","Type":"ContainerDied","Data":"9af4c1582127e78dc66395947ff9cb62c444c1d10f40abb30a04767bef12f8d8"} Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.878395 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.878400 4997 scope.go:117] "RemoveContainer" containerID="45f674f05294bc21c57daee4316ef766b40f5822bb721afd5acc8925ff46098d" Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.896954 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.896934227 podStartE2EDuration="2.896934227s" podCreationTimestamp="2025-12-05 08:59:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:59:35.895625581 +0000 UTC m=+7476.424532852" watchObservedRunningTime="2025-12-05 08:59:35.896934227 +0000 UTC m=+7476.425841498" Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.924190 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.943703 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.957587 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 08:59:35 crc kubenswrapper[4997]: E1205 08:59:35.958846 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cadd4122-9bc4-4e70-97d4-7e062feed5de" containerName="nova-cell1-conductor-conductor" Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.958872 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="cadd4122-9bc4-4e70-97d4-7e062feed5de" containerName="nova-cell1-conductor-conductor" Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.959392 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="cadd4122-9bc4-4e70-97d4-7e062feed5de" containerName="nova-cell1-conductor-conductor" Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.960514 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.964681 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 08:59:35 crc kubenswrapper[4997]: I1205 08:59:35.996375 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 08:59:36 crc kubenswrapper[4997]: I1205 08:59:36.057187 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgzmq\" (UniqueName: \"kubernetes.io/projected/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-kube-api-access-lgzmq\") pod \"nova-cell1-conductor-0\" (UID: \"7a6e7c78-6f77-4d5e-96f8-95f67debcd97\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:59:36 crc kubenswrapper[4997]: I1205 08:59:36.057469 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7a6e7c78-6f77-4d5e-96f8-95f67debcd97\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:59:36 crc kubenswrapper[4997]: I1205 08:59:36.057553 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7a6e7c78-6f77-4d5e-96f8-95f67debcd97\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:59:36 crc kubenswrapper[4997]: I1205 08:59:36.154582 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:59:36 crc kubenswrapper[4997]: I1205 08:59:36.159323 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgzmq\" (UniqueName: \"kubernetes.io/projected/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-kube-api-access-lgzmq\") pod \"nova-cell1-conductor-0\" (UID: \"7a6e7c78-6f77-4d5e-96f8-95f67debcd97\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:59:36 crc kubenswrapper[4997]: I1205 08:59:36.159432 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7a6e7c78-6f77-4d5e-96f8-95f67debcd97\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:59:36 crc kubenswrapper[4997]: I1205 08:59:36.159466 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7a6e7c78-6f77-4d5e-96f8-95f67debcd97\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:59:36 crc kubenswrapper[4997]: I1205 08:59:36.173134 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7a6e7c78-6f77-4d5e-96f8-95f67debcd97\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:59:36 crc kubenswrapper[4997]: I1205 08:59:36.173610 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7a6e7c78-6f77-4d5e-96f8-95f67debcd97\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:59:36 crc kubenswrapper[4997]: I1205 08:59:36.174904 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgzmq\" (UniqueName: \"kubernetes.io/projected/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-kube-api-access-lgzmq\") pod \"nova-cell1-conductor-0\" (UID: \"7a6e7c78-6f77-4d5e-96f8-95f67debcd97\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:59:36 crc kubenswrapper[4997]: I1205 08:59:36.281102 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 08:59:36 crc kubenswrapper[4997]: W1205 08:59:36.796159 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a6e7c78_6f77_4d5e_96f8_95f67debcd97.slice/crio-aa55d63292154af2420a002e5b662be1664497a35a635f46e63107192049e0f8 WatchSource:0}: Error finding container aa55d63292154af2420a002e5b662be1664497a35a635f46e63107192049e0f8: Status 404 returned error can't find the container with id aa55d63292154af2420a002e5b662be1664497a35a635f46e63107192049e0f8 Dec 05 08:59:36 crc kubenswrapper[4997]: I1205 08:59:36.805750 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 08:59:36 crc kubenswrapper[4997]: I1205 08:59:36.888378 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7a6e7c78-6f77-4d5e-96f8-95f67debcd97","Type":"ContainerStarted","Data":"aa55d63292154af2420a002e5b662be1664497a35a635f46e63107192049e0f8"} Dec 05 08:59:37 crc kubenswrapper[4997]: I1205 08:59:37.749918 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 08:59:37 crc kubenswrapper[4997]: E1205 08:59:37.750447 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:59:37 crc kubenswrapper[4997]: I1205 08:59:37.761757 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cadd4122-9bc4-4e70-97d4-7e062feed5de" path="/var/lib/kubelet/pods/cadd4122-9bc4-4e70-97d4-7e062feed5de/volumes" Dec 05 08:59:37 crc kubenswrapper[4997]: I1205 08:59:37.906900 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7a6e7c78-6f77-4d5e-96f8-95f67debcd97","Type":"ContainerStarted","Data":"a765187e12ab594edbd3c92b89d19f7b86dabdc712477b007ef3882121061020"} Dec 05 08:59:37 crc kubenswrapper[4997]: I1205 08:59:37.907069 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 05 08:59:37 crc kubenswrapper[4997]: I1205 08:59:37.930602 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.930578817 podStartE2EDuration="2.930578817s" podCreationTimestamp="2025-12-05 08:59:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:59:37.92737489 +0000 UTC m=+7478.456282161" watchObservedRunningTime="2025-12-05 08:59:37.930578817 +0000 UTC m=+7478.459486078" Dec 05 08:59:38 crc kubenswrapper[4997]: I1205 08:59:38.333702 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 08:59:38 crc kubenswrapper[4997]: I1205 08:59:38.333776 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 08:59:39 crc kubenswrapper[4997]: I1205 08:59:39.200833 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 08:59:41 crc kubenswrapper[4997]: I1205 08:59:41.154365 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:59:41 crc kubenswrapper[4997]: I1205 08:59:41.166975 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:59:41 crc kubenswrapper[4997]: I1205 08:59:41.329400 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 05 08:59:41 crc kubenswrapper[4997]: I1205 08:59:41.946921 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:59:43 crc kubenswrapper[4997]: I1205 08:59:43.289518 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 08:59:43 crc kubenswrapper[4997]: I1205 08:59:43.289955 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 08:59:43 crc kubenswrapper[4997]: I1205 08:59:43.334223 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 08:59:43 crc kubenswrapper[4997]: I1205 08:59:43.334296 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 08:59:43 crc kubenswrapper[4997]: I1205 08:59:43.399776 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 05 08:59:44 crc kubenswrapper[4997]: I1205 08:59:44.200288 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 08:59:44 crc kubenswrapper[4997]: I1205 08:59:44.236494 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 08:59:44 crc kubenswrapper[4997]: I1205 08:59:44.372789 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f3cbe3c8-d087-461e-a663-c0e8229640b4" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.97:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:59:44 crc kubenswrapper[4997]: I1205 08:59:44.372789 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f3cbe3c8-d087-461e-a663-c0e8229640b4" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.97:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:59:44 crc kubenswrapper[4997]: I1205 08:59:44.454776 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="79d909f6-46e3-401d-aafa-3d649b632284" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.98:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:59:44 crc kubenswrapper[4997]: I1205 08:59:44.454776 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="79d909f6-46e3-401d-aafa-3d649b632284" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.98:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:59:45 crc kubenswrapper[4997]: I1205 08:59:45.012479 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 08:59:51 crc kubenswrapper[4997]: I1205 08:59:51.749571 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 08:59:51 crc kubenswrapper[4997]: E1205 08:59:51.750377 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.043423 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.045798 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.048688 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.061515 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.158217 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-scripts\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.158535 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.158720 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.158891 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.158945 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tnk5\" (UniqueName: \"kubernetes.io/projected/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-kube-api-access-7tnk5\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.159007 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-config-data\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.261108 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-scripts\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.261417 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.261634 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.262051 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.262124 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.262142 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tnk5\" (UniqueName: \"kubernetes.io/projected/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-kube-api-access-7tnk5\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.262247 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-config-data\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.267395 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-scripts\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.267528 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.270786 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-config-data\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.271400 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.282839 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tnk5\" (UniqueName: \"kubernetes.io/projected/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-kube-api-access-7tnk5\") pod \"cinder-scheduler-0\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.374673 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 08:59:52 crc kubenswrapper[4997]: I1205 08:59:52.717806 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 08:59:52 crc kubenswrapper[4997]: W1205 08:59:52.718784 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ed72646_a4c9_41c2_9b1e_d23f802e44a8.slice/crio-feca09416d9dbffdb2687336ff3449dbcdda41a95a6c550631da96c512b50bc1 WatchSource:0}: Error finding container feca09416d9dbffdb2687336ff3449dbcdda41a95a6c550631da96c512b50bc1: Status 404 returned error can't find the container with id feca09416d9dbffdb2687336ff3449dbcdda41a95a6c550631da96c512b50bc1 Dec 05 08:59:53 crc kubenswrapper[4997]: I1205 08:59:53.038313 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0ed72646-a4c9-41c2-9b1e-d23f802e44a8","Type":"ContainerStarted","Data":"feca09416d9dbffdb2687336ff3449dbcdda41a95a6c550631da96c512b50bc1"} Dec 05 08:59:53 crc kubenswrapper[4997]: I1205 08:59:53.303658 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 08:59:53 crc kubenswrapper[4997]: I1205 08:59:53.304030 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 08:59:53 crc kubenswrapper[4997]: I1205 08:59:53.306132 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 08:59:53 crc kubenswrapper[4997]: I1205 08:59:53.309858 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 08:59:53 crc kubenswrapper[4997]: I1205 08:59:53.341480 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 08:59:53 crc kubenswrapper[4997]: I1205 08:59:53.341818 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 08:59:53 crc kubenswrapper[4997]: I1205 08:59:53.348062 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.068159 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.069091 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="af352d23-3b54-41ce-b382-73611152bf30" containerName="cinder-api-log" containerID="cri-o://cbe02905034056ee64697ed4e7969882b6b240d233968b88185095818be54cf0" gracePeriod=30 Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.069503 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="af352d23-3b54-41ce-b382-73611152bf30" containerName="cinder-api" containerID="cri-o://7e61220eb02cd00838d606d86785fd88e5260eb8bd3d4f005cc85dfe57840c64" gracePeriod=30 Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.075046 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0ed72646-a4c9-41c2-9b1e-d23f802e44a8","Type":"ContainerStarted","Data":"b66d30eb08873b7e36e93f855133156974c6584679281841aa07162eba405459"} Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.075705 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.084554 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.086247 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.672673 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.675027 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.677566 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.700567 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.707741 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.711346 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.717233 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.733686 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.810875 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eba14735-b01b-4358-918f-f8f2735f2925-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.810926 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eba14735-b01b-4358-918f-f8f2735f2925-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.810953 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.810972 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-dev\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.810991 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811006 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-dev\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811021 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-run\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811048 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eba14735-b01b-4358-918f-f8f2735f2925-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811065 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eba14735-b01b-4358-918f-f8f2735f2925-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811085 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811098 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-lib-modules\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811118 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811132 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7990da72-c93f-40dd-ae1c-2e5af98502ad-config-data-custom\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811148 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811170 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811184 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-sys\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811207 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7990da72-c93f-40dd-ae1c-2e5af98502ad-ceph\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811222 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811239 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67b76\" (UniqueName: \"kubernetes.io/projected/eba14735-b01b-4358-918f-f8f2735f2925-kube-api-access-67b76\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811263 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-etc-nvme\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811316 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-sys\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811333 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-run\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811357 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811374 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811387 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811409 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7990da72-c93f-40dd-ae1c-2e5af98502ad-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811434 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811449 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/eba14735-b01b-4358-918f-f8f2735f2925-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811469 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w545m\" (UniqueName: \"kubernetes.io/projected/7990da72-c93f-40dd-ae1c-2e5af98502ad-kube-api-access-w545m\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811492 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7990da72-c93f-40dd-ae1c-2e5af98502ad-config-data\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811517 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7990da72-c93f-40dd-ae1c-2e5af98502ad-scripts\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.811537 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.912731 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.912789 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.912826 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7990da72-c93f-40dd-ae1c-2e5af98502ad-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.912858 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.912894 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/eba14735-b01b-4358-918f-f8f2735f2925-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.912929 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w545m\" (UniqueName: \"kubernetes.io/projected/7990da72-c93f-40dd-ae1c-2e5af98502ad-kube-api-access-w545m\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.912963 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7990da72-c93f-40dd-ae1c-2e5af98502ad-config-data\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913002 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7990da72-c93f-40dd-ae1c-2e5af98502ad-scripts\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913031 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913060 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eba14735-b01b-4358-918f-f8f2735f2925-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913092 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eba14735-b01b-4358-918f-f8f2735f2925-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913123 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913147 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-dev\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913174 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913189 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-dev\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913205 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-run\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913246 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eba14735-b01b-4358-918f-f8f2735f2925-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913271 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eba14735-b01b-4358-918f-f8f2735f2925-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913291 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913311 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-lib-modules\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913336 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913358 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7990da72-c93f-40dd-ae1c-2e5af98502ad-config-data-custom\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913383 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913416 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913438 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-sys\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913460 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7990da72-c93f-40dd-ae1c-2e5af98502ad-ceph\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913479 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913504 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67b76\" (UniqueName: \"kubernetes.io/projected/eba14735-b01b-4358-918f-f8f2735f2925-kube-api-access-67b76\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913537 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-etc-nvme\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913586 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-sys\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913609 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-run\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913655 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913757 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.913802 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.914061 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.914100 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-dev\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.914248 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-dev\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.914248 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.914273 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.914293 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-run\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.914374 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.914481 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-sys\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.914954 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.915132 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.915183 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.915231 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-lib-modules\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.915257 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.915277 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.915309 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-sys\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.915324 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.915342 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7990da72-c93f-40dd-ae1c-2e5af98502ad-etc-nvme\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.915367 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/eba14735-b01b-4358-918f-f8f2735f2925-run\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.922343 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eba14735-b01b-4358-918f-f8f2735f2925-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.933902 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7990da72-c93f-40dd-ae1c-2e5af98502ad-config-data\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.934453 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7990da72-c93f-40dd-ae1c-2e5af98502ad-scripts\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.934796 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/eba14735-b01b-4358-918f-f8f2735f2925-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.934948 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7990da72-c93f-40dd-ae1c-2e5af98502ad-ceph\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.935211 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eba14735-b01b-4358-918f-f8f2735f2925-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.935220 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eba14735-b01b-4358-918f-f8f2735f2925-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.935294 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eba14735-b01b-4358-918f-f8f2735f2925-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.935407 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7990da72-c93f-40dd-ae1c-2e5af98502ad-config-data-custom\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.935531 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7990da72-c93f-40dd-ae1c-2e5af98502ad-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.948024 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67b76\" (UniqueName: \"kubernetes.io/projected/eba14735-b01b-4358-918f-f8f2735f2925-kube-api-access-67b76\") pod \"cinder-volume-volume1-0\" (UID: \"eba14735-b01b-4358-918f-f8f2735f2925\") " pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.945800 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w545m\" (UniqueName: \"kubernetes.io/projected/7990da72-c93f-40dd-ae1c-2e5af98502ad-kube-api-access-w545m\") pod \"cinder-backup-0\" (UID: \"7990da72-c93f-40dd-ae1c-2e5af98502ad\") " pod="openstack/cinder-backup-0" Dec 05 08:59:54 crc kubenswrapper[4997]: I1205 08:59:54.997286 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 05 08:59:55 crc kubenswrapper[4997]: I1205 08:59:55.054265 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 05 08:59:55 crc kubenswrapper[4997]: I1205 08:59:55.086857 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0ed72646-a4c9-41c2-9b1e-d23f802e44a8","Type":"ContainerStarted","Data":"3d6a35a124be621eba7dfd3b9fbce2881263b1be63bc2f3e11226c88d275fedf"} Dec 05 08:59:55 crc kubenswrapper[4997]: I1205 08:59:55.093590 4997 generic.go:334] "Generic (PLEG): container finished" podID="af352d23-3b54-41ce-b382-73611152bf30" containerID="cbe02905034056ee64697ed4e7969882b6b240d233968b88185095818be54cf0" exitCode=143 Dec 05 08:59:55 crc kubenswrapper[4997]: I1205 08:59:55.093651 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"af352d23-3b54-41ce-b382-73611152bf30","Type":"ContainerDied","Data":"cbe02905034056ee64697ed4e7969882b6b240d233968b88185095818be54cf0"} Dec 05 08:59:55 crc kubenswrapper[4997]: I1205 08:59:55.647586 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.36779036 podStartE2EDuration="3.647564585s" podCreationTimestamp="2025-12-05 08:59:52 +0000 UTC" firstStartedPulling="2025-12-05 08:59:52.721794954 +0000 UTC m=+7493.250702215" lastFinishedPulling="2025-12-05 08:59:53.001569179 +0000 UTC m=+7493.530476440" observedRunningTime="2025-12-05 08:59:55.115937257 +0000 UTC m=+7495.644844528" watchObservedRunningTime="2025-12-05 08:59:55.647564585 +0000 UTC m=+7496.176471846" Dec 05 08:59:55 crc kubenswrapper[4997]: I1205 08:59:55.655637 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 05 08:59:55 crc kubenswrapper[4997]: W1205 08:59:55.656377 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeba14735_b01b_4358_918f_f8f2735f2925.slice/crio-92b8ed0fa9e3522ec5b644c819e02bcca97856ad21c6a8b37803f0586c6f494c WatchSource:0}: Error finding container 92b8ed0fa9e3522ec5b644c819e02bcca97856ad21c6a8b37803f0586c6f494c: Status 404 returned error can't find the container with id 92b8ed0fa9e3522ec5b644c819e02bcca97856ad21c6a8b37803f0586c6f494c Dec 05 08:59:55 crc kubenswrapper[4997]: W1205 08:59:55.742710 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7990da72_c93f_40dd_ae1c_2e5af98502ad.slice/crio-60cc549709e9fa268c5d78eb8d11ce157095a6f695cd98631a9cbe4fafd6410c WatchSource:0}: Error finding container 60cc549709e9fa268c5d78eb8d11ce157095a6f695cd98631a9cbe4fafd6410c: Status 404 returned error can't find the container with id 60cc549709e9fa268c5d78eb8d11ce157095a6f695cd98631a9cbe4fafd6410c Dec 05 08:59:55 crc kubenswrapper[4997]: I1205 08:59:55.744700 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 05 08:59:56 crc kubenswrapper[4997]: I1205 08:59:56.104093 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"7990da72-c93f-40dd-ae1c-2e5af98502ad","Type":"ContainerStarted","Data":"60cc549709e9fa268c5d78eb8d11ce157095a6f695cd98631a9cbe4fafd6410c"} Dec 05 08:59:56 crc kubenswrapper[4997]: I1205 08:59:56.105696 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"eba14735-b01b-4358-918f-f8f2735f2925","Type":"ContainerStarted","Data":"92b8ed0fa9e3522ec5b644c819e02bcca97856ad21c6a8b37803f0586c6f494c"} Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.125217 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"7990da72-c93f-40dd-ae1c-2e5af98502ad","Type":"ContainerStarted","Data":"91e003460cdb22d80792e33915cce45eb15183108dca89aaa1ce8255a0357102"} Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.125890 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"7990da72-c93f-40dd-ae1c-2e5af98502ad","Type":"ContainerStarted","Data":"f2d1b7dc9b46f4552999cd5e3ebb6936d753dc65f544ee921fda2a3ca552a340"} Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.130071 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"eba14735-b01b-4358-918f-f8f2735f2925","Type":"ContainerStarted","Data":"0ae8c749cd0a013599d5b2dec77fc769136adbb9a6ea0d2535f2e61f3040b4ba"} Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.130114 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"eba14735-b01b-4358-918f-f8f2735f2925","Type":"ContainerStarted","Data":"09e5ece23471842f666ad1a5420d8a90f0d175ca354d58db86ab4f5acec01f58"} Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.161816 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.712954712 podStartE2EDuration="3.161796107s" podCreationTimestamp="2025-12-05 08:59:54 +0000 UTC" firstStartedPulling="2025-12-05 08:59:55.744968663 +0000 UTC m=+7496.273875934" lastFinishedPulling="2025-12-05 08:59:56.193810068 +0000 UTC m=+7496.722717329" observedRunningTime="2025-12-05 08:59:57.152821884 +0000 UTC m=+7497.681729155" watchObservedRunningTime="2025-12-05 08:59:57.161796107 +0000 UTC m=+7497.690703368" Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.182059 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.649659542 podStartE2EDuration="3.182042562s" podCreationTimestamp="2025-12-05 08:59:54 +0000 UTC" firstStartedPulling="2025-12-05 08:59:55.659041715 +0000 UTC m=+7496.187948976" lastFinishedPulling="2025-12-05 08:59:56.191424735 +0000 UTC m=+7496.720331996" observedRunningTime="2025-12-05 08:59:57.179139814 +0000 UTC m=+7497.708047105" watchObservedRunningTime="2025-12-05 08:59:57.182042562 +0000 UTC m=+7497.710949823" Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.236694 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="af352d23-3b54-41ce-b382-73611152bf30" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.1.95:8776/healthcheck\": read tcp 10.217.0.2:49748->10.217.1.95:8776: read: connection reset by peer" Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.376214 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.765429 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.950950 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-scripts\") pod \"af352d23-3b54-41ce-b382-73611152bf30\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.951006 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af352d23-3b54-41ce-b382-73611152bf30-logs\") pod \"af352d23-3b54-41ce-b382-73611152bf30\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.951055 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af352d23-3b54-41ce-b382-73611152bf30-etc-machine-id\") pod \"af352d23-3b54-41ce-b382-73611152bf30\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.951088 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-config-data-custom\") pod \"af352d23-3b54-41ce-b382-73611152bf30\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.951127 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-combined-ca-bundle\") pod \"af352d23-3b54-41ce-b382-73611152bf30\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.951170 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rhbw\" (UniqueName: \"kubernetes.io/projected/af352d23-3b54-41ce-b382-73611152bf30-kube-api-access-6rhbw\") pod \"af352d23-3b54-41ce-b382-73611152bf30\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.951225 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-config-data\") pod \"af352d23-3b54-41ce-b382-73611152bf30\" (UID: \"af352d23-3b54-41ce-b382-73611152bf30\") " Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.954175 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af352d23-3b54-41ce-b382-73611152bf30-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "af352d23-3b54-41ce-b382-73611152bf30" (UID: "af352d23-3b54-41ce-b382-73611152bf30"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.954323 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af352d23-3b54-41ce-b382-73611152bf30-logs" (OuterVolumeSpecName: "logs") pod "af352d23-3b54-41ce-b382-73611152bf30" (UID: "af352d23-3b54-41ce-b382-73611152bf30"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.959306 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af352d23-3b54-41ce-b382-73611152bf30-kube-api-access-6rhbw" (OuterVolumeSpecName: "kube-api-access-6rhbw") pod "af352d23-3b54-41ce-b382-73611152bf30" (UID: "af352d23-3b54-41ce-b382-73611152bf30"). InnerVolumeSpecName "kube-api-access-6rhbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.959360 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-scripts" (OuterVolumeSpecName: "scripts") pod "af352d23-3b54-41ce-b382-73611152bf30" (UID: "af352d23-3b54-41ce-b382-73611152bf30"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.970813 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "af352d23-3b54-41ce-b382-73611152bf30" (UID: "af352d23-3b54-41ce-b382-73611152bf30"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:57 crc kubenswrapper[4997]: I1205 08:59:57.978970 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af352d23-3b54-41ce-b382-73611152bf30" (UID: "af352d23-3b54-41ce-b382-73611152bf30"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.003827 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-config-data" (OuterVolumeSpecName: "config-data") pod "af352d23-3b54-41ce-b382-73611152bf30" (UID: "af352d23-3b54-41ce-b382-73611152bf30"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.052950 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.052985 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.052995 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af352d23-3b54-41ce-b382-73611152bf30-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.053003 4997 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af352d23-3b54-41ce-b382-73611152bf30-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.053017 4997 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.053026 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af352d23-3b54-41ce-b382-73611152bf30-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.053035 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rhbw\" (UniqueName: \"kubernetes.io/projected/af352d23-3b54-41ce-b382-73611152bf30-kube-api-access-6rhbw\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.140935 4997 generic.go:334] "Generic (PLEG): container finished" podID="af352d23-3b54-41ce-b382-73611152bf30" containerID="7e61220eb02cd00838d606d86785fd88e5260eb8bd3d4f005cc85dfe57840c64" exitCode=0 Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.141826 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.142745 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"af352d23-3b54-41ce-b382-73611152bf30","Type":"ContainerDied","Data":"7e61220eb02cd00838d606d86785fd88e5260eb8bd3d4f005cc85dfe57840c64"} Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.142799 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"af352d23-3b54-41ce-b382-73611152bf30","Type":"ContainerDied","Data":"da6506909722af71f54899090e6cba69a4ff8a1c87d8f1c1c616bf51ee990af5"} Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.142817 4997 scope.go:117] "RemoveContainer" containerID="7e61220eb02cd00838d606d86785fd88e5260eb8bd3d4f005cc85dfe57840c64" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.174932 4997 scope.go:117] "RemoveContainer" containerID="cbe02905034056ee64697ed4e7969882b6b240d233968b88185095818be54cf0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.214882 4997 scope.go:117] "RemoveContainer" containerID="7e61220eb02cd00838d606d86785fd88e5260eb8bd3d4f005cc85dfe57840c64" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.215006 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:59:58 crc kubenswrapper[4997]: E1205 08:59:58.215288 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e61220eb02cd00838d606d86785fd88e5260eb8bd3d4f005cc85dfe57840c64\": container with ID starting with 7e61220eb02cd00838d606d86785fd88e5260eb8bd3d4f005cc85dfe57840c64 not found: ID does not exist" containerID="7e61220eb02cd00838d606d86785fd88e5260eb8bd3d4f005cc85dfe57840c64" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.215347 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e61220eb02cd00838d606d86785fd88e5260eb8bd3d4f005cc85dfe57840c64"} err="failed to get container status \"7e61220eb02cd00838d606d86785fd88e5260eb8bd3d4f005cc85dfe57840c64\": rpc error: code = NotFound desc = could not find container \"7e61220eb02cd00838d606d86785fd88e5260eb8bd3d4f005cc85dfe57840c64\": container with ID starting with 7e61220eb02cd00838d606d86785fd88e5260eb8bd3d4f005cc85dfe57840c64 not found: ID does not exist" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.215372 4997 scope.go:117] "RemoveContainer" containerID="cbe02905034056ee64697ed4e7969882b6b240d233968b88185095818be54cf0" Dec 05 08:59:58 crc kubenswrapper[4997]: E1205 08:59:58.215641 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbe02905034056ee64697ed4e7969882b6b240d233968b88185095818be54cf0\": container with ID starting with cbe02905034056ee64697ed4e7969882b6b240d233968b88185095818be54cf0 not found: ID does not exist" containerID="cbe02905034056ee64697ed4e7969882b6b240d233968b88185095818be54cf0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.215669 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbe02905034056ee64697ed4e7969882b6b240d233968b88185095818be54cf0"} err="failed to get container status \"cbe02905034056ee64697ed4e7969882b6b240d233968b88185095818be54cf0\": rpc error: code = NotFound desc = could not find container \"cbe02905034056ee64697ed4e7969882b6b240d233968b88185095818be54cf0\": container with ID starting with cbe02905034056ee64697ed4e7969882b6b240d233968b88185095818be54cf0 not found: ID does not exist" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.228852 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.239944 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:59:58 crc kubenswrapper[4997]: E1205 08:59:58.240428 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af352d23-3b54-41ce-b382-73611152bf30" containerName="cinder-api-log" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.240448 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="af352d23-3b54-41ce-b382-73611152bf30" containerName="cinder-api-log" Dec 05 08:59:58 crc kubenswrapper[4997]: E1205 08:59:58.240471 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af352d23-3b54-41ce-b382-73611152bf30" containerName="cinder-api" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.240480 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="af352d23-3b54-41ce-b382-73611152bf30" containerName="cinder-api" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.240758 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="af352d23-3b54-41ce-b382-73611152bf30" containerName="cinder-api-log" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.240773 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="af352d23-3b54-41ce-b382-73611152bf30" containerName="cinder-api" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.242016 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.243943 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.250857 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.358552 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66529\" (UniqueName: \"kubernetes.io/projected/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-kube-api-access-66529\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.359002 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-scripts\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.359043 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-logs\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.359065 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-config-data-custom\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.359082 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-config-data\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.359111 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.359137 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.461215 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-scripts\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.461308 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-logs\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.461343 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-config-data-custom\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.461362 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-config-data\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.461396 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.461422 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.461483 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66529\" (UniqueName: \"kubernetes.io/projected/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-kube-api-access-66529\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.462849 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.463344 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-logs\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.471975 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.472302 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-scripts\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.472597 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-config-data-custom\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.473679 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-config-data\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.488346 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66529\" (UniqueName: \"kubernetes.io/projected/9de7a6bf-87ed-44b2-9a59-e91ef6f53740-kube-api-access-66529\") pod \"cinder-api-0\" (UID: \"9de7a6bf-87ed-44b2-9a59-e91ef6f53740\") " pod="openstack/cinder-api-0" Dec 05 08:59:58 crc kubenswrapper[4997]: I1205 08:59:58.567201 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 08:59:59 crc kubenswrapper[4997]: I1205 08:59:59.064228 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:59:59 crc kubenswrapper[4997]: I1205 08:59:59.154367 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9de7a6bf-87ed-44b2-9a59-e91ef6f53740","Type":"ContainerStarted","Data":"81714091ef5683eebb01c40ba0e23529139f83906ba4fdb9683a3834e7e0c1db"} Dec 05 08:59:59 crc kubenswrapper[4997]: I1205 08:59:59.790425 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af352d23-3b54-41ce-b382-73611152bf30" path="/var/lib/kubelet/pods/af352d23-3b54-41ce-b382-73611152bf30/volumes" Dec 05 08:59:59 crc kubenswrapper[4997]: I1205 08:59:59.998340 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.054523 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.167461 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t"] Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.169103 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t" Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.173275 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.173454 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.196703 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t"] Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.210993 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9de7a6bf-87ed-44b2-9a59-e91ef6f53740","Type":"ContainerStarted","Data":"3fea805ad4df9ba949c82b7e29cd5fef32f57e7cd87542f2abba41aeb739e6f4"} Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.240220 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc60482-e3d1-455f-9d81-58ef71b08668-secret-volume\") pod \"collect-profiles-29415420-cqs2t\" (UID: \"4dc60482-e3d1-455f-9d81-58ef71b08668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t" Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.240376 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc60482-e3d1-455f-9d81-58ef71b08668-config-volume\") pod \"collect-profiles-29415420-cqs2t\" (UID: \"4dc60482-e3d1-455f-9d81-58ef71b08668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t" Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.240401 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmmfr\" (UniqueName: \"kubernetes.io/projected/4dc60482-e3d1-455f-9d81-58ef71b08668-kube-api-access-gmmfr\") pod \"collect-profiles-29415420-cqs2t\" (UID: \"4dc60482-e3d1-455f-9d81-58ef71b08668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t" Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.342471 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc60482-e3d1-455f-9d81-58ef71b08668-secret-volume\") pod \"collect-profiles-29415420-cqs2t\" (UID: \"4dc60482-e3d1-455f-9d81-58ef71b08668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t" Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.342692 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc60482-e3d1-455f-9d81-58ef71b08668-config-volume\") pod \"collect-profiles-29415420-cqs2t\" (UID: \"4dc60482-e3d1-455f-9d81-58ef71b08668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t" Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.342720 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmmfr\" (UniqueName: \"kubernetes.io/projected/4dc60482-e3d1-455f-9d81-58ef71b08668-kube-api-access-gmmfr\") pod \"collect-profiles-29415420-cqs2t\" (UID: \"4dc60482-e3d1-455f-9d81-58ef71b08668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t" Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.343673 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc60482-e3d1-455f-9d81-58ef71b08668-config-volume\") pod \"collect-profiles-29415420-cqs2t\" (UID: \"4dc60482-e3d1-455f-9d81-58ef71b08668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t" Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.361180 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc60482-e3d1-455f-9d81-58ef71b08668-secret-volume\") pod \"collect-profiles-29415420-cqs2t\" (UID: \"4dc60482-e3d1-455f-9d81-58ef71b08668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t" Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.368521 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmmfr\" (UniqueName: \"kubernetes.io/projected/4dc60482-e3d1-455f-9d81-58ef71b08668-kube-api-access-gmmfr\") pod \"collect-profiles-29415420-cqs2t\" (UID: \"4dc60482-e3d1-455f-9d81-58ef71b08668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t" Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.504881 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t" Dec 05 09:00:00 crc kubenswrapper[4997]: I1205 09:00:00.826004 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t"] Dec 05 09:00:01 crc kubenswrapper[4997]: I1205 09:00:01.225951 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t" event={"ID":"4dc60482-e3d1-455f-9d81-58ef71b08668","Type":"ContainerStarted","Data":"e92ba1096f1801b80d0ee209fb07cd20aaed5df90ee61f5e60c6db5269749d9e"} Dec 05 09:00:01 crc kubenswrapper[4997]: I1205 09:00:01.226344 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t" event={"ID":"4dc60482-e3d1-455f-9d81-58ef71b08668","Type":"ContainerStarted","Data":"ede88a95d4026c756e3fede1115ffdd228190a35804194fda008ab9fe9e289bc"} Dec 05 09:00:01 crc kubenswrapper[4997]: I1205 09:00:01.227870 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9de7a6bf-87ed-44b2-9a59-e91ef6f53740","Type":"ContainerStarted","Data":"19cdea38df20377d6ea228dab44af829700fcaf906f92b300959010a18a93b67"} Dec 05 09:00:01 crc kubenswrapper[4997]: I1205 09:00:01.228114 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 05 09:00:01 crc kubenswrapper[4997]: I1205 09:00:01.262436 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.262418315 podStartE2EDuration="3.262418315s" podCreationTimestamp="2025-12-05 08:59:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:00:01.257856852 +0000 UTC m=+7501.786764133" watchObservedRunningTime="2025-12-05 09:00:01.262418315 +0000 UTC m=+7501.791325576" Dec 05 09:00:02 crc kubenswrapper[4997]: I1205 09:00:02.237234 4997 generic.go:334] "Generic (PLEG): container finished" podID="4dc60482-e3d1-455f-9d81-58ef71b08668" containerID="e92ba1096f1801b80d0ee209fb07cd20aaed5df90ee61f5e60c6db5269749d9e" exitCode=0 Dec 05 09:00:02 crc kubenswrapper[4997]: I1205 09:00:02.237363 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t" event={"ID":"4dc60482-e3d1-455f-9d81-58ef71b08668","Type":"ContainerDied","Data":"e92ba1096f1801b80d0ee209fb07cd20aaed5df90ee61f5e60c6db5269749d9e"} Dec 05 09:00:02 crc kubenswrapper[4997]: I1205 09:00:02.593677 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t" Dec 05 09:00:02 crc kubenswrapper[4997]: I1205 09:00:02.605601 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 05 09:00:02 crc kubenswrapper[4997]: I1205 09:00:02.658584 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 09:00:02 crc kubenswrapper[4997]: I1205 09:00:02.705026 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc60482-e3d1-455f-9d81-58ef71b08668-secret-volume\") pod \"4dc60482-e3d1-455f-9d81-58ef71b08668\" (UID: \"4dc60482-e3d1-455f-9d81-58ef71b08668\") " Dec 05 09:00:02 crc kubenswrapper[4997]: I1205 09:00:02.705329 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmmfr\" (UniqueName: \"kubernetes.io/projected/4dc60482-e3d1-455f-9d81-58ef71b08668-kube-api-access-gmmfr\") pod \"4dc60482-e3d1-455f-9d81-58ef71b08668\" (UID: \"4dc60482-e3d1-455f-9d81-58ef71b08668\") " Dec 05 09:00:02 crc kubenswrapper[4997]: I1205 09:00:02.705369 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc60482-e3d1-455f-9d81-58ef71b08668-config-volume\") pod \"4dc60482-e3d1-455f-9d81-58ef71b08668\" (UID: \"4dc60482-e3d1-455f-9d81-58ef71b08668\") " Dec 05 09:00:02 crc kubenswrapper[4997]: I1205 09:00:02.707624 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dc60482-e3d1-455f-9d81-58ef71b08668-config-volume" (OuterVolumeSpecName: "config-volume") pod "4dc60482-e3d1-455f-9d81-58ef71b08668" (UID: "4dc60482-e3d1-455f-9d81-58ef71b08668"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:00:02 crc kubenswrapper[4997]: I1205 09:00:02.712966 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dc60482-e3d1-455f-9d81-58ef71b08668-kube-api-access-gmmfr" (OuterVolumeSpecName: "kube-api-access-gmmfr") pod "4dc60482-e3d1-455f-9d81-58ef71b08668" (UID: "4dc60482-e3d1-455f-9d81-58ef71b08668"). InnerVolumeSpecName "kube-api-access-gmmfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:00:02 crc kubenswrapper[4997]: I1205 09:00:02.712977 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dc60482-e3d1-455f-9d81-58ef71b08668-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4dc60482-e3d1-455f-9d81-58ef71b08668" (UID: "4dc60482-e3d1-455f-9d81-58ef71b08668"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:00:02 crc kubenswrapper[4997]: I1205 09:00:02.749149 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:00:02 crc kubenswrapper[4997]: E1205 09:00:02.749509 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:00:02 crc kubenswrapper[4997]: I1205 09:00:02.809489 4997 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc60482-e3d1-455f-9d81-58ef71b08668-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:02 crc kubenswrapper[4997]: I1205 09:00:02.809527 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmmfr\" (UniqueName: \"kubernetes.io/projected/4dc60482-e3d1-455f-9d81-58ef71b08668-kube-api-access-gmmfr\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:02 crc kubenswrapper[4997]: I1205 09:00:02.809539 4997 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc60482-e3d1-455f-9d81-58ef71b08668-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:03 crc kubenswrapper[4997]: I1205 09:00:03.247544 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t" event={"ID":"4dc60482-e3d1-455f-9d81-58ef71b08668","Type":"ContainerDied","Data":"ede88a95d4026c756e3fede1115ffdd228190a35804194fda008ab9fe9e289bc"} Dec 05 09:00:03 crc kubenswrapper[4997]: I1205 09:00:03.249093 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ede88a95d4026c756e3fede1115ffdd228190a35804194fda008ab9fe9e289bc" Dec 05 09:00:03 crc kubenswrapper[4997]: I1205 09:00:03.247680 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="0ed72646-a4c9-41c2-9b1e-d23f802e44a8" containerName="cinder-scheduler" containerID="cri-o://b66d30eb08873b7e36e93f855133156974c6584679281841aa07162eba405459" gracePeriod=30 Dec 05 09:00:03 crc kubenswrapper[4997]: I1205 09:00:03.247964 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="0ed72646-a4c9-41c2-9b1e-d23f802e44a8" containerName="probe" containerID="cri-o://3d6a35a124be621eba7dfd3b9fbce2881263b1be63bc2f3e11226c88d275fedf" gracePeriod=30 Dec 05 09:00:03 crc kubenswrapper[4997]: I1205 09:00:03.247563 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t" Dec 05 09:00:03 crc kubenswrapper[4997]: I1205 09:00:03.673731 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864"] Dec 05 09:00:03 crc kubenswrapper[4997]: I1205 09:00:03.683548 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415375-9k864"] Dec 05 09:00:03 crc kubenswrapper[4997]: I1205 09:00:03.759692 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0228be20-bfff-4d5b-97d9-8b5f8ace9d5f" path="/var/lib/kubelet/pods/0228be20-bfff-4d5b-97d9-8b5f8ace9d5f/volumes" Dec 05 09:00:04 crc kubenswrapper[4997]: I1205 09:00:04.258319 4997 generic.go:334] "Generic (PLEG): container finished" podID="0ed72646-a4c9-41c2-9b1e-d23f802e44a8" containerID="3d6a35a124be621eba7dfd3b9fbce2881263b1be63bc2f3e11226c88d275fedf" exitCode=0 Dec 05 09:00:04 crc kubenswrapper[4997]: I1205 09:00:04.258369 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0ed72646-a4c9-41c2-9b1e-d23f802e44a8","Type":"ContainerDied","Data":"3d6a35a124be621eba7dfd3b9fbce2881263b1be63bc2f3e11226c88d275fedf"} Dec 05 09:00:05 crc kubenswrapper[4997]: I1205 09:00:05.224921 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Dec 05 09:00:05 crc kubenswrapper[4997]: I1205 09:00:05.298057 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.140942 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.273754 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tnk5\" (UniqueName: \"kubernetes.io/projected/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-kube-api-access-7tnk5\") pod \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.273849 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-combined-ca-bundle\") pod \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.273896 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-config-data\") pod \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.273938 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-config-data-custom\") pod \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.273957 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-scripts\") pod \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.273981 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-etc-machine-id\") pod \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\" (UID: \"0ed72646-a4c9-41c2-9b1e-d23f802e44a8\") " Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.274847 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0ed72646-a4c9-41c2-9b1e-d23f802e44a8" (UID: "0ed72646-a4c9-41c2-9b1e-d23f802e44a8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.281835 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0ed72646-a4c9-41c2-9b1e-d23f802e44a8" (UID: "0ed72646-a4c9-41c2-9b1e-d23f802e44a8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.281869 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-scripts" (OuterVolumeSpecName: "scripts") pod "0ed72646-a4c9-41c2-9b1e-d23f802e44a8" (UID: "0ed72646-a4c9-41c2-9b1e-d23f802e44a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.285525 4997 generic.go:334] "Generic (PLEG): container finished" podID="0ed72646-a4c9-41c2-9b1e-d23f802e44a8" containerID="b66d30eb08873b7e36e93f855133156974c6584679281841aa07162eba405459" exitCode=0 Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.285570 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0ed72646-a4c9-41c2-9b1e-d23f802e44a8","Type":"ContainerDied","Data":"b66d30eb08873b7e36e93f855133156974c6584679281841aa07162eba405459"} Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.285602 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0ed72646-a4c9-41c2-9b1e-d23f802e44a8","Type":"ContainerDied","Data":"feca09416d9dbffdb2687336ff3449dbcdda41a95a6c550631da96c512b50bc1"} Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.285641 4997 scope.go:117] "RemoveContainer" containerID="3d6a35a124be621eba7dfd3b9fbce2881263b1be63bc2f3e11226c88d275fedf" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.285799 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.307057 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-kube-api-access-7tnk5" (OuterVolumeSpecName: "kube-api-access-7tnk5") pod "0ed72646-a4c9-41c2-9b1e-d23f802e44a8" (UID: "0ed72646-a4c9-41c2-9b1e-d23f802e44a8"). InnerVolumeSpecName "kube-api-access-7tnk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.364149 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ed72646-a4c9-41c2-9b1e-d23f802e44a8" (UID: "0ed72646-a4c9-41c2-9b1e-d23f802e44a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.376168 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tnk5\" (UniqueName: \"kubernetes.io/projected/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-kube-api-access-7tnk5\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.376202 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.376215 4997 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.376226 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.376238 4997 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.397932 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-config-data" (OuterVolumeSpecName: "config-data") pod "0ed72646-a4c9-41c2-9b1e-d23f802e44a8" (UID: "0ed72646-a4c9-41c2-9b1e-d23f802e44a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.409788 4997 scope.go:117] "RemoveContainer" containerID="b66d30eb08873b7e36e93f855133156974c6584679281841aa07162eba405459" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.430895 4997 scope.go:117] "RemoveContainer" containerID="3d6a35a124be621eba7dfd3b9fbce2881263b1be63bc2f3e11226c88d275fedf" Dec 05 09:00:06 crc kubenswrapper[4997]: E1205 09:00:06.431342 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d6a35a124be621eba7dfd3b9fbce2881263b1be63bc2f3e11226c88d275fedf\": container with ID starting with 3d6a35a124be621eba7dfd3b9fbce2881263b1be63bc2f3e11226c88d275fedf not found: ID does not exist" containerID="3d6a35a124be621eba7dfd3b9fbce2881263b1be63bc2f3e11226c88d275fedf" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.431373 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d6a35a124be621eba7dfd3b9fbce2881263b1be63bc2f3e11226c88d275fedf"} err="failed to get container status \"3d6a35a124be621eba7dfd3b9fbce2881263b1be63bc2f3e11226c88d275fedf\": rpc error: code = NotFound desc = could not find container \"3d6a35a124be621eba7dfd3b9fbce2881263b1be63bc2f3e11226c88d275fedf\": container with ID starting with 3d6a35a124be621eba7dfd3b9fbce2881263b1be63bc2f3e11226c88d275fedf not found: ID does not exist" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.431396 4997 scope.go:117] "RemoveContainer" containerID="b66d30eb08873b7e36e93f855133156974c6584679281841aa07162eba405459" Dec 05 09:00:06 crc kubenswrapper[4997]: E1205 09:00:06.431572 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b66d30eb08873b7e36e93f855133156974c6584679281841aa07162eba405459\": container with ID starting with b66d30eb08873b7e36e93f855133156974c6584679281841aa07162eba405459 not found: ID does not exist" containerID="b66d30eb08873b7e36e93f855133156974c6584679281841aa07162eba405459" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.431592 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b66d30eb08873b7e36e93f855133156974c6584679281841aa07162eba405459"} err="failed to get container status \"b66d30eb08873b7e36e93f855133156974c6584679281841aa07162eba405459\": rpc error: code = NotFound desc = could not find container \"b66d30eb08873b7e36e93f855133156974c6584679281841aa07162eba405459\": container with ID starting with b66d30eb08873b7e36e93f855133156974c6584679281841aa07162eba405459 not found: ID does not exist" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.478629 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed72646-a4c9-41c2-9b1e-d23f802e44a8-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.624390 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.633053 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.655837 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 09:00:06 crc kubenswrapper[4997]: E1205 09:00:06.656243 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ed72646-a4c9-41c2-9b1e-d23f802e44a8" containerName="probe" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.656264 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ed72646-a4c9-41c2-9b1e-d23f802e44a8" containerName="probe" Dec 05 09:00:06 crc kubenswrapper[4997]: E1205 09:00:06.656294 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dc60482-e3d1-455f-9d81-58ef71b08668" containerName="collect-profiles" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.656301 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dc60482-e3d1-455f-9d81-58ef71b08668" containerName="collect-profiles" Dec 05 09:00:06 crc kubenswrapper[4997]: E1205 09:00:06.656314 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ed72646-a4c9-41c2-9b1e-d23f802e44a8" containerName="cinder-scheduler" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.656321 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ed72646-a4c9-41c2-9b1e-d23f802e44a8" containerName="cinder-scheduler" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.656474 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ed72646-a4c9-41c2-9b1e-d23f802e44a8" containerName="probe" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.656490 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ed72646-a4c9-41c2-9b1e-d23f802e44a8" containerName="cinder-scheduler" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.656505 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dc60482-e3d1-455f-9d81-58ef71b08668" containerName="collect-profiles" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.659506 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.661959 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.665784 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.783540 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/958ea1bf-c540-4450-a408-c225c3ba1dd3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.783701 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/958ea1bf-c540-4450-a408-c225c3ba1dd3-config-data\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.783737 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpgns\" (UniqueName: \"kubernetes.io/projected/958ea1bf-c540-4450-a408-c225c3ba1dd3-kube-api-access-vpgns\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.783763 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958ea1bf-c540-4450-a408-c225c3ba1dd3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.783781 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/958ea1bf-c540-4450-a408-c225c3ba1dd3-scripts\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.783816 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/958ea1bf-c540-4450-a408-c225c3ba1dd3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.885840 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958ea1bf-c540-4450-a408-c225c3ba1dd3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.885939 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/958ea1bf-c540-4450-a408-c225c3ba1dd3-scripts\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.886067 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/958ea1bf-c540-4450-a408-c225c3ba1dd3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.886248 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/958ea1bf-c540-4450-a408-c225c3ba1dd3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.886396 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/958ea1bf-c540-4450-a408-c225c3ba1dd3-config-data\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.886479 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpgns\" (UniqueName: \"kubernetes.io/projected/958ea1bf-c540-4450-a408-c225c3ba1dd3-kube-api-access-vpgns\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.887072 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/958ea1bf-c540-4450-a408-c225c3ba1dd3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.891556 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/958ea1bf-c540-4450-a408-c225c3ba1dd3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.892075 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/958ea1bf-c540-4450-a408-c225c3ba1dd3-config-data\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.892826 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958ea1bf-c540-4450-a408-c225c3ba1dd3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.893330 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/958ea1bf-c540-4450-a408-c225c3ba1dd3-scripts\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.905442 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpgns\" (UniqueName: \"kubernetes.io/projected/958ea1bf-c540-4450-a408-c225c3ba1dd3-kube-api-access-vpgns\") pod \"cinder-scheduler-0\" (UID: \"958ea1bf-c540-4450-a408-c225c3ba1dd3\") " pod="openstack/cinder-scheduler-0" Dec 05 09:00:06 crc kubenswrapper[4997]: I1205 09:00:06.996138 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 09:00:07 crc kubenswrapper[4997]: I1205 09:00:07.220997 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 09:00:07 crc kubenswrapper[4997]: I1205 09:00:07.301602 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"958ea1bf-c540-4450-a408-c225c3ba1dd3","Type":"ContainerStarted","Data":"2cedde0205d44701f7a9b80cc5f91c10e9e34f7ce4f0973c2c888f94746ce055"} Dec 05 09:00:07 crc kubenswrapper[4997]: I1205 09:00:07.765089 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ed72646-a4c9-41c2-9b1e-d23f802e44a8" path="/var/lib/kubelet/pods/0ed72646-a4c9-41c2-9b1e-d23f802e44a8/volumes" Dec 05 09:00:08 crc kubenswrapper[4997]: I1205 09:00:08.330823 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"958ea1bf-c540-4450-a408-c225c3ba1dd3","Type":"ContainerStarted","Data":"68ae7b18b479da7c3b473b8b35fed8172ecaf75bee8539f6fbd77e0344451b34"} Dec 05 09:00:09 crc kubenswrapper[4997]: I1205 09:00:09.350457 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"958ea1bf-c540-4450-a408-c225c3ba1dd3","Type":"ContainerStarted","Data":"43e22c8ec9d0c8dbd725079f93b69c9272bda7f54ea7e3317088cc9fa001e21b"} Dec 05 09:00:09 crc kubenswrapper[4997]: I1205 09:00:09.380683 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.380662625 podStartE2EDuration="3.380662625s" podCreationTimestamp="2025-12-05 09:00:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:00:09.374820966 +0000 UTC m=+7509.903728247" watchObservedRunningTime="2025-12-05 09:00:09.380662625 +0000 UTC m=+7509.909569886" Dec 05 09:00:10 crc kubenswrapper[4997]: I1205 09:00:10.493276 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 05 09:00:11 crc kubenswrapper[4997]: I1205 09:00:11.997199 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 05 09:00:13 crc kubenswrapper[4997]: I1205 09:00:13.750017 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:00:13 crc kubenswrapper[4997]: E1205 09:00:13.750849 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:00:15 crc kubenswrapper[4997]: I1205 09:00:15.781856 4997 scope.go:117] "RemoveContainer" containerID="0ccda522da9ed3273e605f81df796f6d14bb94a65119065bdd8959c257ab6867" Dec 05 09:00:17 crc kubenswrapper[4997]: I1205 09:00:17.202886 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 05 09:00:24 crc kubenswrapper[4997]: I1205 09:00:24.748442 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:00:24 crc kubenswrapper[4997]: E1205 09:00:24.749365 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:00:39 crc kubenswrapper[4997]: I1205 09:00:39.756215 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:00:39 crc kubenswrapper[4997]: E1205 09:00:39.757048 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:00:49 crc kubenswrapper[4997]: I1205 09:00:49.077093 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-e2cd-account-create-update-c54l4"] Dec 05 09:00:49 crc kubenswrapper[4997]: I1205 09:00:49.087902 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-pmf4c"] Dec 05 09:00:49 crc kubenswrapper[4997]: I1205 09:00:49.096682 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-e2cd-account-create-update-c54l4"] Dec 05 09:00:49 crc kubenswrapper[4997]: I1205 09:00:49.105348 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-pmf4c"] Dec 05 09:00:49 crc kubenswrapper[4997]: I1205 09:00:49.761097 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54fc3d55-a800-4f20-ad4b-eb48952d3660" path="/var/lib/kubelet/pods/54fc3d55-a800-4f20-ad4b-eb48952d3660/volumes" Dec 05 09:00:49 crc kubenswrapper[4997]: I1205 09:00:49.763820 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f94c087-4666-45bd-92b5-8d47209a0bc6" path="/var/lib/kubelet/pods/8f94c087-4666-45bd-92b5-8d47209a0bc6/volumes" Dec 05 09:00:52 crc kubenswrapper[4997]: I1205 09:00:52.749270 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:00:52 crc kubenswrapper[4997]: E1205 09:00:52.749912 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.040341 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-6vpxj"] Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.052206 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-6vpxj"] Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.161267 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29415421-f7jhc"] Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.163376 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415421-f7jhc" Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.175837 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29415421-f7jhc"] Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.219737 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-fernet-keys\") pod \"keystone-cron-29415421-f7jhc\" (UID: \"6787bbc7-67b4-4f03-9302-470b4f539214\") " pod="openstack/keystone-cron-29415421-f7jhc" Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.219796 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-combined-ca-bundle\") pod \"keystone-cron-29415421-f7jhc\" (UID: \"6787bbc7-67b4-4f03-9302-470b4f539214\") " pod="openstack/keystone-cron-29415421-f7jhc" Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.219893 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djq7r\" (UniqueName: \"kubernetes.io/projected/6787bbc7-67b4-4f03-9302-470b4f539214-kube-api-access-djq7r\") pod \"keystone-cron-29415421-f7jhc\" (UID: \"6787bbc7-67b4-4f03-9302-470b4f539214\") " pod="openstack/keystone-cron-29415421-f7jhc" Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.219952 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-config-data\") pod \"keystone-cron-29415421-f7jhc\" (UID: \"6787bbc7-67b4-4f03-9302-470b4f539214\") " pod="openstack/keystone-cron-29415421-f7jhc" Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.321306 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-fernet-keys\") pod \"keystone-cron-29415421-f7jhc\" (UID: \"6787bbc7-67b4-4f03-9302-470b4f539214\") " pod="openstack/keystone-cron-29415421-f7jhc" Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.321372 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-combined-ca-bundle\") pod \"keystone-cron-29415421-f7jhc\" (UID: \"6787bbc7-67b4-4f03-9302-470b4f539214\") " pod="openstack/keystone-cron-29415421-f7jhc" Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.321429 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djq7r\" (UniqueName: \"kubernetes.io/projected/6787bbc7-67b4-4f03-9302-470b4f539214-kube-api-access-djq7r\") pod \"keystone-cron-29415421-f7jhc\" (UID: \"6787bbc7-67b4-4f03-9302-470b4f539214\") " pod="openstack/keystone-cron-29415421-f7jhc" Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.321472 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-config-data\") pod \"keystone-cron-29415421-f7jhc\" (UID: \"6787bbc7-67b4-4f03-9302-470b4f539214\") " pod="openstack/keystone-cron-29415421-f7jhc" Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.326742 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-fernet-keys\") pod \"keystone-cron-29415421-f7jhc\" (UID: \"6787bbc7-67b4-4f03-9302-470b4f539214\") " pod="openstack/keystone-cron-29415421-f7jhc" Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.326918 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-config-data\") pod \"keystone-cron-29415421-f7jhc\" (UID: \"6787bbc7-67b4-4f03-9302-470b4f539214\") " pod="openstack/keystone-cron-29415421-f7jhc" Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.327280 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-combined-ca-bundle\") pod \"keystone-cron-29415421-f7jhc\" (UID: \"6787bbc7-67b4-4f03-9302-470b4f539214\") " pod="openstack/keystone-cron-29415421-f7jhc" Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.340499 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djq7r\" (UniqueName: \"kubernetes.io/projected/6787bbc7-67b4-4f03-9302-470b4f539214-kube-api-access-djq7r\") pod \"keystone-cron-29415421-f7jhc\" (UID: \"6787bbc7-67b4-4f03-9302-470b4f539214\") " pod="openstack/keystone-cron-29415421-f7jhc" Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.488772 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415421-f7jhc" Dec 05 09:01:00 crc kubenswrapper[4997]: I1205 09:01:00.994540 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29415421-f7jhc"] Dec 05 09:01:01 crc kubenswrapper[4997]: I1205 09:01:01.759994 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b68a777-809c-4794-bb7f-6b336a78130c" path="/var/lib/kubelet/pods/0b68a777-809c-4794-bb7f-6b336a78130c/volumes" Dec 05 09:01:01 crc kubenswrapper[4997]: I1205 09:01:01.860509 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415421-f7jhc" event={"ID":"6787bbc7-67b4-4f03-9302-470b4f539214","Type":"ContainerStarted","Data":"c4bb99009681da4db101a28cd25df91ecee892bbee035f4cab40048c26ddc771"} Dec 05 09:01:01 crc kubenswrapper[4997]: I1205 09:01:01.860554 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415421-f7jhc" event={"ID":"6787bbc7-67b4-4f03-9302-470b4f539214","Type":"ContainerStarted","Data":"04450452c0b2b4dc74752f72cf332d33d98bdca30bd7ec8349913ca60608b963"} Dec 05 09:01:01 crc kubenswrapper[4997]: I1205 09:01:01.878675 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29415421-f7jhc" podStartSLOduration=1.878655584 podStartE2EDuration="1.878655584s" podCreationTimestamp="2025-12-05 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:01:01.877162404 +0000 UTC m=+7562.406069685" watchObservedRunningTime="2025-12-05 09:01:01.878655584 +0000 UTC m=+7562.407562845" Dec 05 09:01:03 crc kubenswrapper[4997]: I1205 09:01:03.879441 4997 generic.go:334] "Generic (PLEG): container finished" podID="6787bbc7-67b4-4f03-9302-470b4f539214" containerID="c4bb99009681da4db101a28cd25df91ecee892bbee035f4cab40048c26ddc771" exitCode=0 Dec 05 09:01:03 crc kubenswrapper[4997]: I1205 09:01:03.879489 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415421-f7jhc" event={"ID":"6787bbc7-67b4-4f03-9302-470b4f539214","Type":"ContainerDied","Data":"c4bb99009681da4db101a28cd25df91ecee892bbee035f4cab40048c26ddc771"} Dec 05 09:01:05 crc kubenswrapper[4997]: I1205 09:01:05.284724 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415421-f7jhc" Dec 05 09:01:05 crc kubenswrapper[4997]: I1205 09:01:05.418200 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-combined-ca-bundle\") pod \"6787bbc7-67b4-4f03-9302-470b4f539214\" (UID: \"6787bbc7-67b4-4f03-9302-470b4f539214\") " Dec 05 09:01:05 crc kubenswrapper[4997]: I1205 09:01:05.418243 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-fernet-keys\") pod \"6787bbc7-67b4-4f03-9302-470b4f539214\" (UID: \"6787bbc7-67b4-4f03-9302-470b4f539214\") " Dec 05 09:01:05 crc kubenswrapper[4997]: I1205 09:01:05.418305 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djq7r\" (UniqueName: \"kubernetes.io/projected/6787bbc7-67b4-4f03-9302-470b4f539214-kube-api-access-djq7r\") pod \"6787bbc7-67b4-4f03-9302-470b4f539214\" (UID: \"6787bbc7-67b4-4f03-9302-470b4f539214\") " Dec 05 09:01:05 crc kubenswrapper[4997]: I1205 09:01:05.418387 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-config-data\") pod \"6787bbc7-67b4-4f03-9302-470b4f539214\" (UID: \"6787bbc7-67b4-4f03-9302-470b4f539214\") " Dec 05 09:01:05 crc kubenswrapper[4997]: I1205 09:01:05.425460 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6787bbc7-67b4-4f03-9302-470b4f539214-kube-api-access-djq7r" (OuterVolumeSpecName: "kube-api-access-djq7r") pod "6787bbc7-67b4-4f03-9302-470b4f539214" (UID: "6787bbc7-67b4-4f03-9302-470b4f539214"). InnerVolumeSpecName "kube-api-access-djq7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:01:05 crc kubenswrapper[4997]: I1205 09:01:05.435976 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6787bbc7-67b4-4f03-9302-470b4f539214" (UID: "6787bbc7-67b4-4f03-9302-470b4f539214"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:01:05 crc kubenswrapper[4997]: I1205 09:01:05.451228 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6787bbc7-67b4-4f03-9302-470b4f539214" (UID: "6787bbc7-67b4-4f03-9302-470b4f539214"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:01:05 crc kubenswrapper[4997]: I1205 09:01:05.482948 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-config-data" (OuterVolumeSpecName: "config-data") pod "6787bbc7-67b4-4f03-9302-470b4f539214" (UID: "6787bbc7-67b4-4f03-9302-470b4f539214"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:01:05 crc kubenswrapper[4997]: I1205 09:01:05.520566 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:05 crc kubenswrapper[4997]: I1205 09:01:05.521422 4997 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:05 crc kubenswrapper[4997]: I1205 09:01:05.521454 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djq7r\" (UniqueName: \"kubernetes.io/projected/6787bbc7-67b4-4f03-9302-470b4f539214-kube-api-access-djq7r\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:05 crc kubenswrapper[4997]: I1205 09:01:05.521465 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6787bbc7-67b4-4f03-9302-470b4f539214-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:05 crc kubenswrapper[4997]: I1205 09:01:05.901066 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415421-f7jhc" event={"ID":"6787bbc7-67b4-4f03-9302-470b4f539214","Type":"ContainerDied","Data":"04450452c0b2b4dc74752f72cf332d33d98bdca30bd7ec8349913ca60608b963"} Dec 05 09:01:05 crc kubenswrapper[4997]: I1205 09:01:05.901117 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415421-f7jhc" Dec 05 09:01:05 crc kubenswrapper[4997]: I1205 09:01:05.901126 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04450452c0b2b4dc74752f72cf332d33d98bdca30bd7ec8349913ca60608b963" Dec 05 09:01:07 crc kubenswrapper[4997]: I1205 09:01:07.750785 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:01:07 crc kubenswrapper[4997]: E1205 09:01:07.751635 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:01:15 crc kubenswrapper[4997]: I1205 09:01:15.971207 4997 scope.go:117] "RemoveContainer" containerID="3ed6e708fc661912301a30eb43674bccd45365ffa4096fdfa3fb1cf18c6ee779" Dec 05 09:01:16 crc kubenswrapper[4997]: I1205 09:01:16.030825 4997 scope.go:117] "RemoveContainer" containerID="afe5d97d369a654bc9dbb3bbb1778a60d7e857b4f56127530b272e3a87343b85" Dec 05 09:01:16 crc kubenswrapper[4997]: I1205 09:01:16.070122 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-sxn7d"] Dec 05 09:01:16 crc kubenswrapper[4997]: I1205 09:01:16.072776 4997 scope.go:117] "RemoveContainer" containerID="ccccf89fa650efef7e7ae39fcfbbf18cb328eab8e34c99bb541980db90e9a2e8" Dec 05 09:01:16 crc kubenswrapper[4997]: I1205 09:01:16.081517 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-sxn7d"] Dec 05 09:01:17 crc kubenswrapper[4997]: I1205 09:01:17.760097 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d1240db-51c8-4e89-9165-d6a27b973908" path="/var/lib/kubelet/pods/8d1240db-51c8-4e89-9165-d6a27b973908/volumes" Dec 05 09:01:21 crc kubenswrapper[4997]: I1205 09:01:21.749884 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:01:21 crc kubenswrapper[4997]: E1205 09:01:21.750838 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:01:32 crc kubenswrapper[4997]: I1205 09:01:32.749812 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:01:32 crc kubenswrapper[4997]: E1205 09:01:32.750575 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:01:43 crc kubenswrapper[4997]: I1205 09:01:43.749193 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:01:43 crc kubenswrapper[4997]: E1205 09:01:43.751348 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.666756 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-77c64b89f5-b5lmn"] Dec 05 09:01:48 crc kubenswrapper[4997]: E1205 09:01:48.669104 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6787bbc7-67b4-4f03-9302-470b4f539214" containerName="keystone-cron" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.669200 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6787bbc7-67b4-4f03-9302-470b4f539214" containerName="keystone-cron" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.669483 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6787bbc7-67b4-4f03-9302-470b4f539214" containerName="keystone-cron" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.682597 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-77c64b89f5-b5lmn"] Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.682966 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.692259 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-84b87" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.692674 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.692961 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.693282 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.743481 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.744123 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="733c0bd0-5736-4e03-a1f1-511bec38afa5" containerName="glance-log" containerID="cri-o://8b51d4553f38927a8858bdd05f11beca552b9840963ba79d51f84d4acf38c78d" gracePeriod=30 Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.744369 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="733c0bd0-5736-4e03-a1f1-511bec38afa5" containerName="glance-httpd" containerID="cri-o://90d31f83804ed13a91183dda31d25b53af7d8d17aef41209f1ad59dc5b899cd0" gracePeriod=30 Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.781512 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ec899b91-662f-40d8-8ac2-2b5baa1d792b-horizon-secret-key\") pod \"horizon-77c64b89f5-b5lmn\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.781910 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec899b91-662f-40d8-8ac2-2b5baa1d792b-logs\") pod \"horizon-77c64b89f5-b5lmn\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.781956 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec899b91-662f-40d8-8ac2-2b5baa1d792b-scripts\") pod \"horizon-77c64b89f5-b5lmn\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.781988 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec899b91-662f-40d8-8ac2-2b5baa1d792b-config-data\") pod \"horizon-77c64b89f5-b5lmn\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.782077 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np5x5\" (UniqueName: \"kubernetes.io/projected/ec899b91-662f-40d8-8ac2-2b5baa1d792b-kube-api-access-np5x5\") pod \"horizon-77c64b89f5-b5lmn\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.807517 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5fd6b56965-bqtxr"] Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.809264 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.830485 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.830777 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0983ed44-4f02-4751-9452-5f11f3b93f63" containerName="glance-log" containerID="cri-o://ae9ff29232f7941f7167b41adcc8e22cc47ae1c89feeaf93f421084fd933f3c6" gracePeriod=30 Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.830923 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0983ed44-4f02-4751-9452-5f11f3b93f63" containerName="glance-httpd" containerID="cri-o://ee588007f28e21e91e570b04e85a69b74ab3215ab7c4d7a965388796b2558902" gracePeriod=30 Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.843710 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5fd6b56965-bqtxr"] Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.887908 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz2mq\" (UniqueName: \"kubernetes.io/projected/840981cb-1e16-4feb-bdc8-791ef7766906-kube-api-access-dz2mq\") pod \"horizon-5fd6b56965-bqtxr\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.888012 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ec899b91-662f-40d8-8ac2-2b5baa1d792b-horizon-secret-key\") pod \"horizon-77c64b89f5-b5lmn\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.888037 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/840981cb-1e16-4feb-bdc8-791ef7766906-horizon-secret-key\") pod \"horizon-5fd6b56965-bqtxr\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.888064 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/840981cb-1e16-4feb-bdc8-791ef7766906-config-data\") pod \"horizon-5fd6b56965-bqtxr\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.888093 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec899b91-662f-40d8-8ac2-2b5baa1d792b-logs\") pod \"horizon-77c64b89f5-b5lmn\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.888138 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec899b91-662f-40d8-8ac2-2b5baa1d792b-scripts\") pod \"horizon-77c64b89f5-b5lmn\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.888163 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec899b91-662f-40d8-8ac2-2b5baa1d792b-config-data\") pod \"horizon-77c64b89f5-b5lmn\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.888243 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/840981cb-1e16-4feb-bdc8-791ef7766906-logs\") pod \"horizon-5fd6b56965-bqtxr\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.888286 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/840981cb-1e16-4feb-bdc8-791ef7766906-scripts\") pod \"horizon-5fd6b56965-bqtxr\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.888313 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np5x5\" (UniqueName: \"kubernetes.io/projected/ec899b91-662f-40d8-8ac2-2b5baa1d792b-kube-api-access-np5x5\") pod \"horizon-77c64b89f5-b5lmn\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.891816 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec899b91-662f-40d8-8ac2-2b5baa1d792b-logs\") pod \"horizon-77c64b89f5-b5lmn\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.892791 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec899b91-662f-40d8-8ac2-2b5baa1d792b-scripts\") pod \"horizon-77c64b89f5-b5lmn\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.893018 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec899b91-662f-40d8-8ac2-2b5baa1d792b-config-data\") pod \"horizon-77c64b89f5-b5lmn\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.908034 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ec899b91-662f-40d8-8ac2-2b5baa1d792b-horizon-secret-key\") pod \"horizon-77c64b89f5-b5lmn\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.908196 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np5x5\" (UniqueName: \"kubernetes.io/projected/ec899b91-662f-40d8-8ac2-2b5baa1d792b-kube-api-access-np5x5\") pod \"horizon-77c64b89f5-b5lmn\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.989794 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz2mq\" (UniqueName: \"kubernetes.io/projected/840981cb-1e16-4feb-bdc8-791ef7766906-kube-api-access-dz2mq\") pod \"horizon-5fd6b56965-bqtxr\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.989888 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/840981cb-1e16-4feb-bdc8-791ef7766906-horizon-secret-key\") pod \"horizon-5fd6b56965-bqtxr\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.989913 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/840981cb-1e16-4feb-bdc8-791ef7766906-config-data\") pod \"horizon-5fd6b56965-bqtxr\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.989954 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/840981cb-1e16-4feb-bdc8-791ef7766906-logs\") pod \"horizon-5fd6b56965-bqtxr\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.989986 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/840981cb-1e16-4feb-bdc8-791ef7766906-scripts\") pod \"horizon-5fd6b56965-bqtxr\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.990387 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/840981cb-1e16-4feb-bdc8-791ef7766906-logs\") pod \"horizon-5fd6b56965-bqtxr\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.990746 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/840981cb-1e16-4feb-bdc8-791ef7766906-scripts\") pod \"horizon-5fd6b56965-bqtxr\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.991183 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/840981cb-1e16-4feb-bdc8-791ef7766906-config-data\") pod \"horizon-5fd6b56965-bqtxr\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:01:48 crc kubenswrapper[4997]: I1205 09:01:48.993541 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/840981cb-1e16-4feb-bdc8-791ef7766906-horizon-secret-key\") pod \"horizon-5fd6b56965-bqtxr\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.005589 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz2mq\" (UniqueName: \"kubernetes.io/projected/840981cb-1e16-4feb-bdc8-791ef7766906-kube-api-access-dz2mq\") pod \"horizon-5fd6b56965-bqtxr\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.030669 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.137108 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.290973 4997 generic.go:334] "Generic (PLEG): container finished" podID="733c0bd0-5736-4e03-a1f1-511bec38afa5" containerID="8b51d4553f38927a8858bdd05f11beca552b9840963ba79d51f84d4acf38c78d" exitCode=143 Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.291023 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"733c0bd0-5736-4e03-a1f1-511bec38afa5","Type":"ContainerDied","Data":"8b51d4553f38927a8858bdd05f11beca552b9840963ba79d51f84d4acf38c78d"} Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.294113 4997 generic.go:334] "Generic (PLEG): container finished" podID="0983ed44-4f02-4751-9452-5f11f3b93f63" containerID="ae9ff29232f7941f7167b41adcc8e22cc47ae1c89feeaf93f421084fd933f3c6" exitCode=143 Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.294145 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0983ed44-4f02-4751-9452-5f11f3b93f63","Type":"ContainerDied","Data":"ae9ff29232f7941f7167b41adcc8e22cc47ae1c89feeaf93f421084fd933f3c6"} Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.332212 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5fd6b56965-bqtxr"] Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.367569 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-79f85b4d6f-sj7dx"] Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.370172 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.392234 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79f85b4d6f-sj7dx"] Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.414499 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2bfb275-8e76-4add-9610-23777e579acb-logs\") pod \"horizon-79f85b4d6f-sj7dx\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.415094 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f2bfb275-8e76-4add-9610-23777e579acb-config-data\") pod \"horizon-79f85b4d6f-sj7dx\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.415204 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f2bfb275-8e76-4add-9610-23777e579acb-horizon-secret-key\") pod \"horizon-79f85b4d6f-sj7dx\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.415245 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2bfb275-8e76-4add-9610-23777e579acb-scripts\") pod \"horizon-79f85b4d6f-sj7dx\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.415298 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbxjb\" (UniqueName: \"kubernetes.io/projected/f2bfb275-8e76-4add-9610-23777e579acb-kube-api-access-qbxjb\") pod \"horizon-79f85b4d6f-sj7dx\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.516502 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-77c64b89f5-b5lmn"] Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.517311 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2bfb275-8e76-4add-9610-23777e579acb-logs\") pod \"horizon-79f85b4d6f-sj7dx\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.517419 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f2bfb275-8e76-4add-9610-23777e579acb-config-data\") pod \"horizon-79f85b4d6f-sj7dx\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.517493 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f2bfb275-8e76-4add-9610-23777e579acb-horizon-secret-key\") pod \"horizon-79f85b4d6f-sj7dx\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.517522 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2bfb275-8e76-4add-9610-23777e579acb-scripts\") pod \"horizon-79f85b4d6f-sj7dx\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.517562 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbxjb\" (UniqueName: \"kubernetes.io/projected/f2bfb275-8e76-4add-9610-23777e579acb-kube-api-access-qbxjb\") pod \"horizon-79f85b4d6f-sj7dx\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.518399 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2bfb275-8e76-4add-9610-23777e579acb-logs\") pod \"horizon-79f85b4d6f-sj7dx\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.520281 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2bfb275-8e76-4add-9610-23777e579acb-scripts\") pod \"horizon-79f85b4d6f-sj7dx\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.520464 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f2bfb275-8e76-4add-9610-23777e579acb-config-data\") pod \"horizon-79f85b4d6f-sj7dx\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.523270 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f2bfb275-8e76-4add-9610-23777e579acb-horizon-secret-key\") pod \"horizon-79f85b4d6f-sj7dx\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.540832 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbxjb\" (UniqueName: \"kubernetes.io/projected/f2bfb275-8e76-4add-9610-23777e579acb-kube-api-access-qbxjb\") pod \"horizon-79f85b4d6f-sj7dx\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.699229 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:49 crc kubenswrapper[4997]: I1205 09:01:49.770249 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5fd6b56965-bqtxr"] Dec 05 09:01:50 crc kubenswrapper[4997]: I1205 09:01:50.160098 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79f85b4d6f-sj7dx"] Dec 05 09:01:50 crc kubenswrapper[4997]: W1205 09:01:50.161387 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2bfb275_8e76_4add_9610_23777e579acb.slice/crio-1ae6c4d10b89a7895d3c0b4ef62ecbf631c7e4246086e0079ca0c844b31eede0 WatchSource:0}: Error finding container 1ae6c4d10b89a7895d3c0b4ef62ecbf631c7e4246086e0079ca0c844b31eede0: Status 404 returned error can't find the container with id 1ae6c4d10b89a7895d3c0b4ef62ecbf631c7e4246086e0079ca0c844b31eede0 Dec 05 09:01:50 crc kubenswrapper[4997]: I1205 09:01:50.310720 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5fd6b56965-bqtxr" event={"ID":"840981cb-1e16-4feb-bdc8-791ef7766906","Type":"ContainerStarted","Data":"5370e006fc13c7d8fd10309e5d82e5763460942204fcd6e2b6c81bb4558e6501"} Dec 05 09:01:50 crc kubenswrapper[4997]: I1205 09:01:50.319484 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77c64b89f5-b5lmn" event={"ID":"ec899b91-662f-40d8-8ac2-2b5baa1d792b","Type":"ContainerStarted","Data":"e209dd48a6acd3b2d6b07e673c44ee8ace1eb37789d4b7e4f4cdbad25bd9de04"} Dec 05 09:01:50 crc kubenswrapper[4997]: I1205 09:01:50.321237 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f85b4d6f-sj7dx" event={"ID":"f2bfb275-8e76-4add-9610-23777e579acb","Type":"ContainerStarted","Data":"1ae6c4d10b89a7895d3c0b4ef62ecbf631c7e4246086e0079ca0c844b31eede0"} Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.345550 4997 generic.go:334] "Generic (PLEG): container finished" podID="0983ed44-4f02-4751-9452-5f11f3b93f63" containerID="ee588007f28e21e91e570b04e85a69b74ab3215ab7c4d7a965388796b2558902" exitCode=0 Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.347003 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0983ed44-4f02-4751-9452-5f11f3b93f63","Type":"ContainerDied","Data":"ee588007f28e21e91e570b04e85a69b74ab3215ab7c4d7a965388796b2558902"} Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.353929 4997 generic.go:334] "Generic (PLEG): container finished" podID="733c0bd0-5736-4e03-a1f1-511bec38afa5" containerID="90d31f83804ed13a91183dda31d25b53af7d8d17aef41209f1ad59dc5b899cd0" exitCode=0 Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.354008 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"733c0bd0-5736-4e03-a1f1-511bec38afa5","Type":"ContainerDied","Data":"90d31f83804ed13a91183dda31d25b53af7d8d17aef41209f1ad59dc5b899cd0"} Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.654211 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.685140 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-combined-ca-bundle\") pod \"0983ed44-4f02-4751-9452-5f11f3b93f63\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.685190 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-scripts\") pod \"0983ed44-4f02-4751-9452-5f11f3b93f63\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.685213 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-config-data\") pod \"0983ed44-4f02-4751-9452-5f11f3b93f63\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.685232 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0983ed44-4f02-4751-9452-5f11f3b93f63-logs\") pod \"0983ed44-4f02-4751-9452-5f11f3b93f63\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.685283 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0983ed44-4f02-4751-9452-5f11f3b93f63-ceph\") pod \"0983ed44-4f02-4751-9452-5f11f3b93f63\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.685331 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwzn2\" (UniqueName: \"kubernetes.io/projected/0983ed44-4f02-4751-9452-5f11f3b93f63-kube-api-access-kwzn2\") pod \"0983ed44-4f02-4751-9452-5f11f3b93f63\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.685372 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0983ed44-4f02-4751-9452-5f11f3b93f63-httpd-run\") pod \"0983ed44-4f02-4751-9452-5f11f3b93f63\" (UID: \"0983ed44-4f02-4751-9452-5f11f3b93f63\") " Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.686485 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0983ed44-4f02-4751-9452-5f11f3b93f63-logs" (OuterVolumeSpecName: "logs") pod "0983ed44-4f02-4751-9452-5f11f3b93f63" (UID: "0983ed44-4f02-4751-9452-5f11f3b93f63"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.686602 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0983ed44-4f02-4751-9452-5f11f3b93f63-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0983ed44-4f02-4751-9452-5f11f3b93f63" (UID: "0983ed44-4f02-4751-9452-5f11f3b93f63"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.691251 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0983ed44-4f02-4751-9452-5f11f3b93f63-kube-api-access-kwzn2" (OuterVolumeSpecName: "kube-api-access-kwzn2") pod "0983ed44-4f02-4751-9452-5f11f3b93f63" (UID: "0983ed44-4f02-4751-9452-5f11f3b93f63"). InnerVolumeSpecName "kube-api-access-kwzn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.691948 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0983ed44-4f02-4751-9452-5f11f3b93f63-ceph" (OuterVolumeSpecName: "ceph") pod "0983ed44-4f02-4751-9452-5f11f3b93f63" (UID: "0983ed44-4f02-4751-9452-5f11f3b93f63"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.699179 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-scripts" (OuterVolumeSpecName: "scripts") pod "0983ed44-4f02-4751-9452-5f11f3b93f63" (UID: "0983ed44-4f02-4751-9452-5f11f3b93f63"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.716946 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0983ed44-4f02-4751-9452-5f11f3b93f63" (UID: "0983ed44-4f02-4751-9452-5f11f3b93f63"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.754731 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-config-data" (OuterVolumeSpecName: "config-data") pod "0983ed44-4f02-4751-9452-5f11f3b93f63" (UID: "0983ed44-4f02-4751-9452-5f11f3b93f63"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.783029 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.787208 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw87h\" (UniqueName: \"kubernetes.io/projected/733c0bd0-5736-4e03-a1f1-511bec38afa5-kube-api-access-rw87h\") pod \"733c0bd0-5736-4e03-a1f1-511bec38afa5\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.788314 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.788347 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.788360 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0983ed44-4f02-4751-9452-5f11f3b93f63-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.788373 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0983ed44-4f02-4751-9452-5f11f3b93f63-logs\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.788384 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0983ed44-4f02-4751-9452-5f11f3b93f63-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.788398 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwzn2\" (UniqueName: \"kubernetes.io/projected/0983ed44-4f02-4751-9452-5f11f3b93f63-kube-api-access-kwzn2\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.788410 4997 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0983ed44-4f02-4751-9452-5f11f3b93f63-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.791915 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/733c0bd0-5736-4e03-a1f1-511bec38afa5-kube-api-access-rw87h" (OuterVolumeSpecName: "kube-api-access-rw87h") pod "733c0bd0-5736-4e03-a1f1-511bec38afa5" (UID: "733c0bd0-5736-4e03-a1f1-511bec38afa5"). InnerVolumeSpecName "kube-api-access-rw87h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.889082 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-combined-ca-bundle\") pod \"733c0bd0-5736-4e03-a1f1-511bec38afa5\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.889160 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/733c0bd0-5736-4e03-a1f1-511bec38afa5-ceph\") pod \"733c0bd0-5736-4e03-a1f1-511bec38afa5\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.889235 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/733c0bd0-5736-4e03-a1f1-511bec38afa5-httpd-run\") pod \"733c0bd0-5736-4e03-a1f1-511bec38afa5\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.889370 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-config-data\") pod \"733c0bd0-5736-4e03-a1f1-511bec38afa5\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.889490 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/733c0bd0-5736-4e03-a1f1-511bec38afa5-logs\") pod \"733c0bd0-5736-4e03-a1f1-511bec38afa5\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.889513 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-scripts\") pod \"733c0bd0-5736-4e03-a1f1-511bec38afa5\" (UID: \"733c0bd0-5736-4e03-a1f1-511bec38afa5\") " Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.889663 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/733c0bd0-5736-4e03-a1f1-511bec38afa5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "733c0bd0-5736-4e03-a1f1-511bec38afa5" (UID: "733c0bd0-5736-4e03-a1f1-511bec38afa5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.889955 4997 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/733c0bd0-5736-4e03-a1f1-511bec38afa5-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.889964 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/733c0bd0-5736-4e03-a1f1-511bec38afa5-logs" (OuterVolumeSpecName: "logs") pod "733c0bd0-5736-4e03-a1f1-511bec38afa5" (UID: "733c0bd0-5736-4e03-a1f1-511bec38afa5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.889970 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw87h\" (UniqueName: \"kubernetes.io/projected/733c0bd0-5736-4e03-a1f1-511bec38afa5-kube-api-access-rw87h\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.900398 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/733c0bd0-5736-4e03-a1f1-511bec38afa5-ceph" (OuterVolumeSpecName: "ceph") pod "733c0bd0-5736-4e03-a1f1-511bec38afa5" (UID: "733c0bd0-5736-4e03-a1f1-511bec38afa5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.902142 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-scripts" (OuterVolumeSpecName: "scripts") pod "733c0bd0-5736-4e03-a1f1-511bec38afa5" (UID: "733c0bd0-5736-4e03-a1f1-511bec38afa5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.924655 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "733c0bd0-5736-4e03-a1f1-511bec38afa5" (UID: "733c0bd0-5736-4e03-a1f1-511bec38afa5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.936844 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-config-data" (OuterVolumeSpecName: "config-data") pod "733c0bd0-5736-4e03-a1f1-511bec38afa5" (UID: "733c0bd0-5736-4e03-a1f1-511bec38afa5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.991306 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.991341 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/733c0bd0-5736-4e03-a1f1-511bec38afa5-logs\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.991355 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.991365 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/733c0bd0-5736-4e03-a1f1-511bec38afa5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:52 crc kubenswrapper[4997]: I1205 09:01:52.991376 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/733c0bd0-5736-4e03-a1f1-511bec38afa5-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.367048 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"733c0bd0-5736-4e03-a1f1-511bec38afa5","Type":"ContainerDied","Data":"d32bae6e0b7b66e761af3c3c095e3ce992e22a147f70f75c5879e8c61ed921d7"} Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.367101 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.367121 4997 scope.go:117] "RemoveContainer" containerID="90d31f83804ed13a91183dda31d25b53af7d8d17aef41209f1ad59dc5b899cd0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.370280 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0983ed44-4f02-4751-9452-5f11f3b93f63","Type":"ContainerDied","Data":"4a4fe9700e9911685553c713d615525ae0f51e09fa4a569e56ea8eb330876c76"} Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.370358 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.411842 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.422187 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.432448 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.444573 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.456570 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 09:01:53 crc kubenswrapper[4997]: E1205 09:01:53.457977 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0983ed44-4f02-4751-9452-5f11f3b93f63" containerName="glance-log" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.458003 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0983ed44-4f02-4751-9452-5f11f3b93f63" containerName="glance-log" Dec 05 09:01:53 crc kubenswrapper[4997]: E1205 09:01:53.458019 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="733c0bd0-5736-4e03-a1f1-511bec38afa5" containerName="glance-httpd" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.458027 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="733c0bd0-5736-4e03-a1f1-511bec38afa5" containerName="glance-httpd" Dec 05 09:01:53 crc kubenswrapper[4997]: E1205 09:01:53.458057 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0983ed44-4f02-4751-9452-5f11f3b93f63" containerName="glance-httpd" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.458065 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0983ed44-4f02-4751-9452-5f11f3b93f63" containerName="glance-httpd" Dec 05 09:01:53 crc kubenswrapper[4997]: E1205 09:01:53.458083 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="733c0bd0-5736-4e03-a1f1-511bec38afa5" containerName="glance-log" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.458091 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="733c0bd0-5736-4e03-a1f1-511bec38afa5" containerName="glance-log" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.458939 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="733c0bd0-5736-4e03-a1f1-511bec38afa5" containerName="glance-log" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.458959 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0983ed44-4f02-4751-9452-5f11f3b93f63" containerName="glance-log" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.458975 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0983ed44-4f02-4751-9452-5f11f3b93f63" containerName="glance-httpd" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.458993 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="733c0bd0-5736-4e03-a1f1-511bec38afa5" containerName="glance-httpd" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.460333 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.462346 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.463007 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-x844b" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.463169 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.470730 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.472532 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.475029 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.501517 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.505136 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26891501-25d1-4d26-947e-b23b45116b35-config-data\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.505192 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26891501-25d1-4d26-947e-b23b45116b35-logs\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.505237 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26891501-25d1-4d26-947e-b23b45116b35-scripts\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.505270 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26891501-25d1-4d26-947e-b23b45116b35-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.505290 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1c47845d-3866-47ce-9242-5b315f1a565c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.505326 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7fx8\" (UniqueName: \"kubernetes.io/projected/1c47845d-3866-47ce-9242-5b315f1a565c-kube-api-access-p7fx8\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.505368 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xnzs\" (UniqueName: \"kubernetes.io/projected/26891501-25d1-4d26-947e-b23b45116b35-kube-api-access-2xnzs\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.505401 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c47845d-3866-47ce-9242-5b315f1a565c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.505450 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c47845d-3866-47ce-9242-5b315f1a565c-config-data\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.505512 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26891501-25d1-4d26-947e-b23b45116b35-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.505533 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c47845d-3866-47ce-9242-5b315f1a565c-scripts\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.510862 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1c47845d-3866-47ce-9242-5b315f1a565c-ceph\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.510901 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/26891501-25d1-4d26-947e-b23b45116b35-ceph\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.510925 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c47845d-3866-47ce-9242-5b315f1a565c-logs\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.517859 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.613238 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26891501-25d1-4d26-947e-b23b45116b35-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.613608 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c47845d-3866-47ce-9242-5b315f1a565c-scripts\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.613704 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1c47845d-3866-47ce-9242-5b315f1a565c-ceph\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.613728 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/26891501-25d1-4d26-947e-b23b45116b35-ceph\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.613760 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c47845d-3866-47ce-9242-5b315f1a565c-logs\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.613983 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26891501-25d1-4d26-947e-b23b45116b35-config-data\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.614000 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26891501-25d1-4d26-947e-b23b45116b35-logs\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.614852 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c47845d-3866-47ce-9242-5b315f1a565c-logs\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.615938 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26891501-25d1-4d26-947e-b23b45116b35-scripts\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.615999 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26891501-25d1-4d26-947e-b23b45116b35-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.616057 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1c47845d-3866-47ce-9242-5b315f1a565c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.616691 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26891501-25d1-4d26-947e-b23b45116b35-logs\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.616793 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7fx8\" (UniqueName: \"kubernetes.io/projected/1c47845d-3866-47ce-9242-5b315f1a565c-kube-api-access-p7fx8\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.617095 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1c47845d-3866-47ce-9242-5b315f1a565c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.618142 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26891501-25d1-4d26-947e-b23b45116b35-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.618854 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xnzs\" (UniqueName: \"kubernetes.io/projected/26891501-25d1-4d26-947e-b23b45116b35-kube-api-access-2xnzs\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.618911 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c47845d-3866-47ce-9242-5b315f1a565c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.619560 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c47845d-3866-47ce-9242-5b315f1a565c-config-data\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.621832 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1c47845d-3866-47ce-9242-5b315f1a565c-ceph\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.622648 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/26891501-25d1-4d26-947e-b23b45116b35-ceph\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.623034 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26891501-25d1-4d26-947e-b23b45116b35-config-data\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.626987 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c47845d-3866-47ce-9242-5b315f1a565c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.627105 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26891501-25d1-4d26-947e-b23b45116b35-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.631959 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26891501-25d1-4d26-947e-b23b45116b35-scripts\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.640076 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xnzs\" (UniqueName: \"kubernetes.io/projected/26891501-25d1-4d26-947e-b23b45116b35-kube-api-access-2xnzs\") pod \"glance-default-internal-api-0\" (UID: \"26891501-25d1-4d26-947e-b23b45116b35\") " pod="openstack/glance-default-internal-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.641587 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c47845d-3866-47ce-9242-5b315f1a565c-scripts\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.645144 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7fx8\" (UniqueName: \"kubernetes.io/projected/1c47845d-3866-47ce-9242-5b315f1a565c-kube-api-access-p7fx8\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.652805 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c47845d-3866-47ce-9242-5b315f1a565c-config-data\") pod \"glance-default-external-api-0\" (UID: \"1c47845d-3866-47ce-9242-5b315f1a565c\") " pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.776959 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0983ed44-4f02-4751-9452-5f11f3b93f63" path="/var/lib/kubelet/pods/0983ed44-4f02-4751-9452-5f11f3b93f63/volumes" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.778756 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="733c0bd0-5736-4e03-a1f1-511bec38afa5" path="/var/lib/kubelet/pods/733c0bd0-5736-4e03-a1f1-511bec38afa5/volumes" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.788373 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 09:01:53 crc kubenswrapper[4997]: I1205 09:01:53.812249 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 09:01:58 crc kubenswrapper[4997]: I1205 09:01:58.112387 4997 scope.go:117] "RemoveContainer" containerID="8b51d4553f38927a8858bdd05f11beca552b9840963ba79d51f84d4acf38c78d" Dec 05 09:01:58 crc kubenswrapper[4997]: I1205 09:01:58.189072 4997 scope.go:117] "RemoveContainer" containerID="ee588007f28e21e91e570b04e85a69b74ab3215ab7c4d7a965388796b2558902" Dec 05 09:01:58 crc kubenswrapper[4997]: I1205 09:01:58.314212 4997 scope.go:117] "RemoveContainer" containerID="ae9ff29232f7941f7167b41adcc8e22cc47ae1c89feeaf93f421084fd933f3c6" Dec 05 09:01:58 crc kubenswrapper[4997]: I1205 09:01:58.750354 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:01:58 crc kubenswrapper[4997]: E1205 09:01:58.750837 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:01:58 crc kubenswrapper[4997]: I1205 09:01:58.830367 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 09:01:58 crc kubenswrapper[4997]: I1205 09:01:58.986312 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 09:01:58 crc kubenswrapper[4997]: W1205 09:01:58.995425 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26891501_25d1_4d26_947e_b23b45116b35.slice/crio-f93cba742b3c94b9b360ff65fe8bc9e76bfcac11d62f335b74c56aec17cb6c0a WatchSource:0}: Error finding container f93cba742b3c94b9b360ff65fe8bc9e76bfcac11d62f335b74c56aec17cb6c0a: Status 404 returned error can't find the container with id f93cba742b3c94b9b360ff65fe8bc9e76bfcac11d62f335b74c56aec17cb6c0a Dec 05 09:01:59 crc kubenswrapper[4997]: I1205 09:01:59.464821 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"26891501-25d1-4d26-947e-b23b45116b35","Type":"ContainerStarted","Data":"f93cba742b3c94b9b360ff65fe8bc9e76bfcac11d62f335b74c56aec17cb6c0a"} Dec 05 09:01:59 crc kubenswrapper[4997]: I1205 09:01:59.471008 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f85b4d6f-sj7dx" event={"ID":"f2bfb275-8e76-4add-9610-23777e579acb","Type":"ContainerStarted","Data":"f3e3db4d78eefda48aab2857f9352c5a3c9b62108ad86e29c998a741e7145762"} Dec 05 09:01:59 crc kubenswrapper[4997]: I1205 09:01:59.471222 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f85b4d6f-sj7dx" event={"ID":"f2bfb275-8e76-4add-9610-23777e579acb","Type":"ContainerStarted","Data":"4cab7fc650e874a372fb16bdf1f14882c2fa2c3c26fcd6c4da16d0098e44360f"} Dec 05 09:01:59 crc kubenswrapper[4997]: I1205 09:01:59.476690 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1c47845d-3866-47ce-9242-5b315f1a565c","Type":"ContainerStarted","Data":"b2a2499c18236da175deb6c5d30f2a315174999788e95307081fd587b60012d5"} Dec 05 09:01:59 crc kubenswrapper[4997]: I1205 09:01:59.476980 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1c47845d-3866-47ce-9242-5b315f1a565c","Type":"ContainerStarted","Data":"2635900b98f1968294810b41c1b3b7ea91a82e21878f176ee6f2a97dbe52c321"} Dec 05 09:01:59 crc kubenswrapper[4997]: I1205 09:01:59.478949 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5fd6b56965-bqtxr" event={"ID":"840981cb-1e16-4feb-bdc8-791ef7766906","Type":"ContainerStarted","Data":"a854273e2a3148b39803f48efe80a75fb58e87c717a7be646d4e2398148f5a01"} Dec 05 09:01:59 crc kubenswrapper[4997]: I1205 09:01:59.479058 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5fd6b56965-bqtxr" event={"ID":"840981cb-1e16-4feb-bdc8-791ef7766906","Type":"ContainerStarted","Data":"1ed127fab05163a2e8bd84ed6fc5eda788be2ab683968d4699190055c5b1b19b"} Dec 05 09:01:59 crc kubenswrapper[4997]: I1205 09:01:59.479055 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5fd6b56965-bqtxr" podUID="840981cb-1e16-4feb-bdc8-791ef7766906" containerName="horizon-log" containerID="cri-o://1ed127fab05163a2e8bd84ed6fc5eda788be2ab683968d4699190055c5b1b19b" gracePeriod=30 Dec 05 09:01:59 crc kubenswrapper[4997]: I1205 09:01:59.479312 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5fd6b56965-bqtxr" podUID="840981cb-1e16-4feb-bdc8-791ef7766906" containerName="horizon" containerID="cri-o://a854273e2a3148b39803f48efe80a75fb58e87c717a7be646d4e2398148f5a01" gracePeriod=30 Dec 05 09:01:59 crc kubenswrapper[4997]: I1205 09:01:59.491087 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77c64b89f5-b5lmn" event={"ID":"ec899b91-662f-40d8-8ac2-2b5baa1d792b","Type":"ContainerStarted","Data":"b2655ddd009036466f65fb20a1134979ed48bcb2b1f18b966cf71d26fd6bca48"} Dec 05 09:01:59 crc kubenswrapper[4997]: I1205 09:01:59.491238 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77c64b89f5-b5lmn" event={"ID":"ec899b91-662f-40d8-8ac2-2b5baa1d792b","Type":"ContainerStarted","Data":"aa758f39c4bf96e6aae04ea5105924470727290a9b4b0d1b28afdef0af900f47"} Dec 05 09:01:59 crc kubenswrapper[4997]: I1205 09:01:59.514375 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-79f85b4d6f-sj7dx" podStartSLOduration=2.347332195 podStartE2EDuration="10.51435782s" podCreationTimestamp="2025-12-05 09:01:49 +0000 UTC" firstStartedPulling="2025-12-05 09:01:50.165637284 +0000 UTC m=+7610.694544545" lastFinishedPulling="2025-12-05 09:01:58.332662909 +0000 UTC m=+7618.861570170" observedRunningTime="2025-12-05 09:01:59.496102748 +0000 UTC m=+7620.025010009" watchObservedRunningTime="2025-12-05 09:01:59.51435782 +0000 UTC m=+7620.043265081" Dec 05 09:01:59 crc kubenswrapper[4997]: I1205 09:01:59.518092 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-77c64b89f5-b5lmn" podStartSLOduration=2.835254015 podStartE2EDuration="11.518083351s" podCreationTimestamp="2025-12-05 09:01:48 +0000 UTC" firstStartedPulling="2025-12-05 09:01:49.530138655 +0000 UTC m=+7610.059045916" lastFinishedPulling="2025-12-05 09:01:58.212967991 +0000 UTC m=+7618.741875252" observedRunningTime="2025-12-05 09:01:59.515366758 +0000 UTC m=+7620.044274019" watchObservedRunningTime="2025-12-05 09:01:59.518083351 +0000 UTC m=+7620.046990602" Dec 05 09:01:59 crc kubenswrapper[4997]: I1205 09:01:59.554311 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5fd6b56965-bqtxr" podStartSLOduration=3.086546323 podStartE2EDuration="11.554281928s" podCreationTimestamp="2025-12-05 09:01:48 +0000 UTC" firstStartedPulling="2025-12-05 09:01:49.769793298 +0000 UTC m=+7610.298700559" lastFinishedPulling="2025-12-05 09:01:58.237528903 +0000 UTC m=+7618.766436164" observedRunningTime="2025-12-05 09:01:59.543007194 +0000 UTC m=+7620.071914455" watchObservedRunningTime="2025-12-05 09:01:59.554281928 +0000 UTC m=+7620.083189189" Dec 05 09:01:59 crc kubenswrapper[4997]: I1205 09:01:59.699492 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:01:59 crc kubenswrapper[4997]: I1205 09:01:59.699855 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:02:00 crc kubenswrapper[4997]: I1205 09:02:00.516439 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"26891501-25d1-4d26-947e-b23b45116b35","Type":"ContainerStarted","Data":"b1addc1a20608446b048793dad4a3d2f11daedd08c505a9b160ba40346dde18d"} Dec 05 09:02:00 crc kubenswrapper[4997]: I1205 09:02:00.517765 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"26891501-25d1-4d26-947e-b23b45116b35","Type":"ContainerStarted","Data":"583acf59bec0645958c77dd1890a52fc099de5ee3fcd1e8f3940095f5ea14d49"} Dec 05 09:02:00 crc kubenswrapper[4997]: I1205 09:02:00.520991 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1c47845d-3866-47ce-9242-5b315f1a565c","Type":"ContainerStarted","Data":"050a8846e2b02b7cc7ced91a29c78d47a457911123a574bbce55b8bd6d799609"} Dec 05 09:02:00 crc kubenswrapper[4997]: I1205 09:02:00.554340 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.55431578 podStartE2EDuration="7.55431578s" podCreationTimestamp="2025-12-05 09:01:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:02:00.538466352 +0000 UTC m=+7621.067373633" watchObservedRunningTime="2025-12-05 09:02:00.55431578 +0000 UTC m=+7621.083223041" Dec 05 09:02:00 crc kubenswrapper[4997]: I1205 09:02:00.567776 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.567752382 podStartE2EDuration="7.567752382s" podCreationTimestamp="2025-12-05 09:01:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:02:00.553915099 +0000 UTC m=+7621.082822360" watchObservedRunningTime="2025-12-05 09:02:00.567752382 +0000 UTC m=+7621.096659643" Dec 05 09:02:03 crc kubenswrapper[4997]: I1205 09:02:03.789031 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 09:02:03 crc kubenswrapper[4997]: I1205 09:02:03.789635 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 09:02:03 crc kubenswrapper[4997]: I1205 09:02:03.813445 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 09:02:03 crc kubenswrapper[4997]: I1205 09:02:03.813517 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 09:02:03 crc kubenswrapper[4997]: I1205 09:02:03.827513 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 09:02:03 crc kubenswrapper[4997]: I1205 09:02:03.837530 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 09:02:03 crc kubenswrapper[4997]: I1205 09:02:03.854075 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 09:02:03 crc kubenswrapper[4997]: I1205 09:02:03.927551 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 09:02:04 crc kubenswrapper[4997]: I1205 09:02:04.566215 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 09:02:04 crc kubenswrapper[4997]: I1205 09:02:04.566273 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 09:02:04 crc kubenswrapper[4997]: I1205 09:02:04.566309 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 09:02:04 crc kubenswrapper[4997]: I1205 09:02:04.566335 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 09:02:06 crc kubenswrapper[4997]: I1205 09:02:06.578624 4997 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 09:02:06 crc kubenswrapper[4997]: I1205 09:02:06.579784 4997 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 09:02:06 crc kubenswrapper[4997]: I1205 09:02:06.783571 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 09:02:06 crc kubenswrapper[4997]: I1205 09:02:06.785924 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 09:02:06 crc kubenswrapper[4997]: I1205 09:02:06.907641 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 09:02:06 crc kubenswrapper[4997]: I1205 09:02:06.911438 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 09:02:09 crc kubenswrapper[4997]: I1205 09:02:09.031119 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:02:09 crc kubenswrapper[4997]: I1205 09:02:09.031484 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:02:09 crc kubenswrapper[4997]: I1205 09:02:09.032730 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-77c64b89f5-b5lmn" podUID="ec899b91-662f-40d8-8ac2-2b5baa1d792b" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.109:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.109:8080: connect: connection refused" Dec 05 09:02:09 crc kubenswrapper[4997]: I1205 09:02:09.137893 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:02:09 crc kubenswrapper[4997]: I1205 09:02:09.701052 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-79f85b4d6f-sj7dx" podUID="f2bfb275-8e76-4add-9610-23777e579acb" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.111:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.111:8080: connect: connection refused" Dec 05 09:02:13 crc kubenswrapper[4997]: I1205 09:02:13.749229 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:02:13 crc kubenswrapper[4997]: E1205 09:02:13.750010 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:02:16 crc kubenswrapper[4997]: I1205 09:02:16.185486 4997 scope.go:117] "RemoveContainer" containerID="54fc76a663c1f19d294beeaff96a9faae058033865ef0a7de32aa3f7206c65e7" Dec 05 09:02:20 crc kubenswrapper[4997]: I1205 09:02:20.896172 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:02:21 crc kubenswrapper[4997]: I1205 09:02:21.463316 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:02:22 crc kubenswrapper[4997]: I1205 09:02:22.491244 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:02:23 crc kubenswrapper[4997]: I1205 09:02:23.220248 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:02:23 crc kubenswrapper[4997]: I1205 09:02:23.275076 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-77c64b89f5-b5lmn"] Dec 05 09:02:23 crc kubenswrapper[4997]: I1205 09:02:23.275267 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-77c64b89f5-b5lmn" podUID="ec899b91-662f-40d8-8ac2-2b5baa1d792b" containerName="horizon-log" containerID="cri-o://aa758f39c4bf96e6aae04ea5105924470727290a9b4b0d1b28afdef0af900f47" gracePeriod=30 Dec 05 09:02:23 crc kubenswrapper[4997]: I1205 09:02:23.275720 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-77c64b89f5-b5lmn" podUID="ec899b91-662f-40d8-8ac2-2b5baa1d792b" containerName="horizon" containerID="cri-o://b2655ddd009036466f65fb20a1134979ed48bcb2b1f18b966cf71d26fd6bca48" gracePeriod=30 Dec 05 09:02:26 crc kubenswrapper[4997]: I1205 09:02:26.792193 4997 generic.go:334] "Generic (PLEG): container finished" podID="ec899b91-662f-40d8-8ac2-2b5baa1d792b" containerID="b2655ddd009036466f65fb20a1134979ed48bcb2b1f18b966cf71d26fd6bca48" exitCode=0 Dec 05 09:02:26 crc kubenswrapper[4997]: I1205 09:02:26.792245 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77c64b89f5-b5lmn" event={"ID":"ec899b91-662f-40d8-8ac2-2b5baa1d792b","Type":"ContainerDied","Data":"b2655ddd009036466f65fb20a1134979ed48bcb2b1f18b966cf71d26fd6bca48"} Dec 05 09:02:28 crc kubenswrapper[4997]: I1205 09:02:28.749339 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:02:28 crc kubenswrapper[4997]: E1205 09:02:28.750151 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:02:29 crc kubenswrapper[4997]: I1205 09:02:29.032304 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-77c64b89f5-b5lmn" podUID="ec899b91-662f-40d8-8ac2-2b5baa1d792b" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.109:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.109:8080: connect: connection refused" Dec 05 09:02:29 crc kubenswrapper[4997]: I1205 09:02:29.819513 4997 generic.go:334] "Generic (PLEG): container finished" podID="840981cb-1e16-4feb-bdc8-791ef7766906" containerID="a854273e2a3148b39803f48efe80a75fb58e87c717a7be646d4e2398148f5a01" exitCode=137 Dec 05 09:02:29 crc kubenswrapper[4997]: I1205 09:02:29.819953 4997 generic.go:334] "Generic (PLEG): container finished" podID="840981cb-1e16-4feb-bdc8-791ef7766906" containerID="1ed127fab05163a2e8bd84ed6fc5eda788be2ab683968d4699190055c5b1b19b" exitCode=137 Dec 05 09:02:29 crc kubenswrapper[4997]: I1205 09:02:29.820662 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5fd6b56965-bqtxr" event={"ID":"840981cb-1e16-4feb-bdc8-791ef7766906","Type":"ContainerDied","Data":"a854273e2a3148b39803f48efe80a75fb58e87c717a7be646d4e2398148f5a01"} Dec 05 09:02:29 crc kubenswrapper[4997]: I1205 09:02:29.820703 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5fd6b56965-bqtxr" event={"ID":"840981cb-1e16-4feb-bdc8-791ef7766906","Type":"ContainerDied","Data":"1ed127fab05163a2e8bd84ed6fc5eda788be2ab683968d4699190055c5b1b19b"} Dec 05 09:02:29 crc kubenswrapper[4997]: I1205 09:02:29.974169 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.117809 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dz2mq\" (UniqueName: \"kubernetes.io/projected/840981cb-1e16-4feb-bdc8-791ef7766906-kube-api-access-dz2mq\") pod \"840981cb-1e16-4feb-bdc8-791ef7766906\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.118122 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/840981cb-1e16-4feb-bdc8-791ef7766906-config-data\") pod \"840981cb-1e16-4feb-bdc8-791ef7766906\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.118218 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/840981cb-1e16-4feb-bdc8-791ef7766906-logs\") pod \"840981cb-1e16-4feb-bdc8-791ef7766906\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.118261 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/840981cb-1e16-4feb-bdc8-791ef7766906-horizon-secret-key\") pod \"840981cb-1e16-4feb-bdc8-791ef7766906\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.118305 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/840981cb-1e16-4feb-bdc8-791ef7766906-scripts\") pod \"840981cb-1e16-4feb-bdc8-791ef7766906\" (UID: \"840981cb-1e16-4feb-bdc8-791ef7766906\") " Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.120207 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/840981cb-1e16-4feb-bdc8-791ef7766906-logs" (OuterVolumeSpecName: "logs") pod "840981cb-1e16-4feb-bdc8-791ef7766906" (UID: "840981cb-1e16-4feb-bdc8-791ef7766906"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.125498 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/840981cb-1e16-4feb-bdc8-791ef7766906-kube-api-access-dz2mq" (OuterVolumeSpecName: "kube-api-access-dz2mq") pod "840981cb-1e16-4feb-bdc8-791ef7766906" (UID: "840981cb-1e16-4feb-bdc8-791ef7766906"). InnerVolumeSpecName "kube-api-access-dz2mq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.126035 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/840981cb-1e16-4feb-bdc8-791ef7766906-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "840981cb-1e16-4feb-bdc8-791ef7766906" (UID: "840981cb-1e16-4feb-bdc8-791ef7766906"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.142257 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/840981cb-1e16-4feb-bdc8-791ef7766906-scripts" (OuterVolumeSpecName: "scripts") pod "840981cb-1e16-4feb-bdc8-791ef7766906" (UID: "840981cb-1e16-4feb-bdc8-791ef7766906"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.146142 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/840981cb-1e16-4feb-bdc8-791ef7766906-config-data" (OuterVolumeSpecName: "config-data") pod "840981cb-1e16-4feb-bdc8-791ef7766906" (UID: "840981cb-1e16-4feb-bdc8-791ef7766906"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.220501 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dz2mq\" (UniqueName: \"kubernetes.io/projected/840981cb-1e16-4feb-bdc8-791ef7766906-kube-api-access-dz2mq\") on node \"crc\" DevicePath \"\"" Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.220551 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/840981cb-1e16-4feb-bdc8-791ef7766906-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.220562 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/840981cb-1e16-4feb-bdc8-791ef7766906-logs\") on node \"crc\" DevicePath \"\"" Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.220570 4997 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/840981cb-1e16-4feb-bdc8-791ef7766906-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.220579 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/840981cb-1e16-4feb-bdc8-791ef7766906-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.829855 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5fd6b56965-bqtxr" event={"ID":"840981cb-1e16-4feb-bdc8-791ef7766906","Type":"ContainerDied","Data":"5370e006fc13c7d8fd10309e5d82e5763460942204fcd6e2b6c81bb4558e6501"} Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.829899 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5fd6b56965-bqtxr" Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.829908 4997 scope.go:117] "RemoveContainer" containerID="a854273e2a3148b39803f48efe80a75fb58e87c717a7be646d4e2398148f5a01" Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.873043 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5fd6b56965-bqtxr"] Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.885032 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5fd6b56965-bqtxr"] Dec 05 09:02:30 crc kubenswrapper[4997]: I1205 09:02:30.990138 4997 scope.go:117] "RemoveContainer" containerID="1ed127fab05163a2e8bd84ed6fc5eda788be2ab683968d4699190055c5b1b19b" Dec 05 09:02:31 crc kubenswrapper[4997]: I1205 09:02:31.761343 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="840981cb-1e16-4feb-bdc8-791ef7766906" path="/var/lib/kubelet/pods/840981cb-1e16-4feb-bdc8-791ef7766906/volumes" Dec 05 09:02:39 crc kubenswrapper[4997]: I1205 09:02:39.032001 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-77c64b89f5-b5lmn" podUID="ec899b91-662f-40d8-8ac2-2b5baa1d792b" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.109:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.109:8080: connect: connection refused" Dec 05 09:02:39 crc kubenswrapper[4997]: I1205 09:02:39.755942 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:02:39 crc kubenswrapper[4997]: E1205 09:02:39.756274 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:02:49 crc kubenswrapper[4997]: I1205 09:02:49.031352 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-77c64b89f5-b5lmn" podUID="ec899b91-662f-40d8-8ac2-2b5baa1d792b" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.109:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.109:8080: connect: connection refused" Dec 05 09:02:49 crc kubenswrapper[4997]: I1205 09:02:49.031889 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:02:50 crc kubenswrapper[4997]: I1205 09:02:50.749552 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:02:50 crc kubenswrapper[4997]: E1205 09:02:50.749920 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:02:53 crc kubenswrapper[4997]: I1205 09:02:53.746597 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:02:53 crc kubenswrapper[4997]: I1205 09:02:53.786638 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec899b91-662f-40d8-8ac2-2b5baa1d792b-logs\") pod \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " Dec 05 09:02:53 crc kubenswrapper[4997]: I1205 09:02:53.786714 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec899b91-662f-40d8-8ac2-2b5baa1d792b-scripts\") pod \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " Dec 05 09:02:53 crc kubenswrapper[4997]: I1205 09:02:53.786813 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ec899b91-662f-40d8-8ac2-2b5baa1d792b-horizon-secret-key\") pod \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " Dec 05 09:02:53 crc kubenswrapper[4997]: I1205 09:02:53.786840 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-np5x5\" (UniqueName: \"kubernetes.io/projected/ec899b91-662f-40d8-8ac2-2b5baa1d792b-kube-api-access-np5x5\") pod \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " Dec 05 09:02:53 crc kubenswrapper[4997]: I1205 09:02:53.786981 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec899b91-662f-40d8-8ac2-2b5baa1d792b-config-data\") pod \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\" (UID: \"ec899b91-662f-40d8-8ac2-2b5baa1d792b\") " Dec 05 09:02:53 crc kubenswrapper[4997]: I1205 09:02:53.788410 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec899b91-662f-40d8-8ac2-2b5baa1d792b-logs" (OuterVolumeSpecName: "logs") pod "ec899b91-662f-40d8-8ac2-2b5baa1d792b" (UID: "ec899b91-662f-40d8-8ac2-2b5baa1d792b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:02:53 crc kubenswrapper[4997]: I1205 09:02:53.792502 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec899b91-662f-40d8-8ac2-2b5baa1d792b-kube-api-access-np5x5" (OuterVolumeSpecName: "kube-api-access-np5x5") pod "ec899b91-662f-40d8-8ac2-2b5baa1d792b" (UID: "ec899b91-662f-40d8-8ac2-2b5baa1d792b"). InnerVolumeSpecName "kube-api-access-np5x5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:02:53 crc kubenswrapper[4997]: I1205 09:02:53.801271 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec899b91-662f-40d8-8ac2-2b5baa1d792b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ec899b91-662f-40d8-8ac2-2b5baa1d792b" (UID: "ec899b91-662f-40d8-8ac2-2b5baa1d792b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:02:53 crc kubenswrapper[4997]: I1205 09:02:53.810940 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec899b91-662f-40d8-8ac2-2b5baa1d792b-config-data" (OuterVolumeSpecName: "config-data") pod "ec899b91-662f-40d8-8ac2-2b5baa1d792b" (UID: "ec899b91-662f-40d8-8ac2-2b5baa1d792b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:02:53 crc kubenswrapper[4997]: I1205 09:02:53.816226 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec899b91-662f-40d8-8ac2-2b5baa1d792b-scripts" (OuterVolumeSpecName: "scripts") pod "ec899b91-662f-40d8-8ac2-2b5baa1d792b" (UID: "ec899b91-662f-40d8-8ac2-2b5baa1d792b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:02:53 crc kubenswrapper[4997]: I1205 09:02:53.888903 4997 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ec899b91-662f-40d8-8ac2-2b5baa1d792b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:02:53 crc kubenswrapper[4997]: I1205 09:02:53.888945 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-np5x5\" (UniqueName: \"kubernetes.io/projected/ec899b91-662f-40d8-8ac2-2b5baa1d792b-kube-api-access-np5x5\") on node \"crc\" DevicePath \"\"" Dec 05 09:02:53 crc kubenswrapper[4997]: I1205 09:02:53.888957 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec899b91-662f-40d8-8ac2-2b5baa1d792b-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:02:53 crc kubenswrapper[4997]: I1205 09:02:53.888970 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec899b91-662f-40d8-8ac2-2b5baa1d792b-logs\") on node \"crc\" DevicePath \"\"" Dec 05 09:02:53 crc kubenswrapper[4997]: I1205 09:02:53.888978 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec899b91-662f-40d8-8ac2-2b5baa1d792b-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:02:54 crc kubenswrapper[4997]: I1205 09:02:54.050469 4997 generic.go:334] "Generic (PLEG): container finished" podID="ec899b91-662f-40d8-8ac2-2b5baa1d792b" containerID="aa758f39c4bf96e6aae04ea5105924470727290a9b4b0d1b28afdef0af900f47" exitCode=137 Dec 05 09:02:54 crc kubenswrapper[4997]: I1205 09:02:54.050517 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77c64b89f5-b5lmn" event={"ID":"ec899b91-662f-40d8-8ac2-2b5baa1d792b","Type":"ContainerDied","Data":"aa758f39c4bf96e6aae04ea5105924470727290a9b4b0d1b28afdef0af900f47"} Dec 05 09:02:54 crc kubenswrapper[4997]: I1205 09:02:54.050555 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77c64b89f5-b5lmn" Dec 05 09:02:54 crc kubenswrapper[4997]: I1205 09:02:54.050572 4997 scope.go:117] "RemoveContainer" containerID="b2655ddd009036466f65fb20a1134979ed48bcb2b1f18b966cf71d26fd6bca48" Dec 05 09:02:54 crc kubenswrapper[4997]: I1205 09:02:54.050559 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77c64b89f5-b5lmn" event={"ID":"ec899b91-662f-40d8-8ac2-2b5baa1d792b","Type":"ContainerDied","Data":"e209dd48a6acd3b2d6b07e673c44ee8ace1eb37789d4b7e4f4cdbad25bd9de04"} Dec 05 09:02:54 crc kubenswrapper[4997]: I1205 09:02:54.090587 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-77c64b89f5-b5lmn"] Dec 05 09:02:54 crc kubenswrapper[4997]: I1205 09:02:54.099418 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-77c64b89f5-b5lmn"] Dec 05 09:02:54 crc kubenswrapper[4997]: I1205 09:02:54.210178 4997 scope.go:117] "RemoveContainer" containerID="aa758f39c4bf96e6aae04ea5105924470727290a9b4b0d1b28afdef0af900f47" Dec 05 09:02:54 crc kubenswrapper[4997]: I1205 09:02:54.228473 4997 scope.go:117] "RemoveContainer" containerID="b2655ddd009036466f65fb20a1134979ed48bcb2b1f18b966cf71d26fd6bca48" Dec 05 09:02:54 crc kubenswrapper[4997]: E1205 09:02:54.228989 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2655ddd009036466f65fb20a1134979ed48bcb2b1f18b966cf71d26fd6bca48\": container with ID starting with b2655ddd009036466f65fb20a1134979ed48bcb2b1f18b966cf71d26fd6bca48 not found: ID does not exist" containerID="b2655ddd009036466f65fb20a1134979ed48bcb2b1f18b966cf71d26fd6bca48" Dec 05 09:02:54 crc kubenswrapper[4997]: I1205 09:02:54.229031 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2655ddd009036466f65fb20a1134979ed48bcb2b1f18b966cf71d26fd6bca48"} err="failed to get container status \"b2655ddd009036466f65fb20a1134979ed48bcb2b1f18b966cf71d26fd6bca48\": rpc error: code = NotFound desc = could not find container \"b2655ddd009036466f65fb20a1134979ed48bcb2b1f18b966cf71d26fd6bca48\": container with ID starting with b2655ddd009036466f65fb20a1134979ed48bcb2b1f18b966cf71d26fd6bca48 not found: ID does not exist" Dec 05 09:02:54 crc kubenswrapper[4997]: I1205 09:02:54.229057 4997 scope.go:117] "RemoveContainer" containerID="aa758f39c4bf96e6aae04ea5105924470727290a9b4b0d1b28afdef0af900f47" Dec 05 09:02:54 crc kubenswrapper[4997]: E1205 09:02:54.229335 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa758f39c4bf96e6aae04ea5105924470727290a9b4b0d1b28afdef0af900f47\": container with ID starting with aa758f39c4bf96e6aae04ea5105924470727290a9b4b0d1b28afdef0af900f47 not found: ID does not exist" containerID="aa758f39c4bf96e6aae04ea5105924470727290a9b4b0d1b28afdef0af900f47" Dec 05 09:02:54 crc kubenswrapper[4997]: I1205 09:02:54.229425 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa758f39c4bf96e6aae04ea5105924470727290a9b4b0d1b28afdef0af900f47"} err="failed to get container status \"aa758f39c4bf96e6aae04ea5105924470727290a9b4b0d1b28afdef0af900f47\": rpc error: code = NotFound desc = could not find container \"aa758f39c4bf96e6aae04ea5105924470727290a9b4b0d1b28afdef0af900f47\": container with ID starting with aa758f39c4bf96e6aae04ea5105924470727290a9b4b0d1b28afdef0af900f47 not found: ID does not exist" Dec 05 09:02:55 crc kubenswrapper[4997]: I1205 09:02:55.760087 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec899b91-662f-40d8-8ac2-2b5baa1d792b" path="/var/lib/kubelet/pods/ec899b91-662f-40d8-8ac2-2b5baa1d792b/volumes" Dec 05 09:03:01 crc kubenswrapper[4997]: I1205 09:03:01.749236 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:03:01 crc kubenswrapper[4997]: E1205 09:03:01.751079 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.040920 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-678ddff49c-bvhj8"] Dec 05 09:03:06 crc kubenswrapper[4997]: E1205 09:03:06.041711 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec899b91-662f-40d8-8ac2-2b5baa1d792b" containerName="horizon-log" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.041734 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec899b91-662f-40d8-8ac2-2b5baa1d792b" containerName="horizon-log" Dec 05 09:03:06 crc kubenswrapper[4997]: E1205 09:03:06.041745 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="840981cb-1e16-4feb-bdc8-791ef7766906" containerName="horizon-log" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.041751 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="840981cb-1e16-4feb-bdc8-791ef7766906" containerName="horizon-log" Dec 05 09:03:06 crc kubenswrapper[4997]: E1205 09:03:06.041766 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="840981cb-1e16-4feb-bdc8-791ef7766906" containerName="horizon" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.041772 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="840981cb-1e16-4feb-bdc8-791ef7766906" containerName="horizon" Dec 05 09:03:06 crc kubenswrapper[4997]: E1205 09:03:06.041785 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec899b91-662f-40d8-8ac2-2b5baa1d792b" containerName="horizon" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.041790 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec899b91-662f-40d8-8ac2-2b5baa1d792b" containerName="horizon" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.041998 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec899b91-662f-40d8-8ac2-2b5baa1d792b" containerName="horizon" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.042009 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="840981cb-1e16-4feb-bdc8-791ef7766906" containerName="horizon" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.042026 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec899b91-662f-40d8-8ac2-2b5baa1d792b" containerName="horizon-log" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.042035 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="840981cb-1e16-4feb-bdc8-791ef7766906" containerName="horizon-log" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.043200 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.057643 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-678ddff49c-bvhj8"] Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.201734 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2da24366-b581-40f0-94bf-2a9652e2840a-horizon-secret-key\") pod \"horizon-678ddff49c-bvhj8\" (UID: \"2da24366-b581-40f0-94bf-2a9652e2840a\") " pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.201836 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2da24366-b581-40f0-94bf-2a9652e2840a-logs\") pod \"horizon-678ddff49c-bvhj8\" (UID: \"2da24366-b581-40f0-94bf-2a9652e2840a\") " pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.201877 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2da24366-b581-40f0-94bf-2a9652e2840a-scripts\") pod \"horizon-678ddff49c-bvhj8\" (UID: \"2da24366-b581-40f0-94bf-2a9652e2840a\") " pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.201908 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwqdd\" (UniqueName: \"kubernetes.io/projected/2da24366-b581-40f0-94bf-2a9652e2840a-kube-api-access-cwqdd\") pod \"horizon-678ddff49c-bvhj8\" (UID: \"2da24366-b581-40f0-94bf-2a9652e2840a\") " pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.201957 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2da24366-b581-40f0-94bf-2a9652e2840a-config-data\") pod \"horizon-678ddff49c-bvhj8\" (UID: \"2da24366-b581-40f0-94bf-2a9652e2840a\") " pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.303334 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2da24366-b581-40f0-94bf-2a9652e2840a-logs\") pod \"horizon-678ddff49c-bvhj8\" (UID: \"2da24366-b581-40f0-94bf-2a9652e2840a\") " pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.303426 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2da24366-b581-40f0-94bf-2a9652e2840a-scripts\") pod \"horizon-678ddff49c-bvhj8\" (UID: \"2da24366-b581-40f0-94bf-2a9652e2840a\") " pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.303462 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwqdd\" (UniqueName: \"kubernetes.io/projected/2da24366-b581-40f0-94bf-2a9652e2840a-kube-api-access-cwqdd\") pod \"horizon-678ddff49c-bvhj8\" (UID: \"2da24366-b581-40f0-94bf-2a9652e2840a\") " pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.303537 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2da24366-b581-40f0-94bf-2a9652e2840a-config-data\") pod \"horizon-678ddff49c-bvhj8\" (UID: \"2da24366-b581-40f0-94bf-2a9652e2840a\") " pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.303574 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2da24366-b581-40f0-94bf-2a9652e2840a-horizon-secret-key\") pod \"horizon-678ddff49c-bvhj8\" (UID: \"2da24366-b581-40f0-94bf-2a9652e2840a\") " pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.304345 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2da24366-b581-40f0-94bf-2a9652e2840a-logs\") pod \"horizon-678ddff49c-bvhj8\" (UID: \"2da24366-b581-40f0-94bf-2a9652e2840a\") " pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.304701 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2da24366-b581-40f0-94bf-2a9652e2840a-scripts\") pod \"horizon-678ddff49c-bvhj8\" (UID: \"2da24366-b581-40f0-94bf-2a9652e2840a\") " pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.305196 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2da24366-b581-40f0-94bf-2a9652e2840a-config-data\") pod \"horizon-678ddff49c-bvhj8\" (UID: \"2da24366-b581-40f0-94bf-2a9652e2840a\") " pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.309532 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2da24366-b581-40f0-94bf-2a9652e2840a-horizon-secret-key\") pod \"horizon-678ddff49c-bvhj8\" (UID: \"2da24366-b581-40f0-94bf-2a9652e2840a\") " pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.326126 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwqdd\" (UniqueName: \"kubernetes.io/projected/2da24366-b581-40f0-94bf-2a9652e2840a-kube-api-access-cwqdd\") pod \"horizon-678ddff49c-bvhj8\" (UID: \"2da24366-b581-40f0-94bf-2a9652e2840a\") " pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.359847 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:06 crc kubenswrapper[4997]: I1205 09:03:06.803404 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-678ddff49c-bvhj8"] Dec 05 09:03:06 crc kubenswrapper[4997]: W1205 09:03:06.808234 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2da24366_b581_40f0_94bf_2a9652e2840a.slice/crio-9a964eedb76e886bba58b52b87b7cf4fd8d1c5f38ded303ff19ed8d58efc7070 WatchSource:0}: Error finding container 9a964eedb76e886bba58b52b87b7cf4fd8d1c5f38ded303ff19ed8d58efc7070: Status 404 returned error can't find the container with id 9a964eedb76e886bba58b52b87b7cf4fd8d1c5f38ded303ff19ed8d58efc7070 Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.227659 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-678ddff49c-bvhj8" event={"ID":"2da24366-b581-40f0-94bf-2a9652e2840a","Type":"ContainerStarted","Data":"c182efd9f18432f9e8638c9cefb5feb3236cd20fc4ab8c0d2a6670c35331f054"} Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.227711 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-678ddff49c-bvhj8" event={"ID":"2da24366-b581-40f0-94bf-2a9652e2840a","Type":"ContainerStarted","Data":"c30027e8ff2a843109be80744121e418dfeb723d6d27219d27f780e0eafa5080"} Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.227722 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-678ddff49c-bvhj8" event={"ID":"2da24366-b581-40f0-94bf-2a9652e2840a","Type":"ContainerStarted","Data":"9a964eedb76e886bba58b52b87b7cf4fd8d1c5f38ded303ff19ed8d58efc7070"} Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.258980 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-678ddff49c-bvhj8" podStartSLOduration=2.258955791 podStartE2EDuration="2.258955791s" podCreationTimestamp="2025-12-05 09:03:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:03:07.247118432 +0000 UTC m=+7687.776025703" watchObservedRunningTime="2025-12-05 09:03:07.258955791 +0000 UTC m=+7687.787863052" Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.271567 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-5bmw7"] Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.273206 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-5bmw7" Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.284232 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-5bmw7"] Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.395487 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-139c-account-create-update-5wbwg"] Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.396734 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-139c-account-create-update-5wbwg" Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.402187 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.414711 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-139c-account-create-update-5wbwg"] Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.430461 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66qd8\" (UniqueName: \"kubernetes.io/projected/787316ba-2693-4628-a18a-076ddbf78183-kube-api-access-66qd8\") pod \"heat-db-create-5bmw7\" (UID: \"787316ba-2693-4628-a18a-076ddbf78183\") " pod="openstack/heat-db-create-5bmw7" Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.430548 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/787316ba-2693-4628-a18a-076ddbf78183-operator-scripts\") pod \"heat-db-create-5bmw7\" (UID: \"787316ba-2693-4628-a18a-076ddbf78183\") " pod="openstack/heat-db-create-5bmw7" Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.532008 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/787316ba-2693-4628-a18a-076ddbf78183-operator-scripts\") pod \"heat-db-create-5bmw7\" (UID: \"787316ba-2693-4628-a18a-076ddbf78183\") " pod="openstack/heat-db-create-5bmw7" Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.532101 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608681d2-0ae1-4fb0-ba87-7db90a81e82a-operator-scripts\") pod \"heat-139c-account-create-update-5wbwg\" (UID: \"608681d2-0ae1-4fb0-ba87-7db90a81e82a\") " pod="openstack/heat-139c-account-create-update-5wbwg" Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.532153 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m99ph\" (UniqueName: \"kubernetes.io/projected/608681d2-0ae1-4fb0-ba87-7db90a81e82a-kube-api-access-m99ph\") pod \"heat-139c-account-create-update-5wbwg\" (UID: \"608681d2-0ae1-4fb0-ba87-7db90a81e82a\") " pod="openstack/heat-139c-account-create-update-5wbwg" Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.532219 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66qd8\" (UniqueName: \"kubernetes.io/projected/787316ba-2693-4628-a18a-076ddbf78183-kube-api-access-66qd8\") pod \"heat-db-create-5bmw7\" (UID: \"787316ba-2693-4628-a18a-076ddbf78183\") " pod="openstack/heat-db-create-5bmw7" Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.533269 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/787316ba-2693-4628-a18a-076ddbf78183-operator-scripts\") pod \"heat-db-create-5bmw7\" (UID: \"787316ba-2693-4628-a18a-076ddbf78183\") " pod="openstack/heat-db-create-5bmw7" Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.550053 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66qd8\" (UniqueName: \"kubernetes.io/projected/787316ba-2693-4628-a18a-076ddbf78183-kube-api-access-66qd8\") pod \"heat-db-create-5bmw7\" (UID: \"787316ba-2693-4628-a18a-076ddbf78183\") " pod="openstack/heat-db-create-5bmw7" Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.590785 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-5bmw7" Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.633989 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608681d2-0ae1-4fb0-ba87-7db90a81e82a-operator-scripts\") pod \"heat-139c-account-create-update-5wbwg\" (UID: \"608681d2-0ae1-4fb0-ba87-7db90a81e82a\") " pod="openstack/heat-139c-account-create-update-5wbwg" Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.634085 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m99ph\" (UniqueName: \"kubernetes.io/projected/608681d2-0ae1-4fb0-ba87-7db90a81e82a-kube-api-access-m99ph\") pod \"heat-139c-account-create-update-5wbwg\" (UID: \"608681d2-0ae1-4fb0-ba87-7db90a81e82a\") " pod="openstack/heat-139c-account-create-update-5wbwg" Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.635153 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608681d2-0ae1-4fb0-ba87-7db90a81e82a-operator-scripts\") pod \"heat-139c-account-create-update-5wbwg\" (UID: \"608681d2-0ae1-4fb0-ba87-7db90a81e82a\") " pod="openstack/heat-139c-account-create-update-5wbwg" Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.659252 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m99ph\" (UniqueName: \"kubernetes.io/projected/608681d2-0ae1-4fb0-ba87-7db90a81e82a-kube-api-access-m99ph\") pod \"heat-139c-account-create-update-5wbwg\" (UID: \"608681d2-0ae1-4fb0-ba87-7db90a81e82a\") " pod="openstack/heat-139c-account-create-update-5wbwg" Dec 05 09:03:07 crc kubenswrapper[4997]: I1205 09:03:07.742580 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-139c-account-create-update-5wbwg" Dec 05 09:03:08 crc kubenswrapper[4997]: I1205 09:03:08.029524 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-5bmw7"] Dec 05 09:03:08 crc kubenswrapper[4997]: W1205 09:03:08.032109 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod787316ba_2693_4628_a18a_076ddbf78183.slice/crio-1367107dde00126fb14190f89406791802ad074092bfdb69ef2b3f9309909be2 WatchSource:0}: Error finding container 1367107dde00126fb14190f89406791802ad074092bfdb69ef2b3f9309909be2: Status 404 returned error can't find the container with id 1367107dde00126fb14190f89406791802ad074092bfdb69ef2b3f9309909be2 Dec 05 09:03:08 crc kubenswrapper[4997]: I1205 09:03:08.225373 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-139c-account-create-update-5wbwg"] Dec 05 09:03:08 crc kubenswrapper[4997]: I1205 09:03:08.236839 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-139c-account-create-update-5wbwg" event={"ID":"608681d2-0ae1-4fb0-ba87-7db90a81e82a","Type":"ContainerStarted","Data":"4d7dfb32585e0c3a4717fa2259a2cb9fc824d9a9c00e82007d9c12b00e54c41c"} Dec 05 09:03:08 crc kubenswrapper[4997]: I1205 09:03:08.238635 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-5bmw7" event={"ID":"787316ba-2693-4628-a18a-076ddbf78183","Type":"ContainerStarted","Data":"5ffe0c3c8085a13be53ec7e83f755705efe1bd8ad3ea194984140d28d9e95363"} Dec 05 09:03:08 crc kubenswrapper[4997]: I1205 09:03:08.238686 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-5bmw7" event={"ID":"787316ba-2693-4628-a18a-076ddbf78183","Type":"ContainerStarted","Data":"1367107dde00126fb14190f89406791802ad074092bfdb69ef2b3f9309909be2"} Dec 05 09:03:08 crc kubenswrapper[4997]: I1205 09:03:08.261838 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-5bmw7" podStartSLOduration=1.26181729 podStartE2EDuration="1.26181729s" podCreationTimestamp="2025-12-05 09:03:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:03:08.255879749 +0000 UTC m=+7688.784787060" watchObservedRunningTime="2025-12-05 09:03:08.26181729 +0000 UTC m=+7688.790724551" Dec 05 09:03:09 crc kubenswrapper[4997]: I1205 09:03:09.250766 4997 generic.go:334] "Generic (PLEG): container finished" podID="608681d2-0ae1-4fb0-ba87-7db90a81e82a" containerID="483475421a032ea533f681073ac0ddcd26c08b5c3807f97e64018163865d929c" exitCode=0 Dec 05 09:03:09 crc kubenswrapper[4997]: I1205 09:03:09.250858 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-139c-account-create-update-5wbwg" event={"ID":"608681d2-0ae1-4fb0-ba87-7db90a81e82a","Type":"ContainerDied","Data":"483475421a032ea533f681073ac0ddcd26c08b5c3807f97e64018163865d929c"} Dec 05 09:03:09 crc kubenswrapper[4997]: I1205 09:03:09.254234 4997 generic.go:334] "Generic (PLEG): container finished" podID="787316ba-2693-4628-a18a-076ddbf78183" containerID="5ffe0c3c8085a13be53ec7e83f755705efe1bd8ad3ea194984140d28d9e95363" exitCode=0 Dec 05 09:03:09 crc kubenswrapper[4997]: I1205 09:03:09.254306 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-5bmw7" event={"ID":"787316ba-2693-4628-a18a-076ddbf78183","Type":"ContainerDied","Data":"5ffe0c3c8085a13be53ec7e83f755705efe1bd8ad3ea194984140d28d9e95363"} Dec 05 09:03:10 crc kubenswrapper[4997]: I1205 09:03:10.708207 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-139c-account-create-update-5wbwg" Dec 05 09:03:10 crc kubenswrapper[4997]: I1205 09:03:10.718367 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-5bmw7" Dec 05 09:03:10 crc kubenswrapper[4997]: I1205 09:03:10.818022 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608681d2-0ae1-4fb0-ba87-7db90a81e82a-operator-scripts\") pod \"608681d2-0ae1-4fb0-ba87-7db90a81e82a\" (UID: \"608681d2-0ae1-4fb0-ba87-7db90a81e82a\") " Dec 05 09:03:10 crc kubenswrapper[4997]: I1205 09:03:10.818099 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/787316ba-2693-4628-a18a-076ddbf78183-operator-scripts\") pod \"787316ba-2693-4628-a18a-076ddbf78183\" (UID: \"787316ba-2693-4628-a18a-076ddbf78183\") " Dec 05 09:03:10 crc kubenswrapper[4997]: I1205 09:03:10.818180 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66qd8\" (UniqueName: \"kubernetes.io/projected/787316ba-2693-4628-a18a-076ddbf78183-kube-api-access-66qd8\") pod \"787316ba-2693-4628-a18a-076ddbf78183\" (UID: \"787316ba-2693-4628-a18a-076ddbf78183\") " Dec 05 09:03:10 crc kubenswrapper[4997]: I1205 09:03:10.818313 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m99ph\" (UniqueName: \"kubernetes.io/projected/608681d2-0ae1-4fb0-ba87-7db90a81e82a-kube-api-access-m99ph\") pod \"608681d2-0ae1-4fb0-ba87-7db90a81e82a\" (UID: \"608681d2-0ae1-4fb0-ba87-7db90a81e82a\") " Dec 05 09:03:10 crc kubenswrapper[4997]: I1205 09:03:10.819054 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/787316ba-2693-4628-a18a-076ddbf78183-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "787316ba-2693-4628-a18a-076ddbf78183" (UID: "787316ba-2693-4628-a18a-076ddbf78183"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:03:10 crc kubenswrapper[4997]: I1205 09:03:10.819378 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/608681d2-0ae1-4fb0-ba87-7db90a81e82a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "608681d2-0ae1-4fb0-ba87-7db90a81e82a" (UID: "608681d2-0ae1-4fb0-ba87-7db90a81e82a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:03:10 crc kubenswrapper[4997]: I1205 09:03:10.824280 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/608681d2-0ae1-4fb0-ba87-7db90a81e82a-kube-api-access-m99ph" (OuterVolumeSpecName: "kube-api-access-m99ph") pod "608681d2-0ae1-4fb0-ba87-7db90a81e82a" (UID: "608681d2-0ae1-4fb0-ba87-7db90a81e82a"). InnerVolumeSpecName "kube-api-access-m99ph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:03:10 crc kubenswrapper[4997]: I1205 09:03:10.825711 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/787316ba-2693-4628-a18a-076ddbf78183-kube-api-access-66qd8" (OuterVolumeSpecName: "kube-api-access-66qd8") pod "787316ba-2693-4628-a18a-076ddbf78183" (UID: "787316ba-2693-4628-a18a-076ddbf78183"). InnerVolumeSpecName "kube-api-access-66qd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:03:10 crc kubenswrapper[4997]: I1205 09:03:10.926806 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m99ph\" (UniqueName: \"kubernetes.io/projected/608681d2-0ae1-4fb0-ba87-7db90a81e82a-kube-api-access-m99ph\") on node \"crc\" DevicePath \"\"" Dec 05 09:03:10 crc kubenswrapper[4997]: I1205 09:03:10.926846 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608681d2-0ae1-4fb0-ba87-7db90a81e82a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:03:10 crc kubenswrapper[4997]: I1205 09:03:10.926857 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/787316ba-2693-4628-a18a-076ddbf78183-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:03:10 crc kubenswrapper[4997]: I1205 09:03:10.926867 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66qd8\" (UniqueName: \"kubernetes.io/projected/787316ba-2693-4628-a18a-076ddbf78183-kube-api-access-66qd8\") on node \"crc\" DevicePath \"\"" Dec 05 09:03:11 crc kubenswrapper[4997]: I1205 09:03:11.279407 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-5bmw7" event={"ID":"787316ba-2693-4628-a18a-076ddbf78183","Type":"ContainerDied","Data":"1367107dde00126fb14190f89406791802ad074092bfdb69ef2b3f9309909be2"} Dec 05 09:03:11 crc kubenswrapper[4997]: I1205 09:03:11.279441 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-5bmw7" Dec 05 09:03:11 crc kubenswrapper[4997]: I1205 09:03:11.279459 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1367107dde00126fb14190f89406791802ad074092bfdb69ef2b3f9309909be2" Dec 05 09:03:11 crc kubenswrapper[4997]: I1205 09:03:11.297428 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-139c-account-create-update-5wbwg" event={"ID":"608681d2-0ae1-4fb0-ba87-7db90a81e82a","Type":"ContainerDied","Data":"4d7dfb32585e0c3a4717fa2259a2cb9fc824d9a9c00e82007d9c12b00e54c41c"} Dec 05 09:03:11 crc kubenswrapper[4997]: I1205 09:03:11.297480 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d7dfb32585e0c3a4717fa2259a2cb9fc824d9a9c00e82007d9c12b00e54c41c" Dec 05 09:03:11 crc kubenswrapper[4997]: I1205 09:03:11.297549 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-139c-account-create-update-5wbwg" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.594424 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-fmsj4"] Dec 05 09:03:12 crc kubenswrapper[4997]: E1205 09:03:12.595031 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="787316ba-2693-4628-a18a-076ddbf78183" containerName="mariadb-database-create" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.595050 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="787316ba-2693-4628-a18a-076ddbf78183" containerName="mariadb-database-create" Dec 05 09:03:12 crc kubenswrapper[4997]: E1205 09:03:12.595086 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="608681d2-0ae1-4fb0-ba87-7db90a81e82a" containerName="mariadb-account-create-update" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.595094 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="608681d2-0ae1-4fb0-ba87-7db90a81e82a" containerName="mariadb-account-create-update" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.595332 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="787316ba-2693-4628-a18a-076ddbf78183" containerName="mariadb-database-create" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.595359 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="608681d2-0ae1-4fb0-ba87-7db90a81e82a" containerName="mariadb-account-create-update" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.596151 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-fmsj4" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.601864 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-r7bvc" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.602204 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.637754 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf43729-c369-4311-90b3-57ef233b9c68-combined-ca-bundle\") pod \"heat-db-sync-fmsj4\" (UID: \"ddf43729-c369-4311-90b3-57ef233b9c68\") " pod="openstack/heat-db-sync-fmsj4" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.637853 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq5jm\" (UniqueName: \"kubernetes.io/projected/ddf43729-c369-4311-90b3-57ef233b9c68-kube-api-access-rq5jm\") pod \"heat-db-sync-fmsj4\" (UID: \"ddf43729-c369-4311-90b3-57ef233b9c68\") " pod="openstack/heat-db-sync-fmsj4" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.637928 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf43729-c369-4311-90b3-57ef233b9c68-config-data\") pod \"heat-db-sync-fmsj4\" (UID: \"ddf43729-c369-4311-90b3-57ef233b9c68\") " pod="openstack/heat-db-sync-fmsj4" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.643854 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-fmsj4"] Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.740051 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf43729-c369-4311-90b3-57ef233b9c68-combined-ca-bundle\") pod \"heat-db-sync-fmsj4\" (UID: \"ddf43729-c369-4311-90b3-57ef233b9c68\") " pod="openstack/heat-db-sync-fmsj4" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.740366 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq5jm\" (UniqueName: \"kubernetes.io/projected/ddf43729-c369-4311-90b3-57ef233b9c68-kube-api-access-rq5jm\") pod \"heat-db-sync-fmsj4\" (UID: \"ddf43729-c369-4311-90b3-57ef233b9c68\") " pod="openstack/heat-db-sync-fmsj4" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.740553 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf43729-c369-4311-90b3-57ef233b9c68-config-data\") pod \"heat-db-sync-fmsj4\" (UID: \"ddf43729-c369-4311-90b3-57ef233b9c68\") " pod="openstack/heat-db-sync-fmsj4" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.745391 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf43729-c369-4311-90b3-57ef233b9c68-combined-ca-bundle\") pod \"heat-db-sync-fmsj4\" (UID: \"ddf43729-c369-4311-90b3-57ef233b9c68\") " pod="openstack/heat-db-sync-fmsj4" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.749047 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:03:12 crc kubenswrapper[4997]: E1205 09:03:12.749345 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.754387 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf43729-c369-4311-90b3-57ef233b9c68-config-data\") pod \"heat-db-sync-fmsj4\" (UID: \"ddf43729-c369-4311-90b3-57ef233b9c68\") " pod="openstack/heat-db-sync-fmsj4" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.755286 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq5jm\" (UniqueName: \"kubernetes.io/projected/ddf43729-c369-4311-90b3-57ef233b9c68-kube-api-access-rq5jm\") pod \"heat-db-sync-fmsj4\" (UID: \"ddf43729-c369-4311-90b3-57ef233b9c68\") " pod="openstack/heat-db-sync-fmsj4" Dec 05 09:03:12 crc kubenswrapper[4997]: I1205 09:03:12.933813 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-fmsj4" Dec 05 09:03:13 crc kubenswrapper[4997]: I1205 09:03:13.403450 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-fmsj4"] Dec 05 09:03:13 crc kubenswrapper[4997]: W1205 09:03:13.403877 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddf43729_c369_4311_90b3_57ef233b9c68.slice/crio-0f7ee29889289d5634e4bf48946383809d1a5437aadc6b2246776b709b61d091 WatchSource:0}: Error finding container 0f7ee29889289d5634e4bf48946383809d1a5437aadc6b2246776b709b61d091: Status 404 returned error can't find the container with id 0f7ee29889289d5634e4bf48946383809d1a5437aadc6b2246776b709b61d091 Dec 05 09:03:14 crc kubenswrapper[4997]: I1205 09:03:14.324789 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-fmsj4" event={"ID":"ddf43729-c369-4311-90b3-57ef233b9c68","Type":"ContainerStarted","Data":"0f7ee29889289d5634e4bf48946383809d1a5437aadc6b2246776b709b61d091"} Dec 05 09:03:16 crc kubenswrapper[4997]: I1205 09:03:16.361217 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:16 crc kubenswrapper[4997]: I1205 09:03:16.361285 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:22 crc kubenswrapper[4997]: I1205 09:03:22.392703 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-fmsj4" event={"ID":"ddf43729-c369-4311-90b3-57ef233b9c68","Type":"ContainerStarted","Data":"666c39a9ab896fa3a3d555514d76123cd68f76c17b32bb1eebb272d4037b75c2"} Dec 05 09:03:22 crc kubenswrapper[4997]: I1205 09:03:22.411021 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-fmsj4" podStartSLOduration=1.879269319 podStartE2EDuration="10.41100167s" podCreationTimestamp="2025-12-05 09:03:12 +0000 UTC" firstStartedPulling="2025-12-05 09:03:13.406482007 +0000 UTC m=+7693.935389269" lastFinishedPulling="2025-12-05 09:03:21.938214369 +0000 UTC m=+7702.467121620" observedRunningTime="2025-12-05 09:03:22.410343943 +0000 UTC m=+7702.939251214" watchObservedRunningTime="2025-12-05 09:03:22.41100167 +0000 UTC m=+7702.939908931" Dec 05 09:03:24 crc kubenswrapper[4997]: I1205 09:03:24.414609 4997 generic.go:334] "Generic (PLEG): container finished" podID="ddf43729-c369-4311-90b3-57ef233b9c68" containerID="666c39a9ab896fa3a3d555514d76123cd68f76c17b32bb1eebb272d4037b75c2" exitCode=0 Dec 05 09:03:24 crc kubenswrapper[4997]: I1205 09:03:24.414646 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-fmsj4" event={"ID":"ddf43729-c369-4311-90b3-57ef233b9c68","Type":"ContainerDied","Data":"666c39a9ab896fa3a3d555514d76123cd68f76c17b32bb1eebb272d4037b75c2"} Dec 05 09:03:25 crc kubenswrapper[4997]: I1205 09:03:25.749711 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:03:25 crc kubenswrapper[4997]: E1205 09:03:25.750022 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:03:25 crc kubenswrapper[4997]: I1205 09:03:25.754766 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-fmsj4" Dec 05 09:03:25 crc kubenswrapper[4997]: I1205 09:03:25.924579 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf43729-c369-4311-90b3-57ef233b9c68-combined-ca-bundle\") pod \"ddf43729-c369-4311-90b3-57ef233b9c68\" (UID: \"ddf43729-c369-4311-90b3-57ef233b9c68\") " Dec 05 09:03:25 crc kubenswrapper[4997]: I1205 09:03:25.925175 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf43729-c369-4311-90b3-57ef233b9c68-config-data\") pod \"ddf43729-c369-4311-90b3-57ef233b9c68\" (UID: \"ddf43729-c369-4311-90b3-57ef233b9c68\") " Dec 05 09:03:25 crc kubenswrapper[4997]: I1205 09:03:25.925211 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rq5jm\" (UniqueName: \"kubernetes.io/projected/ddf43729-c369-4311-90b3-57ef233b9c68-kube-api-access-rq5jm\") pod \"ddf43729-c369-4311-90b3-57ef233b9c68\" (UID: \"ddf43729-c369-4311-90b3-57ef233b9c68\") " Dec 05 09:03:25 crc kubenswrapper[4997]: I1205 09:03:25.931208 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddf43729-c369-4311-90b3-57ef233b9c68-kube-api-access-rq5jm" (OuterVolumeSpecName: "kube-api-access-rq5jm") pod "ddf43729-c369-4311-90b3-57ef233b9c68" (UID: "ddf43729-c369-4311-90b3-57ef233b9c68"). InnerVolumeSpecName "kube-api-access-rq5jm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:03:25 crc kubenswrapper[4997]: I1205 09:03:25.949926 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf43729-c369-4311-90b3-57ef233b9c68-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ddf43729-c369-4311-90b3-57ef233b9c68" (UID: "ddf43729-c369-4311-90b3-57ef233b9c68"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:03:26 crc kubenswrapper[4997]: I1205 09:03:26.008812 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf43729-c369-4311-90b3-57ef233b9c68-config-data" (OuterVolumeSpecName: "config-data") pod "ddf43729-c369-4311-90b3-57ef233b9c68" (UID: "ddf43729-c369-4311-90b3-57ef233b9c68"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:03:26 crc kubenswrapper[4997]: I1205 09:03:26.030170 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf43729-c369-4311-90b3-57ef233b9c68-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:03:26 crc kubenswrapper[4997]: I1205 09:03:26.030204 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf43729-c369-4311-90b3-57ef233b9c68-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:03:26 crc kubenswrapper[4997]: I1205 09:03:26.030214 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rq5jm\" (UniqueName: \"kubernetes.io/projected/ddf43729-c369-4311-90b3-57ef233b9c68-kube-api-access-rq5jm\") on node \"crc\" DevicePath \"\"" Dec 05 09:03:26 crc kubenswrapper[4997]: I1205 09:03:26.434584 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-fmsj4" event={"ID":"ddf43729-c369-4311-90b3-57ef233b9c68","Type":"ContainerDied","Data":"0f7ee29889289d5634e4bf48946383809d1a5437aadc6b2246776b709b61d091"} Dec 05 09:03:26 crc kubenswrapper[4997]: I1205 09:03:26.434665 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f7ee29889289d5634e4bf48946383809d1a5437aadc6b2246776b709b61d091" Dec 05 09:03:26 crc kubenswrapper[4997]: I1205 09:03:26.434750 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-fmsj4" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.545380 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-54f49697fd-pwnb8"] Dec 05 09:03:27 crc kubenswrapper[4997]: E1205 09:03:27.547672 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddf43729-c369-4311-90b3-57ef233b9c68" containerName="heat-db-sync" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.547772 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddf43729-c369-4311-90b3-57ef233b9c68" containerName="heat-db-sync" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.548170 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddf43729-c369-4311-90b3-57ef233b9c68" containerName="heat-db-sync" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.549426 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-54f49697fd-pwnb8" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.554601 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.602656 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-r7bvc" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.607976 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.661678 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-65bb7ccfdf-b9g84"] Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.665785 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-65bb7ccfdf-b9g84" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.673780 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-54f49697fd-pwnb8"] Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.674310 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.675379 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb82741-d0c0-42af-ad5b-ac6a840393b8-config-data\") pod \"heat-engine-54f49697fd-pwnb8\" (UID: \"fcb82741-d0c0-42af-ad5b-ac6a840393b8\") " pod="openstack/heat-engine-54f49697fd-pwnb8" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.675443 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x4dm\" (UniqueName: \"kubernetes.io/projected/fcb82741-d0c0-42af-ad5b-ac6a840393b8-kube-api-access-2x4dm\") pod \"heat-engine-54f49697fd-pwnb8\" (UID: \"fcb82741-d0c0-42af-ad5b-ac6a840393b8\") " pod="openstack/heat-engine-54f49697fd-pwnb8" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.675653 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb82741-d0c0-42af-ad5b-ac6a840393b8-combined-ca-bundle\") pod \"heat-engine-54f49697fd-pwnb8\" (UID: \"fcb82741-d0c0-42af-ad5b-ac6a840393b8\") " pod="openstack/heat-engine-54f49697fd-pwnb8" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.675722 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcb82741-d0c0-42af-ad5b-ac6a840393b8-config-data-custom\") pod \"heat-engine-54f49697fd-pwnb8\" (UID: \"fcb82741-d0c0-42af-ad5b-ac6a840393b8\") " pod="openstack/heat-engine-54f49697fd-pwnb8" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.706505 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5774948dd5-879rv"] Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.708083 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5774948dd5-879rv" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.716292 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.738735 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-65bb7ccfdf-b9g84"] Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.776995 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcb82741-d0c0-42af-ad5b-ac6a840393b8-config-data-custom\") pod \"heat-engine-54f49697fd-pwnb8\" (UID: \"fcb82741-d0c0-42af-ad5b-ac6a840393b8\") " pod="openstack/heat-engine-54f49697fd-pwnb8" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.777042 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cf60578b-47bb-474a-b296-56c42105680d-config-data-custom\") pod \"heat-api-65bb7ccfdf-b9g84\" (UID: \"cf60578b-47bb-474a-b296-56c42105680d\") " pod="openstack/heat-api-65bb7ccfdf-b9g84" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.777065 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf60578b-47bb-474a-b296-56c42105680d-config-data\") pod \"heat-api-65bb7ccfdf-b9g84\" (UID: \"cf60578b-47bb-474a-b296-56c42105680d\") " pod="openstack/heat-api-65bb7ccfdf-b9g84" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.777140 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb82741-d0c0-42af-ad5b-ac6a840393b8-config-data\") pod \"heat-engine-54f49697fd-pwnb8\" (UID: \"fcb82741-d0c0-42af-ad5b-ac6a840393b8\") " pod="openstack/heat-engine-54f49697fd-pwnb8" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.777165 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x4dm\" (UniqueName: \"kubernetes.io/projected/fcb82741-d0c0-42af-ad5b-ac6a840393b8-kube-api-access-2x4dm\") pod \"heat-engine-54f49697fd-pwnb8\" (UID: \"fcb82741-d0c0-42af-ad5b-ac6a840393b8\") " pod="openstack/heat-engine-54f49697fd-pwnb8" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.777203 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltv2z\" (UniqueName: \"kubernetes.io/projected/cf60578b-47bb-474a-b296-56c42105680d-kube-api-access-ltv2z\") pod \"heat-api-65bb7ccfdf-b9g84\" (UID: \"cf60578b-47bb-474a-b296-56c42105680d\") " pod="openstack/heat-api-65bb7ccfdf-b9g84" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.777251 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf60578b-47bb-474a-b296-56c42105680d-combined-ca-bundle\") pod \"heat-api-65bb7ccfdf-b9g84\" (UID: \"cf60578b-47bb-474a-b296-56c42105680d\") " pod="openstack/heat-api-65bb7ccfdf-b9g84" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.777281 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb82741-d0c0-42af-ad5b-ac6a840393b8-combined-ca-bundle\") pod \"heat-engine-54f49697fd-pwnb8\" (UID: \"fcb82741-d0c0-42af-ad5b-ac6a840393b8\") " pod="openstack/heat-engine-54f49697fd-pwnb8" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.781092 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5774948dd5-879rv"] Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.790536 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcb82741-d0c0-42af-ad5b-ac6a840393b8-config-data-custom\") pod \"heat-engine-54f49697fd-pwnb8\" (UID: \"fcb82741-d0c0-42af-ad5b-ac6a840393b8\") " pod="openstack/heat-engine-54f49697fd-pwnb8" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.792729 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb82741-d0c0-42af-ad5b-ac6a840393b8-combined-ca-bundle\") pod \"heat-engine-54f49697fd-pwnb8\" (UID: \"fcb82741-d0c0-42af-ad5b-ac6a840393b8\") " pod="openstack/heat-engine-54f49697fd-pwnb8" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.810033 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb82741-d0c0-42af-ad5b-ac6a840393b8-config-data\") pod \"heat-engine-54f49697fd-pwnb8\" (UID: \"fcb82741-d0c0-42af-ad5b-ac6a840393b8\") " pod="openstack/heat-engine-54f49697fd-pwnb8" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.814344 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x4dm\" (UniqueName: \"kubernetes.io/projected/fcb82741-d0c0-42af-ad5b-ac6a840393b8-kube-api-access-2x4dm\") pod \"heat-engine-54f49697fd-pwnb8\" (UID: \"fcb82741-d0c0-42af-ad5b-ac6a840393b8\") " pod="openstack/heat-engine-54f49697fd-pwnb8" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.879466 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cf60578b-47bb-474a-b296-56c42105680d-config-data-custom\") pod \"heat-api-65bb7ccfdf-b9g84\" (UID: \"cf60578b-47bb-474a-b296-56c42105680d\") " pod="openstack/heat-api-65bb7ccfdf-b9g84" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.879555 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf60578b-47bb-474a-b296-56c42105680d-config-data\") pod \"heat-api-65bb7ccfdf-b9g84\" (UID: \"cf60578b-47bb-474a-b296-56c42105680d\") " pod="openstack/heat-api-65bb7ccfdf-b9g84" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.879700 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc75t\" (UniqueName: \"kubernetes.io/projected/5f69f8d5-e82e-4db6-aad0-e302e807b2b0-kube-api-access-pc75t\") pod \"heat-cfnapi-5774948dd5-879rv\" (UID: \"5f69f8d5-e82e-4db6-aad0-e302e807b2b0\") " pod="openstack/heat-cfnapi-5774948dd5-879rv" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.879761 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f69f8d5-e82e-4db6-aad0-e302e807b2b0-combined-ca-bundle\") pod \"heat-cfnapi-5774948dd5-879rv\" (UID: \"5f69f8d5-e82e-4db6-aad0-e302e807b2b0\") " pod="openstack/heat-cfnapi-5774948dd5-879rv" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.879868 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f69f8d5-e82e-4db6-aad0-e302e807b2b0-config-data\") pod \"heat-cfnapi-5774948dd5-879rv\" (UID: \"5f69f8d5-e82e-4db6-aad0-e302e807b2b0\") " pod="openstack/heat-cfnapi-5774948dd5-879rv" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.879901 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltv2z\" (UniqueName: \"kubernetes.io/projected/cf60578b-47bb-474a-b296-56c42105680d-kube-api-access-ltv2z\") pod \"heat-api-65bb7ccfdf-b9g84\" (UID: \"cf60578b-47bb-474a-b296-56c42105680d\") " pod="openstack/heat-api-65bb7ccfdf-b9g84" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.879990 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf60578b-47bb-474a-b296-56c42105680d-combined-ca-bundle\") pod \"heat-api-65bb7ccfdf-b9g84\" (UID: \"cf60578b-47bb-474a-b296-56c42105680d\") " pod="openstack/heat-api-65bb7ccfdf-b9g84" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.880060 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f69f8d5-e82e-4db6-aad0-e302e807b2b0-config-data-custom\") pod \"heat-cfnapi-5774948dd5-879rv\" (UID: \"5f69f8d5-e82e-4db6-aad0-e302e807b2b0\") " pod="openstack/heat-cfnapi-5774948dd5-879rv" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.885188 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cf60578b-47bb-474a-b296-56c42105680d-config-data-custom\") pod \"heat-api-65bb7ccfdf-b9g84\" (UID: \"cf60578b-47bb-474a-b296-56c42105680d\") " pod="openstack/heat-api-65bb7ccfdf-b9g84" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.888801 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf60578b-47bb-474a-b296-56c42105680d-config-data\") pod \"heat-api-65bb7ccfdf-b9g84\" (UID: \"cf60578b-47bb-474a-b296-56c42105680d\") " pod="openstack/heat-api-65bb7ccfdf-b9g84" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.891423 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf60578b-47bb-474a-b296-56c42105680d-combined-ca-bundle\") pod \"heat-api-65bb7ccfdf-b9g84\" (UID: \"cf60578b-47bb-474a-b296-56c42105680d\") " pod="openstack/heat-api-65bb7ccfdf-b9g84" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.900686 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltv2z\" (UniqueName: \"kubernetes.io/projected/cf60578b-47bb-474a-b296-56c42105680d-kube-api-access-ltv2z\") pod \"heat-api-65bb7ccfdf-b9g84\" (UID: \"cf60578b-47bb-474a-b296-56c42105680d\") " pod="openstack/heat-api-65bb7ccfdf-b9g84" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.932968 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-54f49697fd-pwnb8" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.982058 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f69f8d5-e82e-4db6-aad0-e302e807b2b0-config-data\") pod \"heat-cfnapi-5774948dd5-879rv\" (UID: \"5f69f8d5-e82e-4db6-aad0-e302e807b2b0\") " pod="openstack/heat-cfnapi-5774948dd5-879rv" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.982179 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f69f8d5-e82e-4db6-aad0-e302e807b2b0-config-data-custom\") pod \"heat-cfnapi-5774948dd5-879rv\" (UID: \"5f69f8d5-e82e-4db6-aad0-e302e807b2b0\") " pod="openstack/heat-cfnapi-5774948dd5-879rv" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.982258 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc75t\" (UniqueName: \"kubernetes.io/projected/5f69f8d5-e82e-4db6-aad0-e302e807b2b0-kube-api-access-pc75t\") pod \"heat-cfnapi-5774948dd5-879rv\" (UID: \"5f69f8d5-e82e-4db6-aad0-e302e807b2b0\") " pod="openstack/heat-cfnapi-5774948dd5-879rv" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.982283 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f69f8d5-e82e-4db6-aad0-e302e807b2b0-combined-ca-bundle\") pod \"heat-cfnapi-5774948dd5-879rv\" (UID: \"5f69f8d5-e82e-4db6-aad0-e302e807b2b0\") " pod="openstack/heat-cfnapi-5774948dd5-879rv" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.987748 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f69f8d5-e82e-4db6-aad0-e302e807b2b0-config-data\") pod \"heat-cfnapi-5774948dd5-879rv\" (UID: \"5f69f8d5-e82e-4db6-aad0-e302e807b2b0\") " pod="openstack/heat-cfnapi-5774948dd5-879rv" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.989812 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f69f8d5-e82e-4db6-aad0-e302e807b2b0-combined-ca-bundle\") pod \"heat-cfnapi-5774948dd5-879rv\" (UID: \"5f69f8d5-e82e-4db6-aad0-e302e807b2b0\") " pod="openstack/heat-cfnapi-5774948dd5-879rv" Dec 05 09:03:27 crc kubenswrapper[4997]: I1205 09:03:27.991766 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f69f8d5-e82e-4db6-aad0-e302e807b2b0-config-data-custom\") pod \"heat-cfnapi-5774948dd5-879rv\" (UID: \"5f69f8d5-e82e-4db6-aad0-e302e807b2b0\") " pod="openstack/heat-cfnapi-5774948dd5-879rv" Dec 05 09:03:28 crc kubenswrapper[4997]: I1205 09:03:28.031706 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc75t\" (UniqueName: \"kubernetes.io/projected/5f69f8d5-e82e-4db6-aad0-e302e807b2b0-kube-api-access-pc75t\") pod \"heat-cfnapi-5774948dd5-879rv\" (UID: \"5f69f8d5-e82e-4db6-aad0-e302e807b2b0\") " pod="openstack/heat-cfnapi-5774948dd5-879rv" Dec 05 09:03:28 crc kubenswrapper[4997]: I1205 09:03:28.052454 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-65bb7ccfdf-b9g84" Dec 05 09:03:28 crc kubenswrapper[4997]: I1205 09:03:28.074236 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5774948dd5-879rv" Dec 05 09:03:28 crc kubenswrapper[4997]: I1205 09:03:28.414414 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:28 crc kubenswrapper[4997]: I1205 09:03:28.474611 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-54f49697fd-pwnb8"] Dec 05 09:03:28 crc kubenswrapper[4997]: I1205 09:03:28.487574 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-65bb7ccfdf-b9g84"] Dec 05 09:03:28 crc kubenswrapper[4997]: W1205 09:03:28.583498 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f69f8d5_e82e_4db6_aad0_e302e807b2b0.slice/crio-be3211bc9c601dc645b641b313035cc57b28f0e3771d5135783afd6a2a4c2453 WatchSource:0}: Error finding container be3211bc9c601dc645b641b313035cc57b28f0e3771d5135783afd6a2a4c2453: Status 404 returned error can't find the container with id be3211bc9c601dc645b641b313035cc57b28f0e3771d5135783afd6a2a4c2453 Dec 05 09:03:28 crc kubenswrapper[4997]: I1205 09:03:28.583807 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5774948dd5-879rv"] Dec 05 09:03:29 crc kubenswrapper[4997]: I1205 09:03:29.470538 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5774948dd5-879rv" event={"ID":"5f69f8d5-e82e-4db6-aad0-e302e807b2b0","Type":"ContainerStarted","Data":"be3211bc9c601dc645b641b313035cc57b28f0e3771d5135783afd6a2a4c2453"} Dec 05 09:03:29 crc kubenswrapper[4997]: I1205 09:03:29.472696 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-65bb7ccfdf-b9g84" event={"ID":"cf60578b-47bb-474a-b296-56c42105680d","Type":"ContainerStarted","Data":"f2f4370c6b4ad68e215edb3f44877553cb14e5cced7f17c0fa4e314ef7767e01"} Dec 05 09:03:29 crc kubenswrapper[4997]: I1205 09:03:29.473899 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-54f49697fd-pwnb8" event={"ID":"fcb82741-d0c0-42af-ad5b-ac6a840393b8","Type":"ContainerStarted","Data":"9357e48542fcd8be12e411d801735ffaa068df2431b1e950eda3603c6889da6d"} Dec 05 09:03:29 crc kubenswrapper[4997]: I1205 09:03:29.473927 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-54f49697fd-pwnb8" event={"ID":"fcb82741-d0c0-42af-ad5b-ac6a840393b8","Type":"ContainerStarted","Data":"60908f1e661f75513d820cd4188b4029a793259781ff32f6a33ec96ddad8a65f"} Dec 05 09:03:29 crc kubenswrapper[4997]: I1205 09:03:29.475082 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-54f49697fd-pwnb8" Dec 05 09:03:29 crc kubenswrapper[4997]: I1205 09:03:29.500328 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-54f49697fd-pwnb8" podStartSLOduration=2.5003079489999998 podStartE2EDuration="2.500307949s" podCreationTimestamp="2025-12-05 09:03:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:03:29.492757966 +0000 UTC m=+7710.021665227" watchObservedRunningTime="2025-12-05 09:03:29.500307949 +0000 UTC m=+7710.029215210" Dec 05 09:03:30 crc kubenswrapper[4997]: I1205 09:03:30.061453 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-c64a-account-create-update-whrd6"] Dec 05 09:03:30 crc kubenswrapper[4997]: I1205 09:03:30.075774 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-vxnqk"] Dec 05 09:03:30 crc kubenswrapper[4997]: I1205 09:03:30.085629 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-c64a-account-create-update-whrd6"] Dec 05 09:03:30 crc kubenswrapper[4997]: I1205 09:03:30.095046 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-vxnqk"] Dec 05 09:03:30 crc kubenswrapper[4997]: I1205 09:03:30.139662 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-678ddff49c-bvhj8" Dec 05 09:03:30 crc kubenswrapper[4997]: I1205 09:03:30.194937 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79f85b4d6f-sj7dx"] Dec 05 09:03:30 crc kubenswrapper[4997]: I1205 09:03:30.195186 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79f85b4d6f-sj7dx" podUID="f2bfb275-8e76-4add-9610-23777e579acb" containerName="horizon-log" containerID="cri-o://4cab7fc650e874a372fb16bdf1f14882c2fa2c3c26fcd6c4da16d0098e44360f" gracePeriod=30 Dec 05 09:03:30 crc kubenswrapper[4997]: I1205 09:03:30.195636 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79f85b4d6f-sj7dx" podUID="f2bfb275-8e76-4add-9610-23777e579acb" containerName="horizon" containerID="cri-o://f3e3db4d78eefda48aab2857f9352c5a3c9b62108ad86e29c998a741e7145762" gracePeriod=30 Dec 05 09:03:31 crc kubenswrapper[4997]: I1205 09:03:31.493378 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-65bb7ccfdf-b9g84" event={"ID":"cf60578b-47bb-474a-b296-56c42105680d","Type":"ContainerStarted","Data":"5703a30d2e0613d364a43a853badc23c069121df1df4fe2bcdc7c4b664eb558a"} Dec 05 09:03:31 crc kubenswrapper[4997]: I1205 09:03:31.494335 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-65bb7ccfdf-b9g84" Dec 05 09:03:31 crc kubenswrapper[4997]: I1205 09:03:31.496046 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5774948dd5-879rv" event={"ID":"5f69f8d5-e82e-4db6-aad0-e302e807b2b0","Type":"ContainerStarted","Data":"f26050a1cea64c4cc2bd4a15cf33531ac0f19b0f361137414c5b86b1b3f9296d"} Dec 05 09:03:31 crc kubenswrapper[4997]: I1205 09:03:31.524335 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-65bb7ccfdf-b9g84" podStartSLOduration=2.51401204 podStartE2EDuration="4.52430519s" podCreationTimestamp="2025-12-05 09:03:27 +0000 UTC" firstStartedPulling="2025-12-05 09:03:28.47699479 +0000 UTC m=+7709.005902051" lastFinishedPulling="2025-12-05 09:03:30.48728794 +0000 UTC m=+7711.016195201" observedRunningTime="2025-12-05 09:03:31.520162108 +0000 UTC m=+7712.049069369" watchObservedRunningTime="2025-12-05 09:03:31.52430519 +0000 UTC m=+7712.053212451" Dec 05 09:03:31 crc kubenswrapper[4997]: I1205 09:03:31.538139 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5774948dd5-879rv" podStartSLOduration=2.633078151 podStartE2EDuration="4.538119172s" podCreationTimestamp="2025-12-05 09:03:27 +0000 UTC" firstStartedPulling="2025-12-05 09:03:28.586086202 +0000 UTC m=+7709.114993463" lastFinishedPulling="2025-12-05 09:03:30.491127223 +0000 UTC m=+7711.020034484" observedRunningTime="2025-12-05 09:03:31.536108707 +0000 UTC m=+7712.065015998" watchObservedRunningTime="2025-12-05 09:03:31.538119172 +0000 UTC m=+7712.067026443" Dec 05 09:03:31 crc kubenswrapper[4997]: I1205 09:03:31.760837 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fa78f95-7c44-48ec-84e5-e34e33ccc0c3" path="/var/lib/kubelet/pods/0fa78f95-7c44-48ec-84e5-e34e33ccc0c3/volumes" Dec 05 09:03:31 crc kubenswrapper[4997]: I1205 09:03:31.761697 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd" path="/var/lib/kubelet/pods/da4fc8f7-5fe3-4f0d-a3ba-4223acc491fd/volumes" Dec 05 09:03:32 crc kubenswrapper[4997]: I1205 09:03:32.504486 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5774948dd5-879rv" Dec 05 09:03:33 crc kubenswrapper[4997]: I1205 09:03:33.515102 4997 generic.go:334] "Generic (PLEG): container finished" podID="f2bfb275-8e76-4add-9610-23777e579acb" containerID="f3e3db4d78eefda48aab2857f9352c5a3c9b62108ad86e29c998a741e7145762" exitCode=0 Dec 05 09:03:33 crc kubenswrapper[4997]: I1205 09:03:33.515192 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f85b4d6f-sj7dx" event={"ID":"f2bfb275-8e76-4add-9610-23777e579acb","Type":"ContainerDied","Data":"f3e3db4d78eefda48aab2857f9352c5a3c9b62108ad86e29c998a741e7145762"} Dec 05 09:03:37 crc kubenswrapper[4997]: I1205 09:03:37.749222 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:03:37 crc kubenswrapper[4997]: E1205 09:03:37.750011 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:03:39 crc kubenswrapper[4997]: I1205 09:03:39.405601 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-65bb7ccfdf-b9g84" Dec 05 09:03:39 crc kubenswrapper[4997]: I1205 09:03:39.453440 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-5774948dd5-879rv" Dec 05 09:03:39 crc kubenswrapper[4997]: I1205 09:03:39.699772 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79f85b4d6f-sj7dx" podUID="f2bfb275-8e76-4add-9610-23777e579acb" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.111:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.111:8080: connect: connection refused" Dec 05 09:03:44 crc kubenswrapper[4997]: I1205 09:03:44.038366 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-fmhcf"] Dec 05 09:03:44 crc kubenswrapper[4997]: I1205 09:03:44.050207 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-fmhcf"] Dec 05 09:03:45 crc kubenswrapper[4997]: I1205 09:03:45.762763 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="797fa92f-8010-4e29-9a0d-9c3045136c25" path="/var/lib/kubelet/pods/797fa92f-8010-4e29-9a0d-9c3045136c25/volumes" Dec 05 09:03:47 crc kubenswrapper[4997]: I1205 09:03:47.963002 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-54f49697fd-pwnb8" Dec 05 09:03:49 crc kubenswrapper[4997]: I1205 09:03:49.700354 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79f85b4d6f-sj7dx" podUID="f2bfb275-8e76-4add-9610-23777e579acb" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.111:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.111:8080: connect: connection refused" Dec 05 09:03:52 crc kubenswrapper[4997]: I1205 09:03:52.749683 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:03:53 crc kubenswrapper[4997]: I1205 09:03:53.705787 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"a4c28f2d7c6c0a58e95d5724c2f234da68b209b35a39f917c1bd66336a29eaeb"} Dec 05 09:03:59 crc kubenswrapper[4997]: I1205 09:03:59.699968 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79f85b4d6f-sj7dx" podUID="f2bfb275-8e76-4add-9610-23777e579acb" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.111:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.111:8080: connect: connection refused" Dec 05 09:03:59 crc kubenswrapper[4997]: I1205 09:03:59.700512 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.635441 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.774465 4997 generic.go:334] "Generic (PLEG): container finished" podID="f2bfb275-8e76-4add-9610-23777e579acb" containerID="4cab7fc650e874a372fb16bdf1f14882c2fa2c3c26fcd6c4da16d0098e44360f" exitCode=137 Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.774510 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f85b4d6f-sj7dx" event={"ID":"f2bfb275-8e76-4add-9610-23777e579acb","Type":"ContainerDied","Data":"4cab7fc650e874a372fb16bdf1f14882c2fa2c3c26fcd6c4da16d0098e44360f"} Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.774516 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79f85b4d6f-sj7dx" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.774537 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f85b4d6f-sj7dx" event={"ID":"f2bfb275-8e76-4add-9610-23777e579acb","Type":"ContainerDied","Data":"1ae6c4d10b89a7895d3c0b4ef62ecbf631c7e4246086e0079ca0c844b31eede0"} Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.774566 4997 scope.go:117] "RemoveContainer" containerID="f3e3db4d78eefda48aab2857f9352c5a3c9b62108ad86e29c998a741e7145762" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.821932 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f2bfb275-8e76-4add-9610-23777e579acb-horizon-secret-key\") pod \"f2bfb275-8e76-4add-9610-23777e579acb\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.822085 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2bfb275-8e76-4add-9610-23777e579acb-logs\") pod \"f2bfb275-8e76-4add-9610-23777e579acb\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.822513 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2bfb275-8e76-4add-9610-23777e579acb-logs" (OuterVolumeSpecName: "logs") pod "f2bfb275-8e76-4add-9610-23777e579acb" (UID: "f2bfb275-8e76-4add-9610-23777e579acb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.822774 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2bfb275-8e76-4add-9610-23777e579acb-scripts\") pod \"f2bfb275-8e76-4add-9610-23777e579acb\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.822967 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbxjb\" (UniqueName: \"kubernetes.io/projected/f2bfb275-8e76-4add-9610-23777e579acb-kube-api-access-qbxjb\") pod \"f2bfb275-8e76-4add-9610-23777e579acb\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.823037 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f2bfb275-8e76-4add-9610-23777e579acb-config-data\") pod \"f2bfb275-8e76-4add-9610-23777e579acb\" (UID: \"f2bfb275-8e76-4add-9610-23777e579acb\") " Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.824736 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2bfb275-8e76-4add-9610-23777e579acb-logs\") on node \"crc\" DevicePath \"\"" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.834678 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2bfb275-8e76-4add-9610-23777e579acb-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f2bfb275-8e76-4add-9610-23777e579acb" (UID: "f2bfb275-8e76-4add-9610-23777e579acb"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.834709 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2bfb275-8e76-4add-9610-23777e579acb-kube-api-access-qbxjb" (OuterVolumeSpecName: "kube-api-access-qbxjb") pod "f2bfb275-8e76-4add-9610-23777e579acb" (UID: "f2bfb275-8e76-4add-9610-23777e579acb"). InnerVolumeSpecName "kube-api-access-qbxjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.848095 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2bfb275-8e76-4add-9610-23777e579acb-scripts" (OuterVolumeSpecName: "scripts") pod "f2bfb275-8e76-4add-9610-23777e579acb" (UID: "f2bfb275-8e76-4add-9610-23777e579acb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.849688 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2bfb275-8e76-4add-9610-23777e579acb-config-data" (OuterVolumeSpecName: "config-data") pod "f2bfb275-8e76-4add-9610-23777e579acb" (UID: "f2bfb275-8e76-4add-9610-23777e579acb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.927667 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2bfb275-8e76-4add-9610-23777e579acb-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.928772 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbxjb\" (UniqueName: \"kubernetes.io/projected/f2bfb275-8e76-4add-9610-23777e579acb-kube-api-access-qbxjb\") on node \"crc\" DevicePath \"\"" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.928801 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f2bfb275-8e76-4add-9610-23777e579acb-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.928814 4997 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f2bfb275-8e76-4add-9610-23777e579acb-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.933936 4997 scope.go:117] "RemoveContainer" containerID="4cab7fc650e874a372fb16bdf1f14882c2fa2c3c26fcd6c4da16d0098e44360f" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.958374 4997 scope.go:117] "RemoveContainer" containerID="f3e3db4d78eefda48aab2857f9352c5a3c9b62108ad86e29c998a741e7145762" Dec 05 09:04:00 crc kubenswrapper[4997]: E1205 09:04:00.958893 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3e3db4d78eefda48aab2857f9352c5a3c9b62108ad86e29c998a741e7145762\": container with ID starting with f3e3db4d78eefda48aab2857f9352c5a3c9b62108ad86e29c998a741e7145762 not found: ID does not exist" containerID="f3e3db4d78eefda48aab2857f9352c5a3c9b62108ad86e29c998a741e7145762" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.958934 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3e3db4d78eefda48aab2857f9352c5a3c9b62108ad86e29c998a741e7145762"} err="failed to get container status \"f3e3db4d78eefda48aab2857f9352c5a3c9b62108ad86e29c998a741e7145762\": rpc error: code = NotFound desc = could not find container \"f3e3db4d78eefda48aab2857f9352c5a3c9b62108ad86e29c998a741e7145762\": container with ID starting with f3e3db4d78eefda48aab2857f9352c5a3c9b62108ad86e29c998a741e7145762 not found: ID does not exist" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.958962 4997 scope.go:117] "RemoveContainer" containerID="4cab7fc650e874a372fb16bdf1f14882c2fa2c3c26fcd6c4da16d0098e44360f" Dec 05 09:04:00 crc kubenswrapper[4997]: E1205 09:04:00.959233 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cab7fc650e874a372fb16bdf1f14882c2fa2c3c26fcd6c4da16d0098e44360f\": container with ID starting with 4cab7fc650e874a372fb16bdf1f14882c2fa2c3c26fcd6c4da16d0098e44360f not found: ID does not exist" containerID="4cab7fc650e874a372fb16bdf1f14882c2fa2c3c26fcd6c4da16d0098e44360f" Dec 05 09:04:00 crc kubenswrapper[4997]: I1205 09:04:00.959263 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cab7fc650e874a372fb16bdf1f14882c2fa2c3c26fcd6c4da16d0098e44360f"} err="failed to get container status \"4cab7fc650e874a372fb16bdf1f14882c2fa2c3c26fcd6c4da16d0098e44360f\": rpc error: code = NotFound desc = could not find container \"4cab7fc650e874a372fb16bdf1f14882c2fa2c3c26fcd6c4da16d0098e44360f\": container with ID starting with 4cab7fc650e874a372fb16bdf1f14882c2fa2c3c26fcd6c4da16d0098e44360f not found: ID does not exist" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.108643 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79f85b4d6f-sj7dx"] Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.116898 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-79f85b4d6f-sj7dx"] Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.302156 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm"] Dec 05 09:04:01 crc kubenswrapper[4997]: E1205 09:04:01.302581 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2bfb275-8e76-4add-9610-23777e579acb" containerName="horizon-log" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.302624 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2bfb275-8e76-4add-9610-23777e579acb" containerName="horizon-log" Dec 05 09:04:01 crc kubenswrapper[4997]: E1205 09:04:01.302674 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2bfb275-8e76-4add-9610-23777e579acb" containerName="horizon" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.302684 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2bfb275-8e76-4add-9610-23777e579acb" containerName="horizon" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.302904 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2bfb275-8e76-4add-9610-23777e579acb" containerName="horizon-log" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.302926 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2bfb275-8e76-4add-9610-23777e579acb" containerName="horizon" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.304602 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.311140 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.313098 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm"] Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.333588 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2414755f-2b83-4d3e-a9d2-16147925fc44-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm\" (UID: \"2414755f-2b83-4d3e-a9d2-16147925fc44\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.333684 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2414755f-2b83-4d3e-a9d2-16147925fc44-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm\" (UID: \"2414755f-2b83-4d3e-a9d2-16147925fc44\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.333779 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bxnh\" (UniqueName: \"kubernetes.io/projected/2414755f-2b83-4d3e-a9d2-16147925fc44-kube-api-access-2bxnh\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm\" (UID: \"2414755f-2b83-4d3e-a9d2-16147925fc44\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.435011 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bxnh\" (UniqueName: \"kubernetes.io/projected/2414755f-2b83-4d3e-a9d2-16147925fc44-kube-api-access-2bxnh\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm\" (UID: \"2414755f-2b83-4d3e-a9d2-16147925fc44\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.435134 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2414755f-2b83-4d3e-a9d2-16147925fc44-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm\" (UID: \"2414755f-2b83-4d3e-a9d2-16147925fc44\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.435178 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2414755f-2b83-4d3e-a9d2-16147925fc44-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm\" (UID: \"2414755f-2b83-4d3e-a9d2-16147925fc44\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.436474 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2414755f-2b83-4d3e-a9d2-16147925fc44-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm\" (UID: \"2414755f-2b83-4d3e-a9d2-16147925fc44\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.436478 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2414755f-2b83-4d3e-a9d2-16147925fc44-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm\" (UID: \"2414755f-2b83-4d3e-a9d2-16147925fc44\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.452639 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bxnh\" (UniqueName: \"kubernetes.io/projected/2414755f-2b83-4d3e-a9d2-16147925fc44-kube-api-access-2bxnh\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm\" (UID: \"2414755f-2b83-4d3e-a9d2-16147925fc44\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.628983 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" Dec 05 09:04:01 crc kubenswrapper[4997]: I1205 09:04:01.768780 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2bfb275-8e76-4add-9610-23777e579acb" path="/var/lib/kubelet/pods/f2bfb275-8e76-4add-9610-23777e579acb/volumes" Dec 05 09:04:02 crc kubenswrapper[4997]: W1205 09:04:02.112899 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2414755f_2b83_4d3e_a9d2_16147925fc44.slice/crio-eebe8434fddde3d87194867cca02594f3a1b8d39a2a737df933b0678631e72a7 WatchSource:0}: Error finding container eebe8434fddde3d87194867cca02594f3a1b8d39a2a737df933b0678631e72a7: Status 404 returned error can't find the container with id eebe8434fddde3d87194867cca02594f3a1b8d39a2a737df933b0678631e72a7 Dec 05 09:04:02 crc kubenswrapper[4997]: I1205 09:04:02.114518 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm"] Dec 05 09:04:02 crc kubenswrapper[4997]: I1205 09:04:02.794110 4997 generic.go:334] "Generic (PLEG): container finished" podID="2414755f-2b83-4d3e-a9d2-16147925fc44" containerID="6f0b72fe8a27f5b8ca169ae69c9d857a186403db9f6bee31241d362f181e0e13" exitCode=0 Dec 05 09:04:02 crc kubenswrapper[4997]: I1205 09:04:02.794147 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" event={"ID":"2414755f-2b83-4d3e-a9d2-16147925fc44","Type":"ContainerDied","Data":"6f0b72fe8a27f5b8ca169ae69c9d857a186403db9f6bee31241d362f181e0e13"} Dec 05 09:04:02 crc kubenswrapper[4997]: I1205 09:04:02.794170 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" event={"ID":"2414755f-2b83-4d3e-a9d2-16147925fc44","Type":"ContainerStarted","Data":"eebe8434fddde3d87194867cca02594f3a1b8d39a2a737df933b0678631e72a7"} Dec 05 09:04:02 crc kubenswrapper[4997]: I1205 09:04:02.797522 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:04:02 crc kubenswrapper[4997]: I1205 09:04:02.977428 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j69xh"] Dec 05 09:04:02 crc kubenswrapper[4997]: I1205 09:04:02.979640 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j69xh" Dec 05 09:04:02 crc kubenswrapper[4997]: I1205 09:04:02.991270 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j69xh"] Dec 05 09:04:03 crc kubenswrapper[4997]: I1205 09:04:03.168853 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n6zb\" (UniqueName: \"kubernetes.io/projected/a5de22f9-fddd-41fd-8a73-f77153be79eb-kube-api-access-7n6zb\") pod \"redhat-operators-j69xh\" (UID: \"a5de22f9-fddd-41fd-8a73-f77153be79eb\") " pod="openshift-marketplace/redhat-operators-j69xh" Dec 05 09:04:03 crc kubenswrapper[4997]: I1205 09:04:03.168947 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5de22f9-fddd-41fd-8a73-f77153be79eb-catalog-content\") pod \"redhat-operators-j69xh\" (UID: \"a5de22f9-fddd-41fd-8a73-f77153be79eb\") " pod="openshift-marketplace/redhat-operators-j69xh" Dec 05 09:04:03 crc kubenswrapper[4997]: I1205 09:04:03.169000 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5de22f9-fddd-41fd-8a73-f77153be79eb-utilities\") pod \"redhat-operators-j69xh\" (UID: \"a5de22f9-fddd-41fd-8a73-f77153be79eb\") " pod="openshift-marketplace/redhat-operators-j69xh" Dec 05 09:04:03 crc kubenswrapper[4997]: I1205 09:04:03.270893 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5de22f9-fddd-41fd-8a73-f77153be79eb-utilities\") pod \"redhat-operators-j69xh\" (UID: \"a5de22f9-fddd-41fd-8a73-f77153be79eb\") " pod="openshift-marketplace/redhat-operators-j69xh" Dec 05 09:04:03 crc kubenswrapper[4997]: I1205 09:04:03.271079 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n6zb\" (UniqueName: \"kubernetes.io/projected/a5de22f9-fddd-41fd-8a73-f77153be79eb-kube-api-access-7n6zb\") pod \"redhat-operators-j69xh\" (UID: \"a5de22f9-fddd-41fd-8a73-f77153be79eb\") " pod="openshift-marketplace/redhat-operators-j69xh" Dec 05 09:04:03 crc kubenswrapper[4997]: I1205 09:04:03.271111 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5de22f9-fddd-41fd-8a73-f77153be79eb-catalog-content\") pod \"redhat-operators-j69xh\" (UID: \"a5de22f9-fddd-41fd-8a73-f77153be79eb\") " pod="openshift-marketplace/redhat-operators-j69xh" Dec 05 09:04:03 crc kubenswrapper[4997]: I1205 09:04:03.271566 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5de22f9-fddd-41fd-8a73-f77153be79eb-utilities\") pod \"redhat-operators-j69xh\" (UID: \"a5de22f9-fddd-41fd-8a73-f77153be79eb\") " pod="openshift-marketplace/redhat-operators-j69xh" Dec 05 09:04:03 crc kubenswrapper[4997]: I1205 09:04:03.271596 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5de22f9-fddd-41fd-8a73-f77153be79eb-catalog-content\") pod \"redhat-operators-j69xh\" (UID: \"a5de22f9-fddd-41fd-8a73-f77153be79eb\") " pod="openshift-marketplace/redhat-operators-j69xh" Dec 05 09:04:03 crc kubenswrapper[4997]: I1205 09:04:03.291421 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n6zb\" (UniqueName: \"kubernetes.io/projected/a5de22f9-fddd-41fd-8a73-f77153be79eb-kube-api-access-7n6zb\") pod \"redhat-operators-j69xh\" (UID: \"a5de22f9-fddd-41fd-8a73-f77153be79eb\") " pod="openshift-marketplace/redhat-operators-j69xh" Dec 05 09:04:03 crc kubenswrapper[4997]: I1205 09:04:03.304979 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j69xh" Dec 05 09:04:04 crc kubenswrapper[4997]: I1205 09:04:03.794171 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j69xh"] Dec 05 09:04:04 crc kubenswrapper[4997]: E1205 09:04:04.282380 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5de22f9_fddd_41fd_8a73_f77153be79eb.slice/crio-c096f4d786ae02310db2396c1e73b98f92bca4d3a957fa8391fa14175e1da83a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5de22f9_fddd_41fd_8a73_f77153be79eb.slice/crio-conmon-c096f4d786ae02310db2396c1e73b98f92bca4d3a957fa8391fa14175e1da83a.scope\": RecentStats: unable to find data in memory cache]" Dec 05 09:04:04 crc kubenswrapper[4997]: I1205 09:04:04.813522 4997 generic.go:334] "Generic (PLEG): container finished" podID="a5de22f9-fddd-41fd-8a73-f77153be79eb" containerID="c096f4d786ae02310db2396c1e73b98f92bca4d3a957fa8391fa14175e1da83a" exitCode=0 Dec 05 09:04:04 crc kubenswrapper[4997]: I1205 09:04:04.813625 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69xh" event={"ID":"a5de22f9-fddd-41fd-8a73-f77153be79eb","Type":"ContainerDied","Data":"c096f4d786ae02310db2396c1e73b98f92bca4d3a957fa8391fa14175e1da83a"} Dec 05 09:04:04 crc kubenswrapper[4997]: I1205 09:04:04.813913 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69xh" event={"ID":"a5de22f9-fddd-41fd-8a73-f77153be79eb","Type":"ContainerStarted","Data":"de1272ac7797c6252eaa586abab865ef52a82b2aa5695724a632d9786300242b"} Dec 05 09:04:04 crc kubenswrapper[4997]: I1205 09:04:04.816547 4997 generic.go:334] "Generic (PLEG): container finished" podID="2414755f-2b83-4d3e-a9d2-16147925fc44" containerID="06cc4e5cb46d2d10f6e0858850a544a98ee8bdb85bffc824515cdec8bc0e438d" exitCode=0 Dec 05 09:04:04 crc kubenswrapper[4997]: I1205 09:04:04.816592 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" event={"ID":"2414755f-2b83-4d3e-a9d2-16147925fc44","Type":"ContainerDied","Data":"06cc4e5cb46d2d10f6e0858850a544a98ee8bdb85bffc824515cdec8bc0e438d"} Dec 05 09:04:05 crc kubenswrapper[4997]: I1205 09:04:05.832984 4997 generic.go:334] "Generic (PLEG): container finished" podID="2414755f-2b83-4d3e-a9d2-16147925fc44" containerID="365fb8d5efc7e649b5e5df99007e21e25cbe1a0a50a16b3822fd5af113bb64bf" exitCode=0 Dec 05 09:04:05 crc kubenswrapper[4997]: I1205 09:04:05.833074 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" event={"ID":"2414755f-2b83-4d3e-a9d2-16147925fc44","Type":"ContainerDied","Data":"365fb8d5efc7e649b5e5df99007e21e25cbe1a0a50a16b3822fd5af113bb64bf"} Dec 05 09:04:06 crc kubenswrapper[4997]: I1205 09:04:06.852771 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69xh" event={"ID":"a5de22f9-fddd-41fd-8a73-f77153be79eb","Type":"ContainerStarted","Data":"0d01833a3d3a275f57ea4c10145a9e9a480cabfe0a2eee0d5d2cc06dc86138bf"} Dec 05 09:04:07 crc kubenswrapper[4997]: I1205 09:04:07.031826 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-q75gz"] Dec 05 09:04:07 crc kubenswrapper[4997]: I1205 09:04:07.054435 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-q75gz"] Dec 05 09:04:07 crc kubenswrapper[4997]: I1205 09:04:07.073729 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-67c3-account-create-update-2mshm"] Dec 05 09:04:07 crc kubenswrapper[4997]: I1205 09:04:07.084113 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-67c3-account-create-update-2mshm"] Dec 05 09:04:07 crc kubenswrapper[4997]: I1205 09:04:07.221776 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" Dec 05 09:04:07 crc kubenswrapper[4997]: I1205 09:04:07.371667 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2414755f-2b83-4d3e-a9d2-16147925fc44-util\") pod \"2414755f-2b83-4d3e-a9d2-16147925fc44\" (UID: \"2414755f-2b83-4d3e-a9d2-16147925fc44\") " Dec 05 09:04:07 crc kubenswrapper[4997]: I1205 09:04:07.371857 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2414755f-2b83-4d3e-a9d2-16147925fc44-bundle\") pod \"2414755f-2b83-4d3e-a9d2-16147925fc44\" (UID: \"2414755f-2b83-4d3e-a9d2-16147925fc44\") " Dec 05 09:04:07 crc kubenswrapper[4997]: I1205 09:04:07.372045 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bxnh\" (UniqueName: \"kubernetes.io/projected/2414755f-2b83-4d3e-a9d2-16147925fc44-kube-api-access-2bxnh\") pod \"2414755f-2b83-4d3e-a9d2-16147925fc44\" (UID: \"2414755f-2b83-4d3e-a9d2-16147925fc44\") " Dec 05 09:04:07 crc kubenswrapper[4997]: I1205 09:04:07.376534 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2414755f-2b83-4d3e-a9d2-16147925fc44-bundle" (OuterVolumeSpecName: "bundle") pod "2414755f-2b83-4d3e-a9d2-16147925fc44" (UID: "2414755f-2b83-4d3e-a9d2-16147925fc44"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:04:07 crc kubenswrapper[4997]: I1205 09:04:07.381291 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2414755f-2b83-4d3e-a9d2-16147925fc44-kube-api-access-2bxnh" (OuterVolumeSpecName: "kube-api-access-2bxnh") pod "2414755f-2b83-4d3e-a9d2-16147925fc44" (UID: "2414755f-2b83-4d3e-a9d2-16147925fc44"). InnerVolumeSpecName "kube-api-access-2bxnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:04:07 crc kubenswrapper[4997]: I1205 09:04:07.479140 4997 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2414755f-2b83-4d3e-a9d2-16147925fc44-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:04:07 crc kubenswrapper[4997]: I1205 09:04:07.479189 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bxnh\" (UniqueName: \"kubernetes.io/projected/2414755f-2b83-4d3e-a9d2-16147925fc44-kube-api-access-2bxnh\") on node \"crc\" DevicePath \"\"" Dec 05 09:04:07 crc kubenswrapper[4997]: I1205 09:04:07.760058 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="538190ea-f66a-408c-9623-ced05985673f" path="/var/lib/kubelet/pods/538190ea-f66a-408c-9623-ced05985673f/volumes" Dec 05 09:04:07 crc kubenswrapper[4997]: I1205 09:04:07.760905 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bde1371d-4c9d-4e91-99a5-dcd2bba4a22e" path="/var/lib/kubelet/pods/bde1371d-4c9d-4e91-99a5-dcd2bba4a22e/volumes" Dec 05 09:04:07 crc kubenswrapper[4997]: I1205 09:04:07.868520 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" Dec 05 09:04:07 crc kubenswrapper[4997]: I1205 09:04:07.868587 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm" event={"ID":"2414755f-2b83-4d3e-a9d2-16147925fc44","Type":"ContainerDied","Data":"eebe8434fddde3d87194867cca02594f3a1b8d39a2a737df933b0678631e72a7"} Dec 05 09:04:07 crc kubenswrapper[4997]: I1205 09:04:07.868721 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eebe8434fddde3d87194867cca02594f3a1b8d39a2a737df933b0678631e72a7" Dec 05 09:04:09 crc kubenswrapper[4997]: I1205 09:04:08.882457 4997 generic.go:334] "Generic (PLEG): container finished" podID="a5de22f9-fddd-41fd-8a73-f77153be79eb" containerID="0d01833a3d3a275f57ea4c10145a9e9a480cabfe0a2eee0d5d2cc06dc86138bf" exitCode=0 Dec 05 09:04:09 crc kubenswrapper[4997]: I1205 09:04:08.882540 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69xh" event={"ID":"a5de22f9-fddd-41fd-8a73-f77153be79eb","Type":"ContainerDied","Data":"0d01833a3d3a275f57ea4c10145a9e9a480cabfe0a2eee0d5d2cc06dc86138bf"} Dec 05 09:04:09 crc kubenswrapper[4997]: I1205 09:04:09.082771 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2414755f-2b83-4d3e-a9d2-16147925fc44-util" (OuterVolumeSpecName: "util") pod "2414755f-2b83-4d3e-a9d2-16147925fc44" (UID: "2414755f-2b83-4d3e-a9d2-16147925fc44"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:04:09 crc kubenswrapper[4997]: I1205 09:04:09.112678 4997 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2414755f-2b83-4d3e-a9d2-16147925fc44-util\") on node \"crc\" DevicePath \"\"" Dec 05 09:04:09 crc kubenswrapper[4997]: I1205 09:04:09.893888 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69xh" event={"ID":"a5de22f9-fddd-41fd-8a73-f77153be79eb","Type":"ContainerStarted","Data":"aaaa317a46da494061c3a7e6a05547e5a372eb429917c38b26e95dc62aabf8a1"} Dec 05 09:04:09 crc kubenswrapper[4997]: I1205 09:04:09.919235 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j69xh" podStartSLOduration=3.200860397 podStartE2EDuration="7.919209057s" podCreationTimestamp="2025-12-05 09:04:02 +0000 UTC" firstStartedPulling="2025-12-05 09:04:04.815240967 +0000 UTC m=+7745.344148228" lastFinishedPulling="2025-12-05 09:04:09.533589627 +0000 UTC m=+7750.062496888" observedRunningTime="2025-12-05 09:04:09.911653363 +0000 UTC m=+7750.440560644" watchObservedRunningTime="2025-12-05 09:04:09.919209057 +0000 UTC m=+7750.448116328" Dec 05 09:04:13 crc kubenswrapper[4997]: I1205 09:04:13.305993 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j69xh" Dec 05 09:04:13 crc kubenswrapper[4997]: I1205 09:04:13.308457 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j69xh" Dec 05 09:04:14 crc kubenswrapper[4997]: I1205 09:04:14.382648 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j69xh" podUID="a5de22f9-fddd-41fd-8a73-f77153be79eb" containerName="registry-server" probeResult="failure" output=< Dec 05 09:04:14 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 09:04:14 crc kubenswrapper[4997]: > Dec 05 09:04:15 crc kubenswrapper[4997]: I1205 09:04:15.106194 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-bcmls"] Dec 05 09:04:15 crc kubenswrapper[4997]: I1205 09:04:15.117323 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-bcmls"] Dec 05 09:04:15 crc kubenswrapper[4997]: I1205 09:04:15.762453 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f5512ca-098c-4d06-b455-3a76338df304" path="/var/lib/kubelet/pods/0f5512ca-098c-4d06-b455-3a76338df304/volumes" Dec 05 09:04:16 crc kubenswrapper[4997]: I1205 09:04:16.445500 4997 scope.go:117] "RemoveContainer" containerID="dca43fc19b1f57dc4e660c95619d128f26c8ff15bbab861870f742c4a77bcfa4" Dec 05 09:04:16 crc kubenswrapper[4997]: I1205 09:04:16.489215 4997 scope.go:117] "RemoveContainer" containerID="db5a163466ab65869c5e21d0cbe1357f5a2f20986d633203d918ed439454d9ce" Dec 05 09:04:16 crc kubenswrapper[4997]: I1205 09:04:16.529835 4997 scope.go:117] "RemoveContainer" containerID="d3a0b14c5b38af42c3a0f352d682859e87679bf2169ad20834a221d60b888743" Dec 05 09:04:16 crc kubenswrapper[4997]: I1205 09:04:16.610949 4997 scope.go:117] "RemoveContainer" containerID="07903e1eeb8d39db262958be291c36d6522a607de50d8c7b9e972a5b3eff1d1d" Dec 05 09:04:16 crc kubenswrapper[4997]: I1205 09:04:16.658547 4997 scope.go:117] "RemoveContainer" containerID="70df3d45cf80aa0a7179ad46c24d0545895732c981a5c6bbcb27a8ee215be892" Dec 05 09:04:16 crc kubenswrapper[4997]: I1205 09:04:16.730847 4997 scope.go:117] "RemoveContainer" containerID="58bd0e4ffef3300c91df54143cbfffb7d5044628d0d6372254962c416b7e6506" Dec 05 09:04:16 crc kubenswrapper[4997]: I1205 09:04:16.821603 4997 scope.go:117] "RemoveContainer" containerID="f76545c8d04a765c1ffec4d2255263ce7b7a54d45a8fca50f1cb6e5bee8214c0" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.172923 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-4586l"] Dec 05 09:04:20 crc kubenswrapper[4997]: E1205 09:04:20.173822 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2414755f-2b83-4d3e-a9d2-16147925fc44" containerName="extract" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.173836 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2414755f-2b83-4d3e-a9d2-16147925fc44" containerName="extract" Dec 05 09:04:20 crc kubenswrapper[4997]: E1205 09:04:20.173863 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2414755f-2b83-4d3e-a9d2-16147925fc44" containerName="util" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.173869 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2414755f-2b83-4d3e-a9d2-16147925fc44" containerName="util" Dec 05 09:04:20 crc kubenswrapper[4997]: E1205 09:04:20.173876 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2414755f-2b83-4d3e-a9d2-16147925fc44" containerName="pull" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.173882 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="2414755f-2b83-4d3e-a9d2-16147925fc44" containerName="pull" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.174083 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="2414755f-2b83-4d3e-a9d2-16147925fc44" containerName="extract" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.174831 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-4586l" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.177898 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.178110 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.182194 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-tb65k" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.191490 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-4586l"] Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.238809 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58lcp\" (UniqueName: \"kubernetes.io/projected/68e3de45-5599-48c3-8b35-73cc4d02250a-kube-api-access-58lcp\") pod \"obo-prometheus-operator-668cf9dfbb-4586l\" (UID: \"68e3de45-5599-48c3-8b35-73cc4d02250a\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-4586l" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.247343 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-qz4cf"] Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.248591 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-qz4cf" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.251525 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.251866 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-r782w" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.277081 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-qz4cf"] Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.290909 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-gtbr9"] Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.293242 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-gtbr9" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.344310 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/77d7110e-dc31-46fe-90d0-a1344dea977e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-58cc577445-gtbr9\" (UID: \"77d7110e-dc31-46fe-90d0-a1344dea977e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-gtbr9" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.344353 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/77d7110e-dc31-46fe-90d0-a1344dea977e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-58cc577445-gtbr9\" (UID: \"77d7110e-dc31-46fe-90d0-a1344dea977e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-gtbr9" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.344469 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58lcp\" (UniqueName: \"kubernetes.io/projected/68e3de45-5599-48c3-8b35-73cc4d02250a-kube-api-access-58lcp\") pod \"obo-prometheus-operator-668cf9dfbb-4586l\" (UID: \"68e3de45-5599-48c3-8b35-73cc4d02250a\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-4586l" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.344551 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c4adb5f1-1596-4833-92e8-8fddef0bdd40-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-58cc577445-qz4cf\" (UID: \"c4adb5f1-1596-4833-92e8-8fddef0bdd40\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-qz4cf" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.344604 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c4adb5f1-1596-4833-92e8-8fddef0bdd40-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-58cc577445-qz4cf\" (UID: \"c4adb5f1-1596-4833-92e8-8fddef0bdd40\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-qz4cf" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.363489 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-gtbr9"] Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.411850 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58lcp\" (UniqueName: \"kubernetes.io/projected/68e3de45-5599-48c3-8b35-73cc4d02250a-kube-api-access-58lcp\") pod \"obo-prometheus-operator-668cf9dfbb-4586l\" (UID: \"68e3de45-5599-48c3-8b35-73cc4d02250a\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-4586l" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.435398 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-xzrfl"] Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.437288 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-xzrfl" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.441092 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-49srj" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.441523 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.446053 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/77d7110e-dc31-46fe-90d0-a1344dea977e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-58cc577445-gtbr9\" (UID: \"77d7110e-dc31-46fe-90d0-a1344dea977e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-gtbr9" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.446100 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/77d7110e-dc31-46fe-90d0-a1344dea977e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-58cc577445-gtbr9\" (UID: \"77d7110e-dc31-46fe-90d0-a1344dea977e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-gtbr9" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.446199 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c4adb5f1-1596-4833-92e8-8fddef0bdd40-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-58cc577445-qz4cf\" (UID: \"c4adb5f1-1596-4833-92e8-8fddef0bdd40\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-qz4cf" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.446227 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c4adb5f1-1596-4833-92e8-8fddef0bdd40-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-58cc577445-qz4cf\" (UID: \"c4adb5f1-1596-4833-92e8-8fddef0bdd40\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-qz4cf" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.450813 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/77d7110e-dc31-46fe-90d0-a1344dea977e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-58cc577445-gtbr9\" (UID: \"77d7110e-dc31-46fe-90d0-a1344dea977e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-gtbr9" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.453396 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-xzrfl"] Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.454726 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c4adb5f1-1596-4833-92e8-8fddef0bdd40-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-58cc577445-qz4cf\" (UID: \"c4adb5f1-1596-4833-92e8-8fddef0bdd40\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-qz4cf" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.460106 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c4adb5f1-1596-4833-92e8-8fddef0bdd40-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-58cc577445-qz4cf\" (UID: \"c4adb5f1-1596-4833-92e8-8fddef0bdd40\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-qz4cf" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.460871 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/77d7110e-dc31-46fe-90d0-a1344dea977e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-58cc577445-gtbr9\" (UID: \"77d7110e-dc31-46fe-90d0-a1344dea977e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-gtbr9" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.501011 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-4586l" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.547936 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xld5j\" (UniqueName: \"kubernetes.io/projected/da0ed268-f4eb-4a9f-8ecf-bd7d318e0f09-kube-api-access-xld5j\") pod \"observability-operator-d8bb48f5d-xzrfl\" (UID: \"da0ed268-f4eb-4a9f-8ecf-bd7d318e0f09\") " pod="openshift-operators/observability-operator-d8bb48f5d-xzrfl" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.548045 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/da0ed268-f4eb-4a9f-8ecf-bd7d318e0f09-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-xzrfl\" (UID: \"da0ed268-f4eb-4a9f-8ecf-bd7d318e0f09\") " pod="openshift-operators/observability-operator-d8bb48f5d-xzrfl" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.572072 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-qz4cf" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.624758 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-rb78r"] Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.626729 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-rb78r" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.631568 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-9gm4g" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.647159 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-gtbr9" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.649754 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xld5j\" (UniqueName: \"kubernetes.io/projected/da0ed268-f4eb-4a9f-8ecf-bd7d318e0f09-kube-api-access-xld5j\") pod \"observability-operator-d8bb48f5d-xzrfl\" (UID: \"da0ed268-f4eb-4a9f-8ecf-bd7d318e0f09\") " pod="openshift-operators/observability-operator-d8bb48f5d-xzrfl" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.649823 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/da0ed268-f4eb-4a9f-8ecf-bd7d318e0f09-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-xzrfl\" (UID: \"da0ed268-f4eb-4a9f-8ecf-bd7d318e0f09\") " pod="openshift-operators/observability-operator-d8bb48f5d-xzrfl" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.661151 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/da0ed268-f4eb-4a9f-8ecf-bd7d318e0f09-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-xzrfl\" (UID: \"da0ed268-f4eb-4a9f-8ecf-bd7d318e0f09\") " pod="openshift-operators/observability-operator-d8bb48f5d-xzrfl" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.674969 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-rb78r"] Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.677553 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xld5j\" (UniqueName: \"kubernetes.io/projected/da0ed268-f4eb-4a9f-8ecf-bd7d318e0f09-kube-api-access-xld5j\") pod \"observability-operator-d8bb48f5d-xzrfl\" (UID: \"da0ed268-f4eb-4a9f-8ecf-bd7d318e0f09\") " pod="openshift-operators/observability-operator-d8bb48f5d-xzrfl" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.752845 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcmrm\" (UniqueName: \"kubernetes.io/projected/9f2abe7e-5864-476a-b5b8-f46b7d7d7350-kube-api-access-gcmrm\") pod \"perses-operator-5446b9c989-rb78r\" (UID: \"9f2abe7e-5864-476a-b5b8-f46b7d7d7350\") " pod="openshift-operators/perses-operator-5446b9c989-rb78r" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.752920 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/9f2abe7e-5864-476a-b5b8-f46b7d7d7350-openshift-service-ca\") pod \"perses-operator-5446b9c989-rb78r\" (UID: \"9f2abe7e-5864-476a-b5b8-f46b7d7d7350\") " pod="openshift-operators/perses-operator-5446b9c989-rb78r" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.856816 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcmrm\" (UniqueName: \"kubernetes.io/projected/9f2abe7e-5864-476a-b5b8-f46b7d7d7350-kube-api-access-gcmrm\") pod \"perses-operator-5446b9c989-rb78r\" (UID: \"9f2abe7e-5864-476a-b5b8-f46b7d7d7350\") " pod="openshift-operators/perses-operator-5446b9c989-rb78r" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.856885 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/9f2abe7e-5864-476a-b5b8-f46b7d7d7350-openshift-service-ca\") pod \"perses-operator-5446b9c989-rb78r\" (UID: \"9f2abe7e-5864-476a-b5b8-f46b7d7d7350\") " pod="openshift-operators/perses-operator-5446b9c989-rb78r" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.857810 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/9f2abe7e-5864-476a-b5b8-f46b7d7d7350-openshift-service-ca\") pod \"perses-operator-5446b9c989-rb78r\" (UID: \"9f2abe7e-5864-476a-b5b8-f46b7d7d7350\") " pod="openshift-operators/perses-operator-5446b9c989-rb78r" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.901869 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcmrm\" (UniqueName: \"kubernetes.io/projected/9f2abe7e-5864-476a-b5b8-f46b7d7d7350-kube-api-access-gcmrm\") pod \"perses-operator-5446b9c989-rb78r\" (UID: \"9f2abe7e-5864-476a-b5b8-f46b7d7d7350\") " pod="openshift-operators/perses-operator-5446b9c989-rb78r" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.908168 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-xzrfl" Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.958876 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-4586l"] Dec 05 09:04:20 crc kubenswrapper[4997]: I1205 09:04:20.970398 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-rb78r" Dec 05 09:04:21 crc kubenswrapper[4997]: I1205 09:04:21.143546 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-qz4cf"] Dec 05 09:04:21 crc kubenswrapper[4997]: I1205 09:04:21.176083 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-gtbr9"] Dec 05 09:04:21 crc kubenswrapper[4997]: I1205 09:04:21.766767 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-rb78r"] Dec 05 09:04:21 crc kubenswrapper[4997]: I1205 09:04:21.845650 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-xzrfl"] Dec 05 09:04:21 crc kubenswrapper[4997]: W1205 09:04:21.845773 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda0ed268_f4eb_4a9f_8ecf_bd7d318e0f09.slice/crio-d785d92a3270eb11924c2e8b8a1929a1760a89d50dfb417cc8ca7c87ea0dc35c WatchSource:0}: Error finding container d785d92a3270eb11924c2e8b8a1929a1760a89d50dfb417cc8ca7c87ea0dc35c: Status 404 returned error can't find the container with id d785d92a3270eb11924c2e8b8a1929a1760a89d50dfb417cc8ca7c87ea0dc35c Dec 05 09:04:22 crc kubenswrapper[4997]: I1205 09:04:22.040021 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-rb78r" event={"ID":"9f2abe7e-5864-476a-b5b8-f46b7d7d7350","Type":"ContainerStarted","Data":"51ef69439001b3cff53c9775865b7f1e62eec67e7b26cfcc6b73e0711a96a44b"} Dec 05 09:04:22 crc kubenswrapper[4997]: I1205 09:04:22.041948 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-gtbr9" event={"ID":"77d7110e-dc31-46fe-90d0-a1344dea977e","Type":"ContainerStarted","Data":"ac70f53633bc6bba6052475d3a00b1281da0ca0027eabbf9340acfcc68d861b8"} Dec 05 09:04:22 crc kubenswrapper[4997]: I1205 09:04:22.043280 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-xzrfl" event={"ID":"da0ed268-f4eb-4a9f-8ecf-bd7d318e0f09","Type":"ContainerStarted","Data":"d785d92a3270eb11924c2e8b8a1929a1760a89d50dfb417cc8ca7c87ea0dc35c"} Dec 05 09:04:22 crc kubenswrapper[4997]: I1205 09:04:22.044676 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-4586l" event={"ID":"68e3de45-5599-48c3-8b35-73cc4d02250a","Type":"ContainerStarted","Data":"816f62767ae738fcc5abb7eb3c36facf4e6e4107c606662ecb7b8ff126eebc40"} Dec 05 09:04:22 crc kubenswrapper[4997]: I1205 09:04:22.045919 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-qz4cf" event={"ID":"c4adb5f1-1596-4833-92e8-8fddef0bdd40","Type":"ContainerStarted","Data":"900518ab88a75dd168b58e2ed90cb1845db26bd42980db78eed2afa61678d2a1"} Dec 05 09:04:23 crc kubenswrapper[4997]: I1205 09:04:23.388893 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j69xh" Dec 05 09:04:23 crc kubenswrapper[4997]: I1205 09:04:23.480375 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j69xh" Dec 05 09:04:23 crc kubenswrapper[4997]: I1205 09:04:23.776677 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j69xh"] Dec 05 09:04:25 crc kubenswrapper[4997]: I1205 09:04:25.091291 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j69xh" podUID="a5de22f9-fddd-41fd-8a73-f77153be79eb" containerName="registry-server" containerID="cri-o://aaaa317a46da494061c3a7e6a05547e5a372eb429917c38b26e95dc62aabf8a1" gracePeriod=2 Dec 05 09:04:26 crc kubenswrapper[4997]: I1205 09:04:26.107782 4997 generic.go:334] "Generic (PLEG): container finished" podID="a5de22f9-fddd-41fd-8a73-f77153be79eb" containerID="aaaa317a46da494061c3a7e6a05547e5a372eb429917c38b26e95dc62aabf8a1" exitCode=0 Dec 05 09:04:26 crc kubenswrapper[4997]: I1205 09:04:26.107867 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69xh" event={"ID":"a5de22f9-fddd-41fd-8a73-f77153be79eb","Type":"ContainerDied","Data":"aaaa317a46da494061c3a7e6a05547e5a372eb429917c38b26e95dc62aabf8a1"} Dec 05 09:04:27 crc kubenswrapper[4997]: I1205 09:04:27.445023 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-2plrj" podUID="24697cfc-2c63-4b44-b4db-88a361476c5f" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:04:28 crc kubenswrapper[4997]: I1205 09:04:28.313899 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j69xh" Dec 05 09:04:28 crc kubenswrapper[4997]: I1205 09:04:28.422723 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5de22f9-fddd-41fd-8a73-f77153be79eb-utilities\") pod \"a5de22f9-fddd-41fd-8a73-f77153be79eb\" (UID: \"a5de22f9-fddd-41fd-8a73-f77153be79eb\") " Dec 05 09:04:28 crc kubenswrapper[4997]: I1205 09:04:28.422833 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7n6zb\" (UniqueName: \"kubernetes.io/projected/a5de22f9-fddd-41fd-8a73-f77153be79eb-kube-api-access-7n6zb\") pod \"a5de22f9-fddd-41fd-8a73-f77153be79eb\" (UID: \"a5de22f9-fddd-41fd-8a73-f77153be79eb\") " Dec 05 09:04:28 crc kubenswrapper[4997]: I1205 09:04:28.422914 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5de22f9-fddd-41fd-8a73-f77153be79eb-catalog-content\") pod \"a5de22f9-fddd-41fd-8a73-f77153be79eb\" (UID: \"a5de22f9-fddd-41fd-8a73-f77153be79eb\") " Dec 05 09:04:28 crc kubenswrapper[4997]: I1205 09:04:28.424447 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5de22f9-fddd-41fd-8a73-f77153be79eb-utilities" (OuterVolumeSpecName: "utilities") pod "a5de22f9-fddd-41fd-8a73-f77153be79eb" (UID: "a5de22f9-fddd-41fd-8a73-f77153be79eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:04:28 crc kubenswrapper[4997]: I1205 09:04:28.435741 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5de22f9-fddd-41fd-8a73-f77153be79eb-kube-api-access-7n6zb" (OuterVolumeSpecName: "kube-api-access-7n6zb") pod "a5de22f9-fddd-41fd-8a73-f77153be79eb" (UID: "a5de22f9-fddd-41fd-8a73-f77153be79eb"). InnerVolumeSpecName "kube-api-access-7n6zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:04:28 crc kubenswrapper[4997]: I1205 09:04:28.524914 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5de22f9-fddd-41fd-8a73-f77153be79eb-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:04:28 crc kubenswrapper[4997]: I1205 09:04:28.526444 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7n6zb\" (UniqueName: \"kubernetes.io/projected/a5de22f9-fddd-41fd-8a73-f77153be79eb-kube-api-access-7n6zb\") on node \"crc\" DevicePath \"\"" Dec 05 09:04:28 crc kubenswrapper[4997]: I1205 09:04:28.539284 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5de22f9-fddd-41fd-8a73-f77153be79eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5de22f9-fddd-41fd-8a73-f77153be79eb" (UID: "a5de22f9-fddd-41fd-8a73-f77153be79eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:04:28 crc kubenswrapper[4997]: I1205 09:04:28.627307 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5de22f9-fddd-41fd-8a73-f77153be79eb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:04:29 crc kubenswrapper[4997]: I1205 09:04:29.194867 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j69xh" event={"ID":"a5de22f9-fddd-41fd-8a73-f77153be79eb","Type":"ContainerDied","Data":"de1272ac7797c6252eaa586abab865ef52a82b2aa5695724a632d9786300242b"} Dec 05 09:04:29 crc kubenswrapper[4997]: I1205 09:04:29.195376 4997 scope.go:117] "RemoveContainer" containerID="aaaa317a46da494061c3a7e6a05547e5a372eb429917c38b26e95dc62aabf8a1" Dec 05 09:04:29 crc kubenswrapper[4997]: I1205 09:04:29.195220 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j69xh" Dec 05 09:04:29 crc kubenswrapper[4997]: I1205 09:04:29.249735 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j69xh"] Dec 05 09:04:29 crc kubenswrapper[4997]: I1205 09:04:29.257725 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j69xh"] Dec 05 09:04:29 crc kubenswrapper[4997]: I1205 09:04:29.764696 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5de22f9-fddd-41fd-8a73-f77153be79eb" path="/var/lib/kubelet/pods/a5de22f9-fddd-41fd-8a73-f77153be79eb/volumes" Dec 05 09:04:31 crc kubenswrapper[4997]: I1205 09:04:31.883927 4997 scope.go:117] "RemoveContainer" containerID="0d01833a3d3a275f57ea4c10145a9e9a480cabfe0a2eee0d5d2cc06dc86138bf" Dec 05 09:04:31 crc kubenswrapper[4997]: I1205 09:04:31.915935 4997 scope.go:117] "RemoveContainer" containerID="c096f4d786ae02310db2396c1e73b98f92bca4d3a957fa8391fa14175e1da83a" Dec 05 09:04:33 crc kubenswrapper[4997]: I1205 09:04:33.235635 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-gtbr9" event={"ID":"77d7110e-dc31-46fe-90d0-a1344dea977e","Type":"ContainerStarted","Data":"ac183048909e2fcc3f4bb5c73ee741b6cbab1d6ac8ea4c79bb2b2502b2352bfe"} Dec 05 09:04:33 crc kubenswrapper[4997]: I1205 09:04:33.237707 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-xzrfl" event={"ID":"da0ed268-f4eb-4a9f-8ecf-bd7d318e0f09","Type":"ContainerStarted","Data":"aa497862d37bf0142da0eb3de1c031d47fced0010ca8fdd3a2c8401769d5661b"} Dec 05 09:04:33 crc kubenswrapper[4997]: I1205 09:04:33.237919 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-xzrfl" Dec 05 09:04:33 crc kubenswrapper[4997]: I1205 09:04:33.240284 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-4586l" event={"ID":"68e3de45-5599-48c3-8b35-73cc4d02250a","Type":"ContainerStarted","Data":"c36f9cae638657674962898f6721acbec276614cc2a295827aa6314156f8518a"} Dec 05 09:04:33 crc kubenswrapper[4997]: I1205 09:04:33.242558 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-qz4cf" event={"ID":"c4adb5f1-1596-4833-92e8-8fddef0bdd40","Type":"ContainerStarted","Data":"a7611b3d59d213b4a3852b90aa3e1b5bae832c81fc4dcdf38bcc039b017ddf92"} Dec 05 09:04:33 crc kubenswrapper[4997]: I1205 09:04:33.244892 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-rb78r" event={"ID":"9f2abe7e-5864-476a-b5b8-f46b7d7d7350","Type":"ContainerStarted","Data":"eaaba1c6a6c994fb5996f5e09387f1b96ea5aac606ba6c386bef48ac6d58a222"} Dec 05 09:04:33 crc kubenswrapper[4997]: I1205 09:04:33.245031 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-rb78r" Dec 05 09:04:33 crc kubenswrapper[4997]: I1205 09:04:33.245666 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-xzrfl" Dec 05 09:04:33 crc kubenswrapper[4997]: I1205 09:04:33.267180 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-gtbr9" podStartSLOduration=2.5709159980000003 podStartE2EDuration="13.267158129s" podCreationTimestamp="2025-12-05 09:04:20 +0000 UTC" firstStartedPulling="2025-12-05 09:04:21.222739987 +0000 UTC m=+7761.751647248" lastFinishedPulling="2025-12-05 09:04:31.918982128 +0000 UTC m=+7772.447889379" observedRunningTime="2025-12-05 09:04:33.261886998 +0000 UTC m=+7773.790794269" watchObservedRunningTime="2025-12-05 09:04:33.267158129 +0000 UTC m=+7773.796065390" Dec 05 09:04:33 crc kubenswrapper[4997]: I1205 09:04:33.290207 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-4586l" podStartSLOduration=2.456638086 podStartE2EDuration="13.290185081s" podCreationTimestamp="2025-12-05 09:04:20 +0000 UTC" firstStartedPulling="2025-12-05 09:04:21.082567436 +0000 UTC m=+7761.611474697" lastFinishedPulling="2025-12-05 09:04:31.916114431 +0000 UTC m=+7772.445021692" observedRunningTime="2025-12-05 09:04:33.285895855 +0000 UTC m=+7773.814803136" watchObservedRunningTime="2025-12-05 09:04:33.290185081 +0000 UTC m=+7773.819092362" Dec 05 09:04:33 crc kubenswrapper[4997]: I1205 09:04:33.425143 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-rb78r" podStartSLOduration=3.280835105 podStartE2EDuration="13.42512141s" podCreationTimestamp="2025-12-05 09:04:20 +0000 UTC" firstStartedPulling="2025-12-05 09:04:21.774656762 +0000 UTC m=+7762.303564023" lastFinishedPulling="2025-12-05 09:04:31.918943067 +0000 UTC m=+7772.447850328" observedRunningTime="2025-12-05 09:04:33.34909359 +0000 UTC m=+7773.878000861" watchObservedRunningTime="2025-12-05 09:04:33.42512141 +0000 UTC m=+7773.954028671" Dec 05 09:04:33 crc kubenswrapper[4997]: I1205 09:04:33.474034 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-xzrfl" podStartSLOduration=3.403940287 podStartE2EDuration="13.474011939s" podCreationTimestamp="2025-12-05 09:04:20 +0000 UTC" firstStartedPulling="2025-12-05 09:04:21.849315587 +0000 UTC m=+7762.378222858" lastFinishedPulling="2025-12-05 09:04:31.919387259 +0000 UTC m=+7772.448294510" observedRunningTime="2025-12-05 09:04:33.462064927 +0000 UTC m=+7773.990972188" watchObservedRunningTime="2025-12-05 09:04:33.474011939 +0000 UTC m=+7774.002919200" Dec 05 09:04:33 crc kubenswrapper[4997]: I1205 09:04:33.516115 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-58cc577445-qz4cf" podStartSLOduration=2.838499456 podStartE2EDuration="13.516090724s" podCreationTimestamp="2025-12-05 09:04:20 +0000 UTC" firstStartedPulling="2025-12-05 09:04:21.216841068 +0000 UTC m=+7761.745748329" lastFinishedPulling="2025-12-05 09:04:31.894432336 +0000 UTC m=+7772.423339597" observedRunningTime="2025-12-05 09:04:33.508763486 +0000 UTC m=+7774.037670767" watchObservedRunningTime="2025-12-05 09:04:33.516090724 +0000 UTC m=+7774.044997995" Dec 05 09:04:40 crc kubenswrapper[4997]: I1205 09:04:40.974787 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-rb78r" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.447128 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.447956 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="d8facf43-b820-4753-a695-34d7e7b89da9" containerName="openstackclient" containerID="cri-o://ed0731958ed9cfb10278b3374d57d6772ac176a47911b4297300d3e8b2cd651d" gracePeriod=2 Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.461167 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.504280 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 05 09:04:43 crc kubenswrapper[4997]: E1205 09:04:43.504709 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5de22f9-fddd-41fd-8a73-f77153be79eb" containerName="extract-content" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.504729 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5de22f9-fddd-41fd-8a73-f77153be79eb" containerName="extract-content" Dec 05 09:04:43 crc kubenswrapper[4997]: E1205 09:04:43.504763 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5de22f9-fddd-41fd-8a73-f77153be79eb" containerName="extract-utilities" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.504770 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5de22f9-fddd-41fd-8a73-f77153be79eb" containerName="extract-utilities" Dec 05 09:04:43 crc kubenswrapper[4997]: E1205 09:04:43.504778 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5de22f9-fddd-41fd-8a73-f77153be79eb" containerName="registry-server" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.504784 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5de22f9-fddd-41fd-8a73-f77153be79eb" containerName="registry-server" Dec 05 09:04:43 crc kubenswrapper[4997]: E1205 09:04:43.504806 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8facf43-b820-4753-a695-34d7e7b89da9" containerName="openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.504812 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8facf43-b820-4753-a695-34d7e7b89da9" containerName="openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.505007 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5de22f9-fddd-41fd-8a73-f77153be79eb" containerName="registry-server" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.505022 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8facf43-b820-4753-a695-34d7e7b89da9" containerName="openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.505881 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.516690 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzgj2\" (UniqueName: \"kubernetes.io/projected/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-kube-api-access-jzgj2\") pod \"openstackclient\" (UID: \"bf551f77-12ea-4b57-9a4c-ce4cc76d9c55\") " pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.516768 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-openstack-config-secret\") pod \"openstackclient\" (UID: \"bf551f77-12ea-4b57-9a4c-ce4cc76d9c55\") " pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.516849 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-openstack-config\") pod \"openstackclient\" (UID: \"bf551f77-12ea-4b57-9a4c-ce4cc76d9c55\") " pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.533510 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.545944 4997 status_manager.go:875] "Failed to update status for pod" pod="openstack/openstackclient" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf551f77-12ea-4b57-9a4c-ce4cc76d9c55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.rdoproject.org/podified-antelope-centos9/openstack-openstackclient:65066e8ca260a75886ae57f157049605\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"openstackclient\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/clouds.yaml\\\",\\\"name\\\":\\\"openstack-config\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/secure.yaml\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/cloudrc\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jzgj2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:04:43Z\\\"}}\" for pod \"openstack\"/\"openstackclient\": pods \"openstackclient\" not found" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.555670 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 05 09:04:43 crc kubenswrapper[4997]: E1205 09:04:43.556455 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-jzgj2 openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="bf551f77-12ea-4b57-9a4c-ce4cc76d9c55" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.561724 4997 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="d8facf43-b820-4753-a695-34d7e7b89da9" podUID="174c1aed-3954-4867-ad3e-0a5d51579490" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.577607 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.613724 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.615379 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.618366 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzgj2\" (UniqueName: \"kubernetes.io/projected/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-kube-api-access-jzgj2\") pod \"openstackclient\" (UID: \"bf551f77-12ea-4b57-9a4c-ce4cc76d9c55\") " pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.618407 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-openstack-config-secret\") pod \"openstackclient\" (UID: \"bf551f77-12ea-4b57-9a4c-ce4cc76d9c55\") " pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.618451 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-openstack-config\") pod \"openstackclient\" (UID: \"bf551f77-12ea-4b57-9a4c-ce4cc76d9c55\") " pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.623177 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-openstack-config\") pod \"openstackclient\" (UID: \"bf551f77-12ea-4b57-9a4c-ce4cc76d9c55\") " pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.624564 4997 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bf551f77-12ea-4b57-9a4c-ce4cc76d9c55" podUID="174c1aed-3954-4867-ad3e-0a5d51579490" Dec 05 09:04:43 crc kubenswrapper[4997]: E1205 09:04:43.624989 4997 projected.go:194] Error preparing data for projected volume kube-api-access-jzgj2 for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (bf551f77-12ea-4b57-9a4c-ce4cc76d9c55) does not match the UID in record. The object might have been deleted and then recreated Dec 05 09:04:43 crc kubenswrapper[4997]: E1205 09:04:43.625145 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-kube-api-access-jzgj2 podName:bf551f77-12ea-4b57-9a4c-ce4cc76d9c55 nodeName:}" failed. No retries permitted until 2025-12-05 09:04:44.125106667 +0000 UTC m=+7784.654013998 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-jzgj2" (UniqueName: "kubernetes.io/projected/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-kube-api-access-jzgj2") pod "openstackclient" (UID: "bf551f77-12ea-4b57-9a4c-ce4cc76d9c55") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (bf551f77-12ea-4b57-9a4c-ce4cc76d9c55) does not match the UID in record. The object might have been deleted and then recreated Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.640604 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.650109 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-openstack-config-secret\") pod \"openstackclient\" (UID: \"bf551f77-12ea-4b57-9a4c-ce4cc76d9c55\") " pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.711695 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.713434 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.717093 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-2dz7b" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.721386 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/174c1aed-3954-4867-ad3e-0a5d51579490-openstack-config-secret\") pod \"openstackclient\" (UID: \"174c1aed-3954-4867-ad3e-0a5d51579490\") " pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.721507 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/174c1aed-3954-4867-ad3e-0a5d51579490-openstack-config\") pod \"openstackclient\" (UID: \"174c1aed-3954-4867-ad3e-0a5d51579490\") " pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.721550 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpk4v\" (UniqueName: \"kubernetes.io/projected/174c1aed-3954-4867-ad3e-0a5d51579490-kube-api-access-gpk4v\") pod \"openstackclient\" (UID: \"174c1aed-3954-4867-ad3e-0a5d51579490\") " pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.735825 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.824277 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/174c1aed-3954-4867-ad3e-0a5d51579490-openstack-config\") pod \"openstackclient\" (UID: \"174c1aed-3954-4867-ad3e-0a5d51579490\") " pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.824375 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpk4v\" (UniqueName: \"kubernetes.io/projected/174c1aed-3954-4867-ad3e-0a5d51579490-kube-api-access-gpk4v\") pod \"openstackclient\" (UID: \"174c1aed-3954-4867-ad3e-0a5d51579490\") " pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.824522 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/174c1aed-3954-4867-ad3e-0a5d51579490-openstack-config-secret\") pod \"openstackclient\" (UID: \"174c1aed-3954-4867-ad3e-0a5d51579490\") " pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.824545 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q652k\" (UniqueName: \"kubernetes.io/projected/53022e5c-c432-461a-bb5a-31844df7c3e9-kube-api-access-q652k\") pod \"kube-state-metrics-0\" (UID: \"53022e5c-c432-461a-bb5a-31844df7c3e9\") " pod="openstack/kube-state-metrics-0" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.827173 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/174c1aed-3954-4867-ad3e-0a5d51579490-openstack-config\") pod \"openstackclient\" (UID: \"174c1aed-3954-4867-ad3e-0a5d51579490\") " pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.836699 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/174c1aed-3954-4867-ad3e-0a5d51579490-openstack-config-secret\") pod \"openstackclient\" (UID: \"174c1aed-3954-4867-ad3e-0a5d51579490\") " pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.873891 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpk4v\" (UniqueName: \"kubernetes.io/projected/174c1aed-3954-4867-ad3e-0a5d51579490-kube-api-access-gpk4v\") pod \"openstackclient\" (UID: \"174c1aed-3954-4867-ad3e-0a5d51579490\") " pod="openstack/openstackclient" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.936401 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q652k\" (UniqueName: \"kubernetes.io/projected/53022e5c-c432-461a-bb5a-31844df7c3e9-kube-api-access-q652k\") pod \"kube-state-metrics-0\" (UID: \"53022e5c-c432-461a-bb5a-31844df7c3e9\") " pod="openstack/kube-state-metrics-0" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.964823 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q652k\" (UniqueName: \"kubernetes.io/projected/53022e5c-c432-461a-bb5a-31844df7c3e9-kube-api-access-q652k\") pod \"kube-state-metrics-0\" (UID: \"53022e5c-c432-461a-bb5a-31844df7c3e9\") " pod="openstack/kube-state-metrics-0" Dec 05 09:04:43 crc kubenswrapper[4997]: I1205 09:04:43.988110 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.062087 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.141085 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzgj2\" (UniqueName: \"kubernetes.io/projected/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-kube-api-access-jzgj2\") pod \"openstackclient\" (UID: \"bf551f77-12ea-4b57-9a4c-ce4cc76d9c55\") " pod="openstack/openstackclient" Dec 05 09:04:44 crc kubenswrapper[4997]: E1205 09:04:44.143745 4997 projected.go:194] Error preparing data for projected volume kube-api-access-jzgj2 for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (bf551f77-12ea-4b57-9a4c-ce4cc76d9c55) does not match the UID in record. The object might have been deleted and then recreated Dec 05 09:04:44 crc kubenswrapper[4997]: E1205 09:04:44.143821 4997 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-kube-api-access-jzgj2 podName:bf551f77-12ea-4b57-9a4c-ce4cc76d9c55 nodeName:}" failed. No retries permitted until 2025-12-05 09:04:45.143804927 +0000 UTC m=+7785.672712188 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-jzgj2" (UniqueName: "kubernetes.io/projected/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-kube-api-access-jzgj2") pod "openstackclient" (UID: "bf551f77-12ea-4b57-9a4c-ce4cc76d9c55") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (bf551f77-12ea-4b57-9a4c-ce4cc76d9c55) does not match the UID in record. The object might have been deleted and then recreated Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.384321 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.388218 4997 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bf551f77-12ea-4b57-9a4c-ce4cc76d9c55" podUID="174c1aed-3954-4867-ad3e-0a5d51579490" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.417741 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.423323 4997 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bf551f77-12ea-4b57-9a4c-ce4cc76d9c55" podUID="174c1aed-3954-4867-ad3e-0a5d51579490" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.453290 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-openstack-config\") pod \"bf551f77-12ea-4b57-9a4c-ce4cc76d9c55\" (UID: \"bf551f77-12ea-4b57-9a4c-ce4cc76d9c55\") " Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.453471 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-openstack-config-secret\") pod \"bf551f77-12ea-4b57-9a4c-ce4cc76d9c55\" (UID: \"bf551f77-12ea-4b57-9a4c-ce4cc76d9c55\") " Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.453911 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzgj2\" (UniqueName: \"kubernetes.io/projected/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-kube-api-access-jzgj2\") on node \"crc\" DevicePath \"\"" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.455349 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "bf551f77-12ea-4b57-9a4c-ce4cc76d9c55" (UID: "bf551f77-12ea-4b57-9a4c-ce4cc76d9c55"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.468371 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "bf551f77-12ea-4b57-9a4c-ce4cc76d9c55" (UID: "bf551f77-12ea-4b57-9a4c-ce4cc76d9c55"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.532679 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.535077 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.545131 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.545305 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.545403 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.545997 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.546112 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-g98rl" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.555505 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7aea731b-86e5-4305-954d-1de96d638b44-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.555573 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/7aea731b-86e5-4305-954d-1de96d638b44-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.555642 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7aea731b-86e5-4305-954d-1de96d638b44-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.555676 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7aea731b-86e5-4305-954d-1de96d638b44-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.555704 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/7aea731b-86e5-4305-954d-1de96d638b44-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.555729 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/7aea731b-86e5-4305-954d-1de96d638b44-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.555761 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk4cs\" (UniqueName: \"kubernetes.io/projected/7aea731b-86e5-4305-954d-1de96d638b44-kube-api-access-fk4cs\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.555838 4997 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.555850 4997 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.559048 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.657183 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/7aea731b-86e5-4305-954d-1de96d638b44-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.657257 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7aea731b-86e5-4305-954d-1de96d638b44-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.657294 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7aea731b-86e5-4305-954d-1de96d638b44-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.657323 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/7aea731b-86e5-4305-954d-1de96d638b44-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.657352 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/7aea731b-86e5-4305-954d-1de96d638b44-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.657382 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk4cs\" (UniqueName: \"kubernetes.io/projected/7aea731b-86e5-4305-954d-1de96d638b44-kube-api-access-fk4cs\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.657436 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7aea731b-86e5-4305-954d-1de96d638b44-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.659992 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/7aea731b-86e5-4305-954d-1de96d638b44-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.661047 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/7aea731b-86e5-4305-954d-1de96d638b44-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.665167 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7aea731b-86e5-4305-954d-1de96d638b44-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.665204 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7aea731b-86e5-4305-954d-1de96d638b44-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.665565 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/7aea731b-86e5-4305-954d-1de96d638b44-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.670005 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7aea731b-86e5-4305-954d-1de96d638b44-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.682120 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk4cs\" (UniqueName: \"kubernetes.io/projected/7aea731b-86e5-4305-954d-1de96d638b44-kube-api-access-fk4cs\") pod \"alertmanager-metric-storage-0\" (UID: \"7aea731b-86e5-4305-954d-1de96d638b44\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.883765 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 05 09:04:44 crc kubenswrapper[4997]: I1205 09:04:44.906966 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.105486 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.149301 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.151989 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.161900 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.161958 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.162089 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-vdvfh" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.162271 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.162314 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.162457 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.211419 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/226e3599-1e1d-4216-b76a-7161e5c57556-config\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.214037 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/226e3599-1e1d-4216-b76a-7161e5c57556-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.214155 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0407ef24-b746-43aa-8e6d-dfff779b820d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0407ef24-b746-43aa-8e6d-dfff779b820d\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.214276 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdrcw\" (UniqueName: \"kubernetes.io/projected/226e3599-1e1d-4216-b76a-7161e5c57556-kube-api-access-gdrcw\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.214417 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/226e3599-1e1d-4216-b76a-7161e5c57556-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.214511 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/226e3599-1e1d-4216-b76a-7161e5c57556-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.214651 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/226e3599-1e1d-4216-b76a-7161e5c57556-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.214849 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/226e3599-1e1d-4216-b76a-7161e5c57556-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.237835 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.318952 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/226e3599-1e1d-4216-b76a-7161e5c57556-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.319178 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/226e3599-1e1d-4216-b76a-7161e5c57556-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.319212 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/226e3599-1e1d-4216-b76a-7161e5c57556-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.319280 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/226e3599-1e1d-4216-b76a-7161e5c57556-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.319310 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/226e3599-1e1d-4216-b76a-7161e5c57556-config\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.319343 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/226e3599-1e1d-4216-b76a-7161e5c57556-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.319373 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0407ef24-b746-43aa-8e6d-dfff779b820d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0407ef24-b746-43aa-8e6d-dfff779b820d\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.319407 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdrcw\" (UniqueName: \"kubernetes.io/projected/226e3599-1e1d-4216-b76a-7161e5c57556-kube-api-access-gdrcw\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.326602 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/226e3599-1e1d-4216-b76a-7161e5c57556-config\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.327116 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/226e3599-1e1d-4216-b76a-7161e5c57556-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.348241 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/226e3599-1e1d-4216-b76a-7161e5c57556-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.355958 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/226e3599-1e1d-4216-b76a-7161e5c57556-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.356395 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/226e3599-1e1d-4216-b76a-7161e5c57556-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.368186 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/226e3599-1e1d-4216-b76a-7161e5c57556-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.389752 4997 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.389793 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0407ef24-b746-43aa-8e6d-dfff779b820d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0407ef24-b746-43aa-8e6d-dfff779b820d\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/da9f312cd1883a785386d62ab96930d11123d5c2075e7b141e09aebb36f24341/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.417121 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdrcw\" (UniqueName: \"kubernetes.io/projected/226e3599-1e1d-4216-b76a-7161e5c57556-kube-api-access-gdrcw\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.424214 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"174c1aed-3954-4867-ad3e-0a5d51579490","Type":"ContainerStarted","Data":"f9517739fdea461069b66c07e5a69aec5e3d3b91cc18addbd4a3959b976cacc8"} Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.431678 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.433932 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"53022e5c-c432-461a-bb5a-31844df7c3e9","Type":"ContainerStarted","Data":"7e4a3fc06ee417c28f29125fd883eaf73ff08239fe43a3b3808aec3c6e51776b"} Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.443095 4997 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bf551f77-12ea-4b57-9a4c-ce4cc76d9c55" podUID="174c1aed-3954-4867-ad3e-0a5d51579490" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.463194 4997 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bf551f77-12ea-4b57-9a4c-ce4cc76d9c55" podUID="174c1aed-3954-4867-ad3e-0a5d51579490" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.639897 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0407ef24-b746-43aa-8e6d-dfff779b820d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0407ef24-b746-43aa-8e6d-dfff779b820d\") pod \"prometheus-metric-storage-0\" (UID: \"226e3599-1e1d-4216-b76a-7161e5c57556\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: E1205 09:04:45.744251 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf551f77_12ea_4b57_9a4c_ce4cc76d9c55.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8facf43_b820_4753_a695_34d7e7b89da9.slice/crio-ed0731958ed9cfb10278b3374d57d6772ac176a47911b4297300d3e8b2cd651d.scope\": RecentStats: unable to find data in memory cache]" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.789577 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf551f77-12ea-4b57-9a4c-ce4cc76d9c55" path="/var/lib/kubelet/pods/bf551f77-12ea-4b57-9a4c-ce4cc76d9c55/volumes" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.821078 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 09:04:45 crc kubenswrapper[4997]: I1205 09:04:45.857557 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 05 09:04:46 crc kubenswrapper[4997]: I1205 09:04:46.440960 4997 generic.go:334] "Generic (PLEG): container finished" podID="d8facf43-b820-4753-a695-34d7e7b89da9" containerID="ed0731958ed9cfb10278b3374d57d6772ac176a47911b4297300d3e8b2cd651d" exitCode=137 Dec 05 09:04:46 crc kubenswrapper[4997]: W1205 09:04:46.552164 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7aea731b_86e5_4305_954d_1de96d638b44.slice/crio-140c152f30c56c6f36573fc51548d010341dc897aeafb3a5bc3f77881b7b1046 WatchSource:0}: Error finding container 140c152f30c56c6f36573fc51548d010341dc897aeafb3a5bc3f77881b7b1046: Status 404 returned error can't find the container with id 140c152f30c56c6f36573fc51548d010341dc897aeafb3a5bc3f77881b7b1046 Dec 05 09:04:47 crc kubenswrapper[4997]: I1205 09:04:47.165770 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:04:47 crc kubenswrapper[4997]: W1205 09:04:47.346895 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod226e3599_1e1d_4216_b76a_7161e5c57556.slice/crio-d6749d252fe3585fdef435e63d81878f696c2c7a8b5c8aad85219fd4fdaf495c WatchSource:0}: Error finding container d6749d252fe3585fdef435e63d81878f696c2c7a8b5c8aad85219fd4fdaf495c: Status 404 returned error can't find the container with id d6749d252fe3585fdef435e63d81878f696c2c7a8b5c8aad85219fd4fdaf495c Dec 05 09:04:47 crc kubenswrapper[4997]: I1205 09:04:47.478991 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"226e3599-1e1d-4216-b76a-7161e5c57556","Type":"ContainerStarted","Data":"d6749d252fe3585fdef435e63d81878f696c2c7a8b5c8aad85219fd4fdaf495c"} Dec 05 09:04:47 crc kubenswrapper[4997]: I1205 09:04:47.496031 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"174c1aed-3954-4867-ad3e-0a5d51579490","Type":"ContainerStarted","Data":"10525da2f57567d3d03af150e72fa44e5d40af6770b35c9716ce707f1cb725b4"} Dec 05 09:04:47 crc kubenswrapper[4997]: I1205 09:04:47.547981 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"7aea731b-86e5-4305-954d-1de96d638b44","Type":"ContainerStarted","Data":"140c152f30c56c6f36573fc51548d010341dc897aeafb3a5bc3f77881b7b1046"} Dec 05 09:04:47 crc kubenswrapper[4997]: I1205 09:04:47.550125 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=4.550094439 podStartE2EDuration="4.550094439s" podCreationTimestamp="2025-12-05 09:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:04:47.526845302 +0000 UTC m=+7788.055752573" watchObservedRunningTime="2025-12-05 09:04:47.550094439 +0000 UTC m=+7788.079001700" Dec 05 09:04:47 crc kubenswrapper[4997]: I1205 09:04:47.834392 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:04:47 crc kubenswrapper[4997]: I1205 09:04:47.894070 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drn6r\" (UniqueName: \"kubernetes.io/projected/d8facf43-b820-4753-a695-34d7e7b89da9-kube-api-access-drn6r\") pod \"d8facf43-b820-4753-a695-34d7e7b89da9\" (UID: \"d8facf43-b820-4753-a695-34d7e7b89da9\") " Dec 05 09:04:47 crc kubenswrapper[4997]: I1205 09:04:47.894276 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d8facf43-b820-4753-a695-34d7e7b89da9-openstack-config\") pod \"d8facf43-b820-4753-a695-34d7e7b89da9\" (UID: \"d8facf43-b820-4753-a695-34d7e7b89da9\") " Dec 05 09:04:47 crc kubenswrapper[4997]: I1205 09:04:47.894358 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8facf43-b820-4753-a695-34d7e7b89da9-openstack-config-secret\") pod \"d8facf43-b820-4753-a695-34d7e7b89da9\" (UID: \"d8facf43-b820-4753-a695-34d7e7b89da9\") " Dec 05 09:04:47 crc kubenswrapper[4997]: I1205 09:04:47.900928 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8facf43-b820-4753-a695-34d7e7b89da9-kube-api-access-drn6r" (OuterVolumeSpecName: "kube-api-access-drn6r") pod "d8facf43-b820-4753-a695-34d7e7b89da9" (UID: "d8facf43-b820-4753-a695-34d7e7b89da9"). InnerVolumeSpecName "kube-api-access-drn6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:04:47 crc kubenswrapper[4997]: I1205 09:04:47.921481 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8facf43-b820-4753-a695-34d7e7b89da9-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "d8facf43-b820-4753-a695-34d7e7b89da9" (UID: "d8facf43-b820-4753-a695-34d7e7b89da9"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:04:47 crc kubenswrapper[4997]: I1205 09:04:47.956204 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8facf43-b820-4753-a695-34d7e7b89da9-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "d8facf43-b820-4753-a695-34d7e7b89da9" (UID: "d8facf43-b820-4753-a695-34d7e7b89da9"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:04:47 crc kubenswrapper[4997]: I1205 09:04:47.999217 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drn6r\" (UniqueName: \"kubernetes.io/projected/d8facf43-b820-4753-a695-34d7e7b89da9-kube-api-access-drn6r\") on node \"crc\" DevicePath \"\"" Dec 05 09:04:47 crc kubenswrapper[4997]: I1205 09:04:47.999270 4997 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d8facf43-b820-4753-a695-34d7e7b89da9-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:04:47 crc kubenswrapper[4997]: I1205 09:04:47.999283 4997 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8facf43-b820-4753-a695-34d7e7b89da9-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 05 09:04:48 crc kubenswrapper[4997]: I1205 09:04:48.561073 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"53022e5c-c432-461a-bb5a-31844df7c3e9","Type":"ContainerStarted","Data":"3690e8bb7f4047027ec5dda45e97bfe280f64079e8b3c9e0fdb77294cfbc9cc6"} Dec 05 09:04:48 crc kubenswrapper[4997]: I1205 09:04:48.561898 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 05 09:04:48 crc kubenswrapper[4997]: I1205 09:04:48.564432 4997 scope.go:117] "RemoveContainer" containerID="ed0731958ed9cfb10278b3374d57d6772ac176a47911b4297300d3e8b2cd651d" Dec 05 09:04:48 crc kubenswrapper[4997]: I1205 09:04:48.564493 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:04:48 crc kubenswrapper[4997]: I1205 09:04:48.586118 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.022388665 podStartE2EDuration="5.586097211s" podCreationTimestamp="2025-12-05 09:04:43 +0000 UTC" firstStartedPulling="2025-12-05 09:04:44.88431349 +0000 UTC m=+7785.413220751" lastFinishedPulling="2025-12-05 09:04:47.448022036 +0000 UTC m=+7787.976929297" observedRunningTime="2025-12-05 09:04:48.575406133 +0000 UTC m=+7789.104313414" watchObservedRunningTime="2025-12-05 09:04:48.586097211 +0000 UTC m=+7789.115004472" Dec 05 09:04:48 crc kubenswrapper[4997]: I1205 09:04:48.589351 4997 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="d8facf43-b820-4753-a695-34d7e7b89da9" podUID="174c1aed-3954-4867-ad3e-0a5d51579490" Dec 05 09:04:49 crc kubenswrapper[4997]: I1205 09:04:49.763365 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8facf43-b820-4753-a695-34d7e7b89da9" path="/var/lib/kubelet/pods/d8facf43-b820-4753-a695-34d7e7b89da9/volumes" Dec 05 09:04:52 crc kubenswrapper[4997]: I1205 09:04:52.624358 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"226e3599-1e1d-4216-b76a-7161e5c57556","Type":"ContainerStarted","Data":"dc15e547394c49282ba1275a403f4d3c25f52c9f02ae2b7939a64b0fbbf67fa2"} Dec 05 09:04:52 crc kubenswrapper[4997]: I1205 09:04:52.626449 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"7aea731b-86e5-4305-954d-1de96d638b44","Type":"ContainerStarted","Data":"3b17cdc304950d0a707b32b9dee0a587f6b27fa2d3a06c47cbe9075691225839"} Dec 05 09:04:54 crc kubenswrapper[4997]: I1205 09:04:54.065641 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 05 09:04:56 crc kubenswrapper[4997]: I1205 09:04:56.056444 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-nlgxp"] Dec 05 09:04:56 crc kubenswrapper[4997]: I1205 09:04:56.059671 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-c014-account-create-update-k8fmb"] Dec 05 09:04:56 crc kubenswrapper[4997]: I1205 09:04:56.071599 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-nlgxp"] Dec 05 09:04:56 crc kubenswrapper[4997]: I1205 09:04:56.082862 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-c014-account-create-update-k8fmb"] Dec 05 09:04:57 crc kubenswrapper[4997]: I1205 09:04:57.760182 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95769d84-8dcc-4626-8f14-e4015d3cfa5e" path="/var/lib/kubelet/pods/95769d84-8dcc-4626-8f14-e4015d3cfa5e/volumes" Dec 05 09:04:57 crc kubenswrapper[4997]: I1205 09:04:57.761079 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc83f439-c922-4812-ba87-59c083f3885d" path="/var/lib/kubelet/pods/bc83f439-c922-4812-ba87-59c083f3885d/volumes" Dec 05 09:04:58 crc kubenswrapper[4997]: I1205 09:04:58.685543 4997 generic.go:334] "Generic (PLEG): container finished" podID="226e3599-1e1d-4216-b76a-7161e5c57556" containerID="dc15e547394c49282ba1275a403f4d3c25f52c9f02ae2b7939a64b0fbbf67fa2" exitCode=0 Dec 05 09:04:58 crc kubenswrapper[4997]: I1205 09:04:58.685631 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"226e3599-1e1d-4216-b76a-7161e5c57556","Type":"ContainerDied","Data":"dc15e547394c49282ba1275a403f4d3c25f52c9f02ae2b7939a64b0fbbf67fa2"} Dec 05 09:04:58 crc kubenswrapper[4997]: I1205 09:04:58.687833 4997 generic.go:334] "Generic (PLEG): container finished" podID="7aea731b-86e5-4305-954d-1de96d638b44" containerID="3b17cdc304950d0a707b32b9dee0a587f6b27fa2d3a06c47cbe9075691225839" exitCode=0 Dec 05 09:04:58 crc kubenswrapper[4997]: I1205 09:04:58.687861 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"7aea731b-86e5-4305-954d-1de96d638b44","Type":"ContainerDied","Data":"3b17cdc304950d0a707b32b9dee0a587f6b27fa2d3a06c47cbe9075691225839"} Dec 05 09:05:11 crc kubenswrapper[4997]: I1205 09:05:11.853295 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"226e3599-1e1d-4216-b76a-7161e5c57556","Type":"ContainerStarted","Data":"5d868dd646ffd2d600e6aa5bf11d02ae0612043aa6c26fcef5fc54c9e47f9919"} Dec 05 09:05:11 crc kubenswrapper[4997]: I1205 09:05:11.856575 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"7aea731b-86e5-4305-954d-1de96d638b44","Type":"ContainerStarted","Data":"0c5a57ac32c9419d167eca6a444e862271e45740e51365d0a4a9d714c3b2f8e7"} Dec 05 09:05:14 crc kubenswrapper[4997]: I1205 09:05:14.887270 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"7aea731b-86e5-4305-954d-1de96d638b44","Type":"ContainerStarted","Data":"eaa2fdff2eab5f1348c2ed7b631077a833ed4e7ae9c52120e1db5d080615af07"} Dec 05 09:05:14 crc kubenswrapper[4997]: I1205 09:05:14.888851 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Dec 05 09:05:14 crc kubenswrapper[4997]: I1205 09:05:14.890419 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Dec 05 09:05:14 crc kubenswrapper[4997]: I1205 09:05:14.914510 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=6.027012845 podStartE2EDuration="30.914440439s" podCreationTimestamp="2025-12-05 09:04:44 +0000 UTC" firstStartedPulling="2025-12-05 09:04:46.58688014 +0000 UTC m=+7787.115787401" lastFinishedPulling="2025-12-05 09:05:11.474307734 +0000 UTC m=+7812.003214995" observedRunningTime="2025-12-05 09:05:14.907761819 +0000 UTC m=+7815.436669090" watchObservedRunningTime="2025-12-05 09:05:14.914440439 +0000 UTC m=+7815.443347710" Dec 05 09:05:15 crc kubenswrapper[4997]: I1205 09:05:15.901517 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"226e3599-1e1d-4216-b76a-7161e5c57556","Type":"ContainerStarted","Data":"485a9d8a664b736386d656e3539a88f48eed61c144ca3f7154481d43354d7ccf"} Dec 05 09:05:17 crc kubenswrapper[4997]: I1205 09:05:17.066141 4997 scope.go:117] "RemoveContainer" containerID="a41934137ae05c0fe072630de94cbe2af26dfc440898cdc6d8306c4988f20cba" Dec 05 09:05:17 crc kubenswrapper[4997]: I1205 09:05:17.442042 4997 scope.go:117] "RemoveContainer" containerID="4eafa7fef7de6ce5f9bf3e3563cb746478083b69063f688fd3f54de37f880b38" Dec 05 09:05:17 crc kubenswrapper[4997]: I1205 09:05:17.921146 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"226e3599-1e1d-4216-b76a-7161e5c57556","Type":"ContainerStarted","Data":"30f56113db60ae9a2fbef2d2910e007bb457e39e1127c8e27c607a976faa4a14"} Dec 05 09:05:17 crc kubenswrapper[4997]: I1205 09:05:17.970091 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.8027731559999998 podStartE2EDuration="33.970065734s" podCreationTimestamp="2025-12-05 09:04:44 +0000 UTC" firstStartedPulling="2025-12-05 09:04:47.358352838 +0000 UTC m=+7787.887260099" lastFinishedPulling="2025-12-05 09:05:17.525645416 +0000 UTC m=+7818.054552677" observedRunningTime="2025-12-05 09:05:17.958389159 +0000 UTC m=+7818.487296430" watchObservedRunningTime="2025-12-05 09:05:17.970065734 +0000 UTC m=+7818.498972995" Dec 05 09:05:20 crc kubenswrapper[4997]: I1205 09:05:20.822380 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 05 09:05:21 crc kubenswrapper[4997]: I1205 09:05:21.040881 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-hjb78"] Dec 05 09:05:21 crc kubenswrapper[4997]: I1205 09:05:21.048480 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-hjb78"] Dec 05 09:05:21 crc kubenswrapper[4997]: I1205 09:05:21.759185 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56f98ba4-c397-45b4-b6fa-3c1b2ea93543" path="/var/lib/kubelet/pods/56f98ba4-c397-45b4-b6fa-3c1b2ea93543/volumes" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.350721 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.356099 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.358201 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.358377 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.374540 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.460541 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51b4105e-c02f-40ac-a542-c2b86d0c6f64-log-httpd\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.460707 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51b4105e-c02f-40ac-a542-c2b86d0c6f64-run-httpd\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.460787 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-config-data\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.461182 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.461326 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-scripts\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.461498 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kjnz\" (UniqueName: \"kubernetes.io/projected/51b4105e-c02f-40ac-a542-c2b86d0c6f64-kube-api-access-9kjnz\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.461731 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.563323 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kjnz\" (UniqueName: \"kubernetes.io/projected/51b4105e-c02f-40ac-a542-c2b86d0c6f64-kube-api-access-9kjnz\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.563754 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.563855 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51b4105e-c02f-40ac-a542-c2b86d0c6f64-log-httpd\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.563959 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51b4105e-c02f-40ac-a542-c2b86d0c6f64-run-httpd\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.564034 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-config-data\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.564126 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.564208 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-scripts\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.564391 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51b4105e-c02f-40ac-a542-c2b86d0c6f64-run-httpd\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.564558 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51b4105e-c02f-40ac-a542-c2b86d0c6f64-log-httpd\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.569889 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.570471 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-scripts\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.583036 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.586594 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-config-data\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.603952 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kjnz\" (UniqueName: \"kubernetes.io/projected/51b4105e-c02f-40ac-a542-c2b86d0c6f64-kube-api-access-9kjnz\") pod \"ceilometer-0\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " pod="openstack/ceilometer-0" Dec 05 09:05:22 crc kubenswrapper[4997]: I1205 09:05:22.687652 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:05:23 crc kubenswrapper[4997]: I1205 09:05:23.221121 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:05:23 crc kubenswrapper[4997]: W1205 09:05:23.228774 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51b4105e_c02f_40ac_a542_c2b86d0c6f64.slice/crio-b16599faab9555e9afb9b63d673b991dbba3a329ec538d59e13fff22a2398c0c WatchSource:0}: Error finding container b16599faab9555e9afb9b63d673b991dbba3a329ec538d59e13fff22a2398c0c: Status 404 returned error can't find the container with id b16599faab9555e9afb9b63d673b991dbba3a329ec538d59e13fff22a2398c0c Dec 05 09:05:23 crc kubenswrapper[4997]: I1205 09:05:23.995958 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51b4105e-c02f-40ac-a542-c2b86d0c6f64","Type":"ContainerStarted","Data":"b16599faab9555e9afb9b63d673b991dbba3a329ec538d59e13fff22a2398c0c"} Dec 05 09:05:28 crc kubenswrapper[4997]: I1205 09:05:28.040294 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51b4105e-c02f-40ac-a542-c2b86d0c6f64","Type":"ContainerStarted","Data":"be7869d770ed2907f9c09b160fc08e7e9b16d2c13a7cb6d244461029f02fa7c9"} Dec 05 09:05:29 crc kubenswrapper[4997]: I1205 09:05:29.052551 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51b4105e-c02f-40ac-a542-c2b86d0c6f64","Type":"ContainerStarted","Data":"dc3a9ec7865fd5c8d34148c68fdd3c653709a6669deb39b13eebe6e530600fe8"} Dec 05 09:05:29 crc kubenswrapper[4997]: I1205 09:05:29.267395 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-749sz"] Dec 05 09:05:29 crc kubenswrapper[4997]: I1205 09:05:29.269684 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-749sz" Dec 05 09:05:29 crc kubenswrapper[4997]: I1205 09:05:29.289825 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-749sz"] Dec 05 09:05:29 crc kubenswrapper[4997]: I1205 09:05:29.303969 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a70d008-7f96-498c-a5d6-5c22cf77a41a-catalog-content\") pod \"certified-operators-749sz\" (UID: \"0a70d008-7f96-498c-a5d6-5c22cf77a41a\") " pod="openshift-marketplace/certified-operators-749sz" Dec 05 09:05:29 crc kubenswrapper[4997]: I1205 09:05:29.304040 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a70d008-7f96-498c-a5d6-5c22cf77a41a-utilities\") pod \"certified-operators-749sz\" (UID: \"0a70d008-7f96-498c-a5d6-5c22cf77a41a\") " pod="openshift-marketplace/certified-operators-749sz" Dec 05 09:05:29 crc kubenswrapper[4997]: I1205 09:05:29.304070 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nnnr\" (UniqueName: \"kubernetes.io/projected/0a70d008-7f96-498c-a5d6-5c22cf77a41a-kube-api-access-6nnnr\") pod \"certified-operators-749sz\" (UID: \"0a70d008-7f96-498c-a5d6-5c22cf77a41a\") " pod="openshift-marketplace/certified-operators-749sz" Dec 05 09:05:29 crc kubenswrapper[4997]: I1205 09:05:29.406699 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a70d008-7f96-498c-a5d6-5c22cf77a41a-catalog-content\") pod \"certified-operators-749sz\" (UID: \"0a70d008-7f96-498c-a5d6-5c22cf77a41a\") " pod="openshift-marketplace/certified-operators-749sz" Dec 05 09:05:29 crc kubenswrapper[4997]: I1205 09:05:29.406748 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a70d008-7f96-498c-a5d6-5c22cf77a41a-utilities\") pod \"certified-operators-749sz\" (UID: \"0a70d008-7f96-498c-a5d6-5c22cf77a41a\") " pod="openshift-marketplace/certified-operators-749sz" Dec 05 09:05:29 crc kubenswrapper[4997]: I1205 09:05:29.406778 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nnnr\" (UniqueName: \"kubernetes.io/projected/0a70d008-7f96-498c-a5d6-5c22cf77a41a-kube-api-access-6nnnr\") pod \"certified-operators-749sz\" (UID: \"0a70d008-7f96-498c-a5d6-5c22cf77a41a\") " pod="openshift-marketplace/certified-operators-749sz" Dec 05 09:05:29 crc kubenswrapper[4997]: I1205 09:05:29.407237 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a70d008-7f96-498c-a5d6-5c22cf77a41a-catalog-content\") pod \"certified-operators-749sz\" (UID: \"0a70d008-7f96-498c-a5d6-5c22cf77a41a\") " pod="openshift-marketplace/certified-operators-749sz" Dec 05 09:05:29 crc kubenswrapper[4997]: I1205 09:05:29.407402 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a70d008-7f96-498c-a5d6-5c22cf77a41a-utilities\") pod \"certified-operators-749sz\" (UID: \"0a70d008-7f96-498c-a5d6-5c22cf77a41a\") " pod="openshift-marketplace/certified-operators-749sz" Dec 05 09:05:29 crc kubenswrapper[4997]: I1205 09:05:29.428763 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nnnr\" (UniqueName: \"kubernetes.io/projected/0a70d008-7f96-498c-a5d6-5c22cf77a41a-kube-api-access-6nnnr\") pod \"certified-operators-749sz\" (UID: \"0a70d008-7f96-498c-a5d6-5c22cf77a41a\") " pod="openshift-marketplace/certified-operators-749sz" Dec 05 09:05:29 crc kubenswrapper[4997]: I1205 09:05:29.603197 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-749sz" Dec 05 09:05:30 crc kubenswrapper[4997]: W1205 09:05:30.258564 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a70d008_7f96_498c_a5d6_5c22cf77a41a.slice/crio-546be805880af13f8c716ae9a2825fcc3c40f83b1e80958cbfa0cbbd8531bcfc WatchSource:0}: Error finding container 546be805880af13f8c716ae9a2825fcc3c40f83b1e80958cbfa0cbbd8531bcfc: Status 404 returned error can't find the container with id 546be805880af13f8c716ae9a2825fcc3c40f83b1e80958cbfa0cbbd8531bcfc Dec 05 09:05:30 crc kubenswrapper[4997]: I1205 09:05:30.261719 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-749sz"] Dec 05 09:05:30 crc kubenswrapper[4997]: I1205 09:05:30.822776 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 05 09:05:30 crc kubenswrapper[4997]: I1205 09:05:30.825118 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 05 09:05:31 crc kubenswrapper[4997]: I1205 09:05:31.070361 4997 generic.go:334] "Generic (PLEG): container finished" podID="0a70d008-7f96-498c-a5d6-5c22cf77a41a" containerID="6563c0d63909645e56a064f79f8aab3497f52630bc1f2a3b282e4513dc484d72" exitCode=0 Dec 05 09:05:31 crc kubenswrapper[4997]: I1205 09:05:31.070420 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-749sz" event={"ID":"0a70d008-7f96-498c-a5d6-5c22cf77a41a","Type":"ContainerDied","Data":"6563c0d63909645e56a064f79f8aab3497f52630bc1f2a3b282e4513dc484d72"} Dec 05 09:05:31 crc kubenswrapper[4997]: I1205 09:05:31.070487 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-749sz" event={"ID":"0a70d008-7f96-498c-a5d6-5c22cf77a41a","Type":"ContainerStarted","Data":"546be805880af13f8c716ae9a2825fcc3c40f83b1e80958cbfa0cbbd8531bcfc"} Dec 05 09:05:31 crc kubenswrapper[4997]: I1205 09:05:31.072655 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51b4105e-c02f-40ac-a542-c2b86d0c6f64","Type":"ContainerStarted","Data":"d0fe7082efc594a2fca09dacffe9a414f09967b43053e8c39fe5cc45ae6b7fa5"} Dec 05 09:05:31 crc kubenswrapper[4997]: I1205 09:05:31.075264 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 05 09:05:32 crc kubenswrapper[4997]: I1205 09:05:32.085098 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-749sz" event={"ID":"0a70d008-7f96-498c-a5d6-5c22cf77a41a","Type":"ContainerStarted","Data":"1fe36e3ba888fed6671eb40dd04623cc2a8f77927cc26aa3db9ed9a687486ee3"} Dec 05 09:05:32 crc kubenswrapper[4997]: I1205 09:05:32.087793 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51b4105e-c02f-40ac-a542-c2b86d0c6f64","Type":"ContainerStarted","Data":"a2a444682897824fb132feaa47224a0610221db6e4e23eef4a3050cd7bc654a9"} Dec 05 09:05:32 crc kubenswrapper[4997]: I1205 09:05:32.125569 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.065735101 podStartE2EDuration="10.125549625s" podCreationTimestamp="2025-12-05 09:05:22 +0000 UTC" firstStartedPulling="2025-12-05 09:05:23.230935916 +0000 UTC m=+7823.759843177" lastFinishedPulling="2025-12-05 09:05:31.29075044 +0000 UTC m=+7831.819657701" observedRunningTime="2025-12-05 09:05:32.124490297 +0000 UTC m=+7832.653397698" watchObservedRunningTime="2025-12-05 09:05:32.125549625 +0000 UTC m=+7832.654456886" Dec 05 09:05:33 crc kubenswrapper[4997]: I1205 09:05:33.100037 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4997]: I1205 09:05:34.111670 4997 generic.go:334] "Generic (PLEG): container finished" podID="0a70d008-7f96-498c-a5d6-5c22cf77a41a" containerID="1fe36e3ba888fed6671eb40dd04623cc2a8f77927cc26aa3db9ed9a687486ee3" exitCode=0 Dec 05 09:05:34 crc kubenswrapper[4997]: I1205 09:05:34.111814 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-749sz" event={"ID":"0a70d008-7f96-498c-a5d6-5c22cf77a41a","Type":"ContainerDied","Data":"1fe36e3ba888fed6671eb40dd04623cc2a8f77927cc26aa3db9ed9a687486ee3"} Dec 05 09:05:38 crc kubenswrapper[4997]: I1205 09:05:38.472240 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-749sz" event={"ID":"0a70d008-7f96-498c-a5d6-5c22cf77a41a","Type":"ContainerStarted","Data":"7ba2e2196342258d475bbc4a06f57bddf679ae966948857c033353e6611460d5"} Dec 05 09:05:38 crc kubenswrapper[4997]: I1205 09:05:38.500405 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-749sz" podStartSLOduration=3.026795612 podStartE2EDuration="9.500382362s" podCreationTimestamp="2025-12-05 09:05:29 +0000 UTC" firstStartedPulling="2025-12-05 09:05:31.072814962 +0000 UTC m=+7831.601722223" lastFinishedPulling="2025-12-05 09:05:37.546401712 +0000 UTC m=+7838.075308973" observedRunningTime="2025-12-05 09:05:38.492103789 +0000 UTC m=+7839.021011080" watchObservedRunningTime="2025-12-05 09:05:38.500382362 +0000 UTC m=+7839.029289623" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.160370 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-jrxqn"] Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.161941 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jrxqn" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.185943 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-jrxqn"] Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.269049 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-bcea-account-create-update-8vzh5"] Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.270675 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-bcea-account-create-update-8vzh5" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.274329 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.292104 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-bcea-account-create-update-8vzh5"] Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.298893 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hflzb\" (UniqueName: \"kubernetes.io/projected/5d315ff5-173e-4469-856c-74ece8f15669-kube-api-access-hflzb\") pod \"aodh-db-create-jrxqn\" (UID: \"5d315ff5-173e-4469-856c-74ece8f15669\") " pod="openstack/aodh-db-create-jrxqn" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.299026 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d315ff5-173e-4469-856c-74ece8f15669-operator-scripts\") pod \"aodh-db-create-jrxqn\" (UID: \"5d315ff5-173e-4469-856c-74ece8f15669\") " pod="openstack/aodh-db-create-jrxqn" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.401848 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0036e14d-14f9-4631-903c-8682e6babada-operator-scripts\") pod \"aodh-bcea-account-create-update-8vzh5\" (UID: \"0036e14d-14f9-4631-903c-8682e6babada\") " pod="openstack/aodh-bcea-account-create-update-8vzh5" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.401938 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hflzb\" (UniqueName: \"kubernetes.io/projected/5d315ff5-173e-4469-856c-74ece8f15669-kube-api-access-hflzb\") pod \"aodh-db-create-jrxqn\" (UID: \"5d315ff5-173e-4469-856c-74ece8f15669\") " pod="openstack/aodh-db-create-jrxqn" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.401992 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r4hl\" (UniqueName: \"kubernetes.io/projected/0036e14d-14f9-4631-903c-8682e6babada-kube-api-access-9r4hl\") pod \"aodh-bcea-account-create-update-8vzh5\" (UID: \"0036e14d-14f9-4631-903c-8682e6babada\") " pod="openstack/aodh-bcea-account-create-update-8vzh5" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.402073 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d315ff5-173e-4469-856c-74ece8f15669-operator-scripts\") pod \"aodh-db-create-jrxqn\" (UID: \"5d315ff5-173e-4469-856c-74ece8f15669\") " pod="openstack/aodh-db-create-jrxqn" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.403093 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d315ff5-173e-4469-856c-74ece8f15669-operator-scripts\") pod \"aodh-db-create-jrxqn\" (UID: \"5d315ff5-173e-4469-856c-74ece8f15669\") " pod="openstack/aodh-db-create-jrxqn" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.431729 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hflzb\" (UniqueName: \"kubernetes.io/projected/5d315ff5-173e-4469-856c-74ece8f15669-kube-api-access-hflzb\") pod \"aodh-db-create-jrxqn\" (UID: \"5d315ff5-173e-4469-856c-74ece8f15669\") " pod="openstack/aodh-db-create-jrxqn" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.481406 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jrxqn" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.503626 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0036e14d-14f9-4631-903c-8682e6babada-operator-scripts\") pod \"aodh-bcea-account-create-update-8vzh5\" (UID: \"0036e14d-14f9-4631-903c-8682e6babada\") " pod="openstack/aodh-bcea-account-create-update-8vzh5" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.503693 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r4hl\" (UniqueName: \"kubernetes.io/projected/0036e14d-14f9-4631-903c-8682e6babada-kube-api-access-9r4hl\") pod \"aodh-bcea-account-create-update-8vzh5\" (UID: \"0036e14d-14f9-4631-903c-8682e6babada\") " pod="openstack/aodh-bcea-account-create-update-8vzh5" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.504479 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0036e14d-14f9-4631-903c-8682e6babada-operator-scripts\") pod \"aodh-bcea-account-create-update-8vzh5\" (UID: \"0036e14d-14f9-4631-903c-8682e6babada\") " pod="openstack/aodh-bcea-account-create-update-8vzh5" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.521607 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r4hl\" (UniqueName: \"kubernetes.io/projected/0036e14d-14f9-4631-903c-8682e6babada-kube-api-access-9r4hl\") pod \"aodh-bcea-account-create-update-8vzh5\" (UID: \"0036e14d-14f9-4631-903c-8682e6babada\") " pod="openstack/aodh-bcea-account-create-update-8vzh5" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.590472 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-bcea-account-create-update-8vzh5" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.604222 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-749sz" Dec 05 09:05:39 crc kubenswrapper[4997]: I1205 09:05:39.605483 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-749sz" Dec 05 09:05:40 crc kubenswrapper[4997]: I1205 09:05:40.074138 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-jrxqn"] Dec 05 09:05:40 crc kubenswrapper[4997]: W1205 09:05:40.081763 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d315ff5_173e_4469_856c_74ece8f15669.slice/crio-c256b2661fa339a4223853b3a5e6aa7e9c4be631bad5866632b94a6bb16b0c2a WatchSource:0}: Error finding container c256b2661fa339a4223853b3a5e6aa7e9c4be631bad5866632b94a6bb16b0c2a: Status 404 returned error can't find the container with id c256b2661fa339a4223853b3a5e6aa7e9c4be631bad5866632b94a6bb16b0c2a Dec 05 09:05:40 crc kubenswrapper[4997]: I1205 09:05:40.272786 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-bcea-account-create-update-8vzh5"] Dec 05 09:05:40 crc kubenswrapper[4997]: W1205 09:05:40.281036 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0036e14d_14f9_4631_903c_8682e6babada.slice/crio-e429fb532005e822682570de4212f65db138fadf8123c4327212c5c0f5f3205a WatchSource:0}: Error finding container e429fb532005e822682570de4212f65db138fadf8123c4327212c5c0f5f3205a: Status 404 returned error can't find the container with id e429fb532005e822682570de4212f65db138fadf8123c4327212c5c0f5f3205a Dec 05 09:05:40 crc kubenswrapper[4997]: I1205 09:05:40.511641 4997 generic.go:334] "Generic (PLEG): container finished" podID="5d315ff5-173e-4469-856c-74ece8f15669" containerID="644de32d915ea6906569388efab86698cefe1488837295c1ca10f7dedb7a5c18" exitCode=0 Dec 05 09:05:40 crc kubenswrapper[4997]: I1205 09:05:40.511720 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-jrxqn" event={"ID":"5d315ff5-173e-4469-856c-74ece8f15669","Type":"ContainerDied","Data":"644de32d915ea6906569388efab86698cefe1488837295c1ca10f7dedb7a5c18"} Dec 05 09:05:40 crc kubenswrapper[4997]: I1205 09:05:40.512008 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-jrxqn" event={"ID":"5d315ff5-173e-4469-856c-74ece8f15669","Type":"ContainerStarted","Data":"c256b2661fa339a4223853b3a5e6aa7e9c4be631bad5866632b94a6bb16b0c2a"} Dec 05 09:05:40 crc kubenswrapper[4997]: I1205 09:05:40.514739 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-bcea-account-create-update-8vzh5" event={"ID":"0036e14d-14f9-4631-903c-8682e6babada","Type":"ContainerStarted","Data":"3597ec59688cc35de0fa3dfb30c5c65b2b793c4c827be3d2df03aa8f54dc9500"} Dec 05 09:05:40 crc kubenswrapper[4997]: I1205 09:05:40.514773 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-bcea-account-create-update-8vzh5" event={"ID":"0036e14d-14f9-4631-903c-8682e6babada","Type":"ContainerStarted","Data":"e429fb532005e822682570de4212f65db138fadf8123c4327212c5c0f5f3205a"} Dec 05 09:05:40 crc kubenswrapper[4997]: I1205 09:05:40.555449 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-bcea-account-create-update-8vzh5" podStartSLOduration=1.55542524 podStartE2EDuration="1.55542524s" podCreationTimestamp="2025-12-05 09:05:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:05:40.544360661 +0000 UTC m=+7841.073267932" watchObservedRunningTime="2025-12-05 09:05:40.55542524 +0000 UTC m=+7841.084332501" Dec 05 09:05:40 crc kubenswrapper[4997]: I1205 09:05:40.759408 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-749sz" podUID="0a70d008-7f96-498c-a5d6-5c22cf77a41a" containerName="registry-server" probeResult="failure" output=< Dec 05 09:05:40 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 09:05:40 crc kubenswrapper[4997]: > Dec 05 09:05:41 crc kubenswrapper[4997]: I1205 09:05:41.526014 4997 generic.go:334] "Generic (PLEG): container finished" podID="0036e14d-14f9-4631-903c-8682e6babada" containerID="3597ec59688cc35de0fa3dfb30c5c65b2b793c4c827be3d2df03aa8f54dc9500" exitCode=0 Dec 05 09:05:41 crc kubenswrapper[4997]: I1205 09:05:41.526078 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-bcea-account-create-update-8vzh5" event={"ID":"0036e14d-14f9-4631-903c-8682e6babada","Type":"ContainerDied","Data":"3597ec59688cc35de0fa3dfb30c5c65b2b793c4c827be3d2df03aa8f54dc9500"} Dec 05 09:05:41 crc kubenswrapper[4997]: I1205 09:05:41.921427 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jrxqn" Dec 05 09:05:42 crc kubenswrapper[4997]: I1205 09:05:42.068401 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hflzb\" (UniqueName: \"kubernetes.io/projected/5d315ff5-173e-4469-856c-74ece8f15669-kube-api-access-hflzb\") pod \"5d315ff5-173e-4469-856c-74ece8f15669\" (UID: \"5d315ff5-173e-4469-856c-74ece8f15669\") " Dec 05 09:05:42 crc kubenswrapper[4997]: I1205 09:05:42.068523 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d315ff5-173e-4469-856c-74ece8f15669-operator-scripts\") pod \"5d315ff5-173e-4469-856c-74ece8f15669\" (UID: \"5d315ff5-173e-4469-856c-74ece8f15669\") " Dec 05 09:05:42 crc kubenswrapper[4997]: I1205 09:05:42.069511 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d315ff5-173e-4469-856c-74ece8f15669-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5d315ff5-173e-4469-856c-74ece8f15669" (UID: "5d315ff5-173e-4469-856c-74ece8f15669"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:05:42 crc kubenswrapper[4997]: I1205 09:05:42.078074 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d315ff5-173e-4469-856c-74ece8f15669-kube-api-access-hflzb" (OuterVolumeSpecName: "kube-api-access-hflzb") pod "5d315ff5-173e-4469-856c-74ece8f15669" (UID: "5d315ff5-173e-4469-856c-74ece8f15669"). InnerVolumeSpecName "kube-api-access-hflzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:05:42 crc kubenswrapper[4997]: I1205 09:05:42.172535 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d315ff5-173e-4469-856c-74ece8f15669-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:42 crc kubenswrapper[4997]: I1205 09:05:42.173571 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hflzb\" (UniqueName: \"kubernetes.io/projected/5d315ff5-173e-4469-856c-74ece8f15669-kube-api-access-hflzb\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:42 crc kubenswrapper[4997]: I1205 09:05:42.538068 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jrxqn" Dec 05 09:05:42 crc kubenswrapper[4997]: I1205 09:05:42.538080 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-jrxqn" event={"ID":"5d315ff5-173e-4469-856c-74ece8f15669","Type":"ContainerDied","Data":"c256b2661fa339a4223853b3a5e6aa7e9c4be631bad5866632b94a6bb16b0c2a"} Dec 05 09:05:42 crc kubenswrapper[4997]: I1205 09:05:42.538132 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c256b2661fa339a4223853b3a5e6aa7e9c4be631bad5866632b94a6bb16b0c2a" Dec 05 09:05:42 crc kubenswrapper[4997]: I1205 09:05:42.902512 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-bcea-account-create-update-8vzh5" Dec 05 09:05:42 crc kubenswrapper[4997]: I1205 09:05:42.990641 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9r4hl\" (UniqueName: \"kubernetes.io/projected/0036e14d-14f9-4631-903c-8682e6babada-kube-api-access-9r4hl\") pod \"0036e14d-14f9-4631-903c-8682e6babada\" (UID: \"0036e14d-14f9-4631-903c-8682e6babada\") " Dec 05 09:05:42 crc kubenswrapper[4997]: I1205 09:05:42.990722 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0036e14d-14f9-4631-903c-8682e6babada-operator-scripts\") pod \"0036e14d-14f9-4631-903c-8682e6babada\" (UID: \"0036e14d-14f9-4631-903c-8682e6babada\") " Dec 05 09:05:42 crc kubenswrapper[4997]: I1205 09:05:42.991575 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0036e14d-14f9-4631-903c-8682e6babada-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0036e14d-14f9-4631-903c-8682e6babada" (UID: "0036e14d-14f9-4631-903c-8682e6babada"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:05:42 crc kubenswrapper[4997]: I1205 09:05:42.995107 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0036e14d-14f9-4631-903c-8682e6babada-kube-api-access-9r4hl" (OuterVolumeSpecName: "kube-api-access-9r4hl") pod "0036e14d-14f9-4631-903c-8682e6babada" (UID: "0036e14d-14f9-4631-903c-8682e6babada"). InnerVolumeSpecName "kube-api-access-9r4hl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:05:43 crc kubenswrapper[4997]: I1205 09:05:43.092683 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9r4hl\" (UniqueName: \"kubernetes.io/projected/0036e14d-14f9-4631-903c-8682e6babada-kube-api-access-9r4hl\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:43 crc kubenswrapper[4997]: I1205 09:05:43.092723 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0036e14d-14f9-4631-903c-8682e6babada-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:43 crc kubenswrapper[4997]: I1205 09:05:43.548074 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-bcea-account-create-update-8vzh5" event={"ID":"0036e14d-14f9-4631-903c-8682e6babada","Type":"ContainerDied","Data":"e429fb532005e822682570de4212f65db138fadf8123c4327212c5c0f5f3205a"} Dec 05 09:05:43 crc kubenswrapper[4997]: I1205 09:05:43.548811 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e429fb532005e822682570de4212f65db138fadf8123c4327212c5c0f5f3205a" Dec 05 09:05:43 crc kubenswrapper[4997]: I1205 09:05:43.548120 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-bcea-account-create-update-8vzh5" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.598651 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-5cpz9"] Dec 05 09:05:44 crc kubenswrapper[4997]: E1205 09:05:44.599196 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d315ff5-173e-4469-856c-74ece8f15669" containerName="mariadb-database-create" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.599230 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d315ff5-173e-4469-856c-74ece8f15669" containerName="mariadb-database-create" Dec 05 09:05:44 crc kubenswrapper[4997]: E1205 09:05:44.599248 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0036e14d-14f9-4631-903c-8682e6babada" containerName="mariadb-account-create-update" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.599257 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0036e14d-14f9-4631-903c-8682e6babada" containerName="mariadb-account-create-update" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.599517 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d315ff5-173e-4469-856c-74ece8f15669" containerName="mariadb-database-create" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.599551 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0036e14d-14f9-4631-903c-8682e6babada" containerName="mariadb-account-create-update" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.600441 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-5cpz9" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.604767 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-dt4w8" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.605120 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.605206 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.605277 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.618518 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-5cpz9"] Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.750396 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-combined-ca-bundle\") pod \"aodh-db-sync-5cpz9\" (UID: \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\") " pod="openstack/aodh-db-sync-5cpz9" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.750734 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-config-data\") pod \"aodh-db-sync-5cpz9\" (UID: \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\") " pod="openstack/aodh-db-sync-5cpz9" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.750911 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9n27\" (UniqueName: \"kubernetes.io/projected/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-kube-api-access-v9n27\") pod \"aodh-db-sync-5cpz9\" (UID: \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\") " pod="openstack/aodh-db-sync-5cpz9" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.751141 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-scripts\") pod \"aodh-db-sync-5cpz9\" (UID: \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\") " pod="openstack/aodh-db-sync-5cpz9" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.853391 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-combined-ca-bundle\") pod \"aodh-db-sync-5cpz9\" (UID: \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\") " pod="openstack/aodh-db-sync-5cpz9" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.853449 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-config-data\") pod \"aodh-db-sync-5cpz9\" (UID: \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\") " pod="openstack/aodh-db-sync-5cpz9" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.853553 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9n27\" (UniqueName: \"kubernetes.io/projected/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-kube-api-access-v9n27\") pod \"aodh-db-sync-5cpz9\" (UID: \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\") " pod="openstack/aodh-db-sync-5cpz9" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.855387 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-scripts\") pod \"aodh-db-sync-5cpz9\" (UID: \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\") " pod="openstack/aodh-db-sync-5cpz9" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.858085 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-scripts\") pod \"aodh-db-sync-5cpz9\" (UID: \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\") " pod="openstack/aodh-db-sync-5cpz9" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.858712 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-combined-ca-bundle\") pod \"aodh-db-sync-5cpz9\" (UID: \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\") " pod="openstack/aodh-db-sync-5cpz9" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.858815 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-config-data\") pod \"aodh-db-sync-5cpz9\" (UID: \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\") " pod="openstack/aodh-db-sync-5cpz9" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.881084 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9n27\" (UniqueName: \"kubernetes.io/projected/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-kube-api-access-v9n27\") pod \"aodh-db-sync-5cpz9\" (UID: \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\") " pod="openstack/aodh-db-sync-5cpz9" Dec 05 09:05:44 crc kubenswrapper[4997]: I1205 09:05:44.922189 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-5cpz9" Dec 05 09:05:45 crc kubenswrapper[4997]: I1205 09:05:45.434996 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-5cpz9"] Dec 05 09:05:45 crc kubenswrapper[4997]: I1205 09:05:45.569249 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-5cpz9" event={"ID":"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9","Type":"ContainerStarted","Data":"69268e2fbcc5e65e4decd97a8e67106b38d1087c3f6ec16ed6b90f58003efbe2"} Dec 05 09:05:49 crc kubenswrapper[4997]: I1205 09:05:49.794165 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-749sz" Dec 05 09:05:49 crc kubenswrapper[4997]: I1205 09:05:49.847299 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-749sz" Dec 05 09:05:50 crc kubenswrapper[4997]: I1205 09:05:50.041370 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-749sz"] Dec 05 09:05:51 crc kubenswrapper[4997]: I1205 09:05:51.646244 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-5cpz9" event={"ID":"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9","Type":"ContainerStarted","Data":"6bfc5691dd87a09f74025bf1c7061f4c87d8119250649b4f0313022fbf2d1cca"} Dec 05 09:05:51 crc kubenswrapper[4997]: I1205 09:05:51.646728 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-749sz" podUID="0a70d008-7f96-498c-a5d6-5c22cf77a41a" containerName="registry-server" containerID="cri-o://7ba2e2196342258d475bbc4a06f57bddf679ae966948857c033353e6611460d5" gracePeriod=2 Dec 05 09:05:51 crc kubenswrapper[4997]: I1205 09:05:51.665685 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-5cpz9" podStartSLOduration=2.524325868 podStartE2EDuration="7.665665046s" podCreationTimestamp="2025-12-05 09:05:44 +0000 UTC" firstStartedPulling="2025-12-05 09:05:45.447069593 +0000 UTC m=+7845.975976864" lastFinishedPulling="2025-12-05 09:05:50.588408781 +0000 UTC m=+7851.117316042" observedRunningTime="2025-12-05 09:05:51.664755122 +0000 UTC m=+7852.193662403" watchObservedRunningTime="2025-12-05 09:05:51.665665046 +0000 UTC m=+7852.194572307" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.173317 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-749sz" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.373738 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a70d008-7f96-498c-a5d6-5c22cf77a41a-utilities\") pod \"0a70d008-7f96-498c-a5d6-5c22cf77a41a\" (UID: \"0a70d008-7f96-498c-a5d6-5c22cf77a41a\") " Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.373842 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nnnr\" (UniqueName: \"kubernetes.io/projected/0a70d008-7f96-498c-a5d6-5c22cf77a41a-kube-api-access-6nnnr\") pod \"0a70d008-7f96-498c-a5d6-5c22cf77a41a\" (UID: \"0a70d008-7f96-498c-a5d6-5c22cf77a41a\") " Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.373978 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a70d008-7f96-498c-a5d6-5c22cf77a41a-catalog-content\") pod \"0a70d008-7f96-498c-a5d6-5c22cf77a41a\" (UID: \"0a70d008-7f96-498c-a5d6-5c22cf77a41a\") " Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.376303 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a70d008-7f96-498c-a5d6-5c22cf77a41a-utilities" (OuterVolumeSpecName: "utilities") pod "0a70d008-7f96-498c-a5d6-5c22cf77a41a" (UID: "0a70d008-7f96-498c-a5d6-5c22cf77a41a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.392315 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a70d008-7f96-498c-a5d6-5c22cf77a41a-kube-api-access-6nnnr" (OuterVolumeSpecName: "kube-api-access-6nnnr") pod "0a70d008-7f96-498c-a5d6-5c22cf77a41a" (UID: "0a70d008-7f96-498c-a5d6-5c22cf77a41a"). InnerVolumeSpecName "kube-api-access-6nnnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.445182 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a70d008-7f96-498c-a5d6-5c22cf77a41a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a70d008-7f96-498c-a5d6-5c22cf77a41a" (UID: "0a70d008-7f96-498c-a5d6-5c22cf77a41a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.476236 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a70d008-7f96-498c-a5d6-5c22cf77a41a-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.476275 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nnnr\" (UniqueName: \"kubernetes.io/projected/0a70d008-7f96-498c-a5d6-5c22cf77a41a-kube-api-access-6nnnr\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.476285 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a70d008-7f96-498c-a5d6-5c22cf77a41a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.658890 4997 generic.go:334] "Generic (PLEG): container finished" podID="0a70d008-7f96-498c-a5d6-5c22cf77a41a" containerID="7ba2e2196342258d475bbc4a06f57bddf679ae966948857c033353e6611460d5" exitCode=0 Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.658951 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-749sz" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.658967 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-749sz" event={"ID":"0a70d008-7f96-498c-a5d6-5c22cf77a41a","Type":"ContainerDied","Data":"7ba2e2196342258d475bbc4a06f57bddf679ae966948857c033353e6611460d5"} Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.659342 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-749sz" event={"ID":"0a70d008-7f96-498c-a5d6-5c22cf77a41a","Type":"ContainerDied","Data":"546be805880af13f8c716ae9a2825fcc3c40f83b1e80958cbfa0cbbd8531bcfc"} Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.659383 4997 scope.go:117] "RemoveContainer" containerID="7ba2e2196342258d475bbc4a06f57bddf679ae966948857c033353e6611460d5" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.688094 4997 scope.go:117] "RemoveContainer" containerID="1fe36e3ba888fed6671eb40dd04623cc2a8f77927cc26aa3db9ed9a687486ee3" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.692703 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-749sz"] Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.699233 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.702847 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-749sz"] Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.732596 4997 scope.go:117] "RemoveContainer" containerID="6563c0d63909645e56a064f79f8aab3497f52630bc1f2a3b282e4513dc484d72" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.765884 4997 scope.go:117] "RemoveContainer" containerID="7ba2e2196342258d475bbc4a06f57bddf679ae966948857c033353e6611460d5" Dec 05 09:05:52 crc kubenswrapper[4997]: E1205 09:05:52.766750 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ba2e2196342258d475bbc4a06f57bddf679ae966948857c033353e6611460d5\": container with ID starting with 7ba2e2196342258d475bbc4a06f57bddf679ae966948857c033353e6611460d5 not found: ID does not exist" containerID="7ba2e2196342258d475bbc4a06f57bddf679ae966948857c033353e6611460d5" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.766796 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ba2e2196342258d475bbc4a06f57bddf679ae966948857c033353e6611460d5"} err="failed to get container status \"7ba2e2196342258d475bbc4a06f57bddf679ae966948857c033353e6611460d5\": rpc error: code = NotFound desc = could not find container \"7ba2e2196342258d475bbc4a06f57bddf679ae966948857c033353e6611460d5\": container with ID starting with 7ba2e2196342258d475bbc4a06f57bddf679ae966948857c033353e6611460d5 not found: ID does not exist" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.766834 4997 scope.go:117] "RemoveContainer" containerID="1fe36e3ba888fed6671eb40dd04623cc2a8f77927cc26aa3db9ed9a687486ee3" Dec 05 09:05:52 crc kubenswrapper[4997]: E1205 09:05:52.767191 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fe36e3ba888fed6671eb40dd04623cc2a8f77927cc26aa3db9ed9a687486ee3\": container with ID starting with 1fe36e3ba888fed6671eb40dd04623cc2a8f77927cc26aa3db9ed9a687486ee3 not found: ID does not exist" containerID="1fe36e3ba888fed6671eb40dd04623cc2a8f77927cc26aa3db9ed9a687486ee3" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.767233 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fe36e3ba888fed6671eb40dd04623cc2a8f77927cc26aa3db9ed9a687486ee3"} err="failed to get container status \"1fe36e3ba888fed6671eb40dd04623cc2a8f77927cc26aa3db9ed9a687486ee3\": rpc error: code = NotFound desc = could not find container \"1fe36e3ba888fed6671eb40dd04623cc2a8f77927cc26aa3db9ed9a687486ee3\": container with ID starting with 1fe36e3ba888fed6671eb40dd04623cc2a8f77927cc26aa3db9ed9a687486ee3 not found: ID does not exist" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.767260 4997 scope.go:117] "RemoveContainer" containerID="6563c0d63909645e56a064f79f8aab3497f52630bc1f2a3b282e4513dc484d72" Dec 05 09:05:52 crc kubenswrapper[4997]: E1205 09:05:52.767566 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6563c0d63909645e56a064f79f8aab3497f52630bc1f2a3b282e4513dc484d72\": container with ID starting with 6563c0d63909645e56a064f79f8aab3497f52630bc1f2a3b282e4513dc484d72 not found: ID does not exist" containerID="6563c0d63909645e56a064f79f8aab3497f52630bc1f2a3b282e4513dc484d72" Dec 05 09:05:52 crc kubenswrapper[4997]: I1205 09:05:52.767595 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6563c0d63909645e56a064f79f8aab3497f52630bc1f2a3b282e4513dc484d72"} err="failed to get container status \"6563c0d63909645e56a064f79f8aab3497f52630bc1f2a3b282e4513dc484d72\": rpc error: code = NotFound desc = could not find container \"6563c0d63909645e56a064f79f8aab3497f52630bc1f2a3b282e4513dc484d72\": container with ID starting with 6563c0d63909645e56a064f79f8aab3497f52630bc1f2a3b282e4513dc484d72 not found: ID does not exist" Dec 05 09:05:53 crc kubenswrapper[4997]: I1205 09:05:53.671413 4997 generic.go:334] "Generic (PLEG): container finished" podID="b2519e54-7c30-40f0-a1eb-f576ae2f6fa9" containerID="6bfc5691dd87a09f74025bf1c7061f4c87d8119250649b4f0313022fbf2d1cca" exitCode=0 Dec 05 09:05:53 crc kubenswrapper[4997]: I1205 09:05:53.671513 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-5cpz9" event={"ID":"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9","Type":"ContainerDied","Data":"6bfc5691dd87a09f74025bf1c7061f4c87d8119250649b4f0313022fbf2d1cca"} Dec 05 09:05:53 crc kubenswrapper[4997]: I1205 09:05:53.763140 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a70d008-7f96-498c-a5d6-5c22cf77a41a" path="/var/lib/kubelet/pods/0a70d008-7f96-498c-a5d6-5c22cf77a41a/volumes" Dec 05 09:05:54 crc kubenswrapper[4997]: I1205 09:05:54.037741 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-3d5f-account-create-update-v4cs5"] Dec 05 09:05:54 crc kubenswrapper[4997]: I1205 09:05:54.047016 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-3d5f-account-create-update-v4cs5"] Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.031782 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-vb5ts"] Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.046640 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-vb5ts"] Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.047313 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-5cpz9" Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.231049 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-combined-ca-bundle\") pod \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\" (UID: \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\") " Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.231183 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-scripts\") pod \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\" (UID: \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\") " Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.231325 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-config-data\") pod \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\" (UID: \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\") " Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.231356 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9n27\" (UniqueName: \"kubernetes.io/projected/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-kube-api-access-v9n27\") pod \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\" (UID: \"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9\") " Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.236446 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-scripts" (OuterVolumeSpecName: "scripts") pod "b2519e54-7c30-40f0-a1eb-f576ae2f6fa9" (UID: "b2519e54-7c30-40f0-a1eb-f576ae2f6fa9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.236682 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-kube-api-access-v9n27" (OuterVolumeSpecName: "kube-api-access-v9n27") pod "b2519e54-7c30-40f0-a1eb-f576ae2f6fa9" (UID: "b2519e54-7c30-40f0-a1eb-f576ae2f6fa9"). InnerVolumeSpecName "kube-api-access-v9n27". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.257218 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-config-data" (OuterVolumeSpecName: "config-data") pod "b2519e54-7c30-40f0-a1eb-f576ae2f6fa9" (UID: "b2519e54-7c30-40f0-a1eb-f576ae2f6fa9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.261839 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2519e54-7c30-40f0-a1eb-f576ae2f6fa9" (UID: "b2519e54-7c30-40f0-a1eb-f576ae2f6fa9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.333920 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.334211 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9n27\" (UniqueName: \"kubernetes.io/projected/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-kube-api-access-v9n27\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.334221 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.334232 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.698904 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-5cpz9" event={"ID":"b2519e54-7c30-40f0-a1eb-f576ae2f6fa9","Type":"ContainerDied","Data":"69268e2fbcc5e65e4decd97a8e67106b38d1087c3f6ec16ed6b90f58003efbe2"} Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.698952 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69268e2fbcc5e65e4decd97a8e67106b38d1087c3f6ec16ed6b90f58003efbe2" Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.698959 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-5cpz9" Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.760106 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ccc1288-f7d9-4963-8efa-1e2392030be5" path="/var/lib/kubelet/pods/4ccc1288-f7d9-4963-8efa-1e2392030be5/volumes" Dec 05 09:05:55 crc kubenswrapper[4997]: I1205 09:05:55.760784 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e40fa138-3d33-42be-b39e-6b8c1533323b" path="/var/lib/kubelet/pods/e40fa138-3d33-42be-b39e-6b8c1533323b/volumes" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.233396 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 05 09:05:59 crc kubenswrapper[4997]: E1205 09:05:59.234103 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a70d008-7f96-498c-a5d6-5c22cf77a41a" containerName="extract-utilities" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.234320 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a70d008-7f96-498c-a5d6-5c22cf77a41a" containerName="extract-utilities" Dec 05 09:05:59 crc kubenswrapper[4997]: E1205 09:05:59.234334 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2519e54-7c30-40f0-a1eb-f576ae2f6fa9" containerName="aodh-db-sync" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.234340 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2519e54-7c30-40f0-a1eb-f576ae2f6fa9" containerName="aodh-db-sync" Dec 05 09:05:59 crc kubenswrapper[4997]: E1205 09:05:59.234374 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a70d008-7f96-498c-a5d6-5c22cf77a41a" containerName="registry-server" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.234380 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a70d008-7f96-498c-a5d6-5c22cf77a41a" containerName="registry-server" Dec 05 09:05:59 crc kubenswrapper[4997]: E1205 09:05:59.234404 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a70d008-7f96-498c-a5d6-5c22cf77a41a" containerName="extract-content" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.234410 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a70d008-7f96-498c-a5d6-5c22cf77a41a" containerName="extract-content" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.234593 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2519e54-7c30-40f0-a1eb-f576ae2f6fa9" containerName="aodh-db-sync" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.234630 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a70d008-7f96-498c-a5d6-5c22cf77a41a" containerName="registry-server" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.237728 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.239742 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.245220 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.245221 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-dt4w8" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.251373 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.410569 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43128621-bd06-4369-8b56-eda49b797194-config-data\") pod \"aodh-0\" (UID: \"43128621-bd06-4369-8b56-eda49b797194\") " pod="openstack/aodh-0" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.410641 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43128621-bd06-4369-8b56-eda49b797194-scripts\") pod \"aodh-0\" (UID: \"43128621-bd06-4369-8b56-eda49b797194\") " pod="openstack/aodh-0" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.410867 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xqrb\" (UniqueName: \"kubernetes.io/projected/43128621-bd06-4369-8b56-eda49b797194-kube-api-access-2xqrb\") pod \"aodh-0\" (UID: \"43128621-bd06-4369-8b56-eda49b797194\") " pod="openstack/aodh-0" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.411011 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43128621-bd06-4369-8b56-eda49b797194-combined-ca-bundle\") pod \"aodh-0\" (UID: \"43128621-bd06-4369-8b56-eda49b797194\") " pod="openstack/aodh-0" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.512822 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43128621-bd06-4369-8b56-eda49b797194-combined-ca-bundle\") pod \"aodh-0\" (UID: \"43128621-bd06-4369-8b56-eda49b797194\") " pod="openstack/aodh-0" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.512947 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43128621-bd06-4369-8b56-eda49b797194-config-data\") pod \"aodh-0\" (UID: \"43128621-bd06-4369-8b56-eda49b797194\") " pod="openstack/aodh-0" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.512979 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43128621-bd06-4369-8b56-eda49b797194-scripts\") pod \"aodh-0\" (UID: \"43128621-bd06-4369-8b56-eda49b797194\") " pod="openstack/aodh-0" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.513030 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xqrb\" (UniqueName: \"kubernetes.io/projected/43128621-bd06-4369-8b56-eda49b797194-kube-api-access-2xqrb\") pod \"aodh-0\" (UID: \"43128621-bd06-4369-8b56-eda49b797194\") " pod="openstack/aodh-0" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.518032 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43128621-bd06-4369-8b56-eda49b797194-scripts\") pod \"aodh-0\" (UID: \"43128621-bd06-4369-8b56-eda49b797194\") " pod="openstack/aodh-0" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.518716 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43128621-bd06-4369-8b56-eda49b797194-config-data\") pod \"aodh-0\" (UID: \"43128621-bd06-4369-8b56-eda49b797194\") " pod="openstack/aodh-0" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.519265 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43128621-bd06-4369-8b56-eda49b797194-combined-ca-bundle\") pod \"aodh-0\" (UID: \"43128621-bd06-4369-8b56-eda49b797194\") " pod="openstack/aodh-0" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.531101 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xqrb\" (UniqueName: \"kubernetes.io/projected/43128621-bd06-4369-8b56-eda49b797194-kube-api-access-2xqrb\") pod \"aodh-0\" (UID: \"43128621-bd06-4369-8b56-eda49b797194\") " pod="openstack/aodh-0" Dec 05 09:05:59 crc kubenswrapper[4997]: I1205 09:05:59.562463 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:06:00 crc kubenswrapper[4997]: W1205 09:06:00.098010 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43128621_bd06_4369_8b56_eda49b797194.slice/crio-caf8c399ff1ba39d4bd9e6e1adf95cc103f945773c7fcf430e1a128f15d9d087 WatchSource:0}: Error finding container caf8c399ff1ba39d4bd9e6e1adf95cc103f945773c7fcf430e1a128f15d9d087: Status 404 returned error can't find the container with id caf8c399ff1ba39d4bd9e6e1adf95cc103f945773c7fcf430e1a128f15d9d087 Dec 05 09:06:00 crc kubenswrapper[4997]: I1205 09:06:00.109063 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 05 09:06:00 crc kubenswrapper[4997]: I1205 09:06:00.749326 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"43128621-bd06-4369-8b56-eda49b797194","Type":"ContainerStarted","Data":"41bb978a0913a5344a680e9430c73260952acbada932b0cc57b7c2338aa18963"} Dec 05 09:06:00 crc kubenswrapper[4997]: I1205 09:06:00.749692 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"43128621-bd06-4369-8b56-eda49b797194","Type":"ContainerStarted","Data":"caf8c399ff1ba39d4bd9e6e1adf95cc103f945773c7fcf430e1a128f15d9d087"} Dec 05 09:06:01 crc kubenswrapper[4997]: I1205 09:06:01.719976 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:06:01 crc kubenswrapper[4997]: I1205 09:06:01.720895 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerName="ceilometer-central-agent" containerID="cri-o://be7869d770ed2907f9c09b160fc08e7e9b16d2c13a7cb6d244461029f02fa7c9" gracePeriod=30 Dec 05 09:06:01 crc kubenswrapper[4997]: I1205 09:06:01.720955 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerName="proxy-httpd" containerID="cri-o://a2a444682897824fb132feaa47224a0610221db6e4e23eef4a3050cd7bc654a9" gracePeriod=30 Dec 05 09:06:01 crc kubenswrapper[4997]: I1205 09:06:01.720983 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerName="sg-core" containerID="cri-o://d0fe7082efc594a2fca09dacffe9a414f09967b43053e8c39fe5cc45ae6b7fa5" gracePeriod=30 Dec 05 09:06:01 crc kubenswrapper[4997]: I1205 09:06:01.720981 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerName="ceilometer-notification-agent" containerID="cri-o://dc3a9ec7865fd5c8d34148c68fdd3c653709a6669deb39b13eebe6e530600fe8" gracePeriod=30 Dec 05 09:06:01 crc kubenswrapper[4997]: I1205 09:06:01.761700 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"43128621-bd06-4369-8b56-eda49b797194","Type":"ContainerStarted","Data":"44393d4e2d1716ab243c8e5323e2a55dfd4de51736eb02879c05cae458359731"} Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.773794 4997 generic.go:334] "Generic (PLEG): container finished" podID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerID="a2a444682897824fb132feaa47224a0610221db6e4e23eef4a3050cd7bc654a9" exitCode=0 Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.774212 4997 generic.go:334] "Generic (PLEG): container finished" podID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerID="d0fe7082efc594a2fca09dacffe9a414f09967b43053e8c39fe5cc45ae6b7fa5" exitCode=2 Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.774220 4997 generic.go:334] "Generic (PLEG): container finished" podID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerID="dc3a9ec7865fd5c8d34148c68fdd3c653709a6669deb39b13eebe6e530600fe8" exitCode=0 Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.774227 4997 generic.go:334] "Generic (PLEG): container finished" podID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerID="be7869d770ed2907f9c09b160fc08e7e9b16d2c13a7cb6d244461029f02fa7c9" exitCode=0 Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.773859 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51b4105e-c02f-40ac-a542-c2b86d0c6f64","Type":"ContainerDied","Data":"a2a444682897824fb132feaa47224a0610221db6e4e23eef4a3050cd7bc654a9"} Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.774261 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51b4105e-c02f-40ac-a542-c2b86d0c6f64","Type":"ContainerDied","Data":"d0fe7082efc594a2fca09dacffe9a414f09967b43053e8c39fe5cc45ae6b7fa5"} Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.774276 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51b4105e-c02f-40ac-a542-c2b86d0c6f64","Type":"ContainerDied","Data":"dc3a9ec7865fd5c8d34148c68fdd3c653709a6669deb39b13eebe6e530600fe8"} Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.774287 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51b4105e-c02f-40ac-a542-c2b86d0c6f64","Type":"ContainerDied","Data":"be7869d770ed2907f9c09b160fc08e7e9b16d2c13a7cb6d244461029f02fa7c9"} Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.774296 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"51b4105e-c02f-40ac-a542-c2b86d0c6f64","Type":"ContainerDied","Data":"b16599faab9555e9afb9b63d673b991dbba3a329ec538d59e13fff22a2398c0c"} Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.774306 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b16599faab9555e9afb9b63d673b991dbba3a329ec538d59e13fff22a2398c0c" Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.854203 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.991210 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-scripts\") pod \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.991430 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51b4105e-c02f-40ac-a542-c2b86d0c6f64-run-httpd\") pod \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.991482 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-combined-ca-bundle\") pod \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.991522 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-config-data\") pod \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.991580 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kjnz\" (UniqueName: \"kubernetes.io/projected/51b4105e-c02f-40ac-a542-c2b86d0c6f64-kube-api-access-9kjnz\") pod \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.991764 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51b4105e-c02f-40ac-a542-c2b86d0c6f64-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "51b4105e-c02f-40ac-a542-c2b86d0c6f64" (UID: "51b4105e-c02f-40ac-a542-c2b86d0c6f64"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.991805 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-sg-core-conf-yaml\") pod \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.991923 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51b4105e-c02f-40ac-a542-c2b86d0c6f64-log-httpd\") pod \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\" (UID: \"51b4105e-c02f-40ac-a542-c2b86d0c6f64\") " Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.992253 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51b4105e-c02f-40ac-a542-c2b86d0c6f64-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "51b4105e-c02f-40ac-a542-c2b86d0c6f64" (UID: "51b4105e-c02f-40ac-a542-c2b86d0c6f64"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.996964 4997 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51b4105e-c02f-40ac-a542-c2b86d0c6f64-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.997004 4997 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51b4105e-c02f-40ac-a542-c2b86d0c6f64-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.999776 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-scripts" (OuterVolumeSpecName: "scripts") pod "51b4105e-c02f-40ac-a542-c2b86d0c6f64" (UID: "51b4105e-c02f-40ac-a542-c2b86d0c6f64"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:06:02 crc kubenswrapper[4997]: I1205 09:06:02.999955 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51b4105e-c02f-40ac-a542-c2b86d0c6f64-kube-api-access-9kjnz" (OuterVolumeSpecName: "kube-api-access-9kjnz") pod "51b4105e-c02f-40ac-a542-c2b86d0c6f64" (UID: "51b4105e-c02f-40ac-a542-c2b86d0c6f64"). InnerVolumeSpecName "kube-api-access-9kjnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.031854 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "51b4105e-c02f-40ac-a542-c2b86d0c6f64" (UID: "51b4105e-c02f-40ac-a542-c2b86d0c6f64"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.075030 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51b4105e-c02f-40ac-a542-c2b86d0c6f64" (UID: "51b4105e-c02f-40ac-a542-c2b86d0c6f64"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.099182 4997 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.099211 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.099223 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.099235 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kjnz\" (UniqueName: \"kubernetes.io/projected/51b4105e-c02f-40ac-a542-c2b86d0c6f64-kube-api-access-9kjnz\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.116013 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-config-data" (OuterVolumeSpecName: "config-data") pod "51b4105e-c02f-40ac-a542-c2b86d0c6f64" (UID: "51b4105e-c02f-40ac-a542-c2b86d0c6f64"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.201320 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b4105e-c02f-40ac-a542-c2b86d0c6f64-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.786985 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.787811 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"43128621-bd06-4369-8b56-eda49b797194","Type":"ContainerStarted","Data":"5fbc71902a96dca351327fb345424d699d88cfd23cfdcd8ab8f53f25690b0818"} Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.846327 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.862871 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.874570 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:06:03 crc kubenswrapper[4997]: E1205 09:06:03.875063 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerName="proxy-httpd" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.875083 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerName="proxy-httpd" Dec 05 09:06:03 crc kubenswrapper[4997]: E1205 09:06:03.875103 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerName="ceilometer-central-agent" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.875110 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerName="ceilometer-central-agent" Dec 05 09:06:03 crc kubenswrapper[4997]: E1205 09:06:03.875117 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerName="ceilometer-notification-agent" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.875123 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerName="ceilometer-notification-agent" Dec 05 09:06:03 crc kubenswrapper[4997]: E1205 09:06:03.875142 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerName="sg-core" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.875148 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerName="sg-core" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.875315 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerName="ceilometer-central-agent" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.875335 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerName="proxy-httpd" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.875345 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerName="ceilometer-notification-agent" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.875360 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" containerName="sg-core" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.877142 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.879703 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.879702 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 09:06:03 crc kubenswrapper[4997]: I1205 09:06:03.906681 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.029156 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-config-data\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.029213 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8347c13-a145-4306-a8a9-f3f8b617f296-log-httpd\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.029266 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hrk7\" (UniqueName: \"kubernetes.io/projected/e8347c13-a145-4306-a8a9-f3f8b617f296-kube-api-access-5hrk7\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.029296 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-scripts\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.029315 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.029337 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8347c13-a145-4306-a8a9-f3f8b617f296-run-httpd\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.029415 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.131345 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.131476 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-config-data\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.131525 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8347c13-a145-4306-a8a9-f3f8b617f296-log-httpd\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.131575 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hrk7\" (UniqueName: \"kubernetes.io/projected/e8347c13-a145-4306-a8a9-f3f8b617f296-kube-api-access-5hrk7\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.131634 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-scripts\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.131662 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.131694 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8347c13-a145-4306-a8a9-f3f8b617f296-run-httpd\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.132332 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8347c13-a145-4306-a8a9-f3f8b617f296-run-httpd\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.133038 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8347c13-a145-4306-a8a9-f3f8b617f296-log-httpd\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.135593 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.136367 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-scripts\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.137107 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-config-data\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.141733 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.149429 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hrk7\" (UniqueName: \"kubernetes.io/projected/e8347c13-a145-4306-a8a9-f3f8b617f296-kube-api-access-5hrk7\") pod \"ceilometer-0\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.313829 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.754349 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:06:04 crc kubenswrapper[4997]: W1205 09:06:04.755734 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8347c13_a145_4306_a8a9_f3f8b617f296.slice/crio-1812c75822b87a74d8547661db0fd405e03871b8baa42ef4f6b0d176880a014e WatchSource:0}: Error finding container 1812c75822b87a74d8547661db0fd405e03871b8baa42ef4f6b0d176880a014e: Status 404 returned error can't find the container with id 1812c75822b87a74d8547661db0fd405e03871b8baa42ef4f6b0d176880a014e Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.796496 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"43128621-bd06-4369-8b56-eda49b797194","Type":"ContainerStarted","Data":"039a17feb9c09e3e472d6d0160cff4513b6c13804336350ce115f32a49d042aa"} Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.799725 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8347c13-a145-4306-a8a9-f3f8b617f296","Type":"ContainerStarted","Data":"1812c75822b87a74d8547661db0fd405e03871b8baa42ef4f6b0d176880a014e"} Dec 05 09:06:04 crc kubenswrapper[4997]: I1205 09:06:04.822301 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.073683403 podStartE2EDuration="5.822278827s" podCreationTimestamp="2025-12-05 09:05:59 +0000 UTC" firstStartedPulling="2025-12-05 09:06:00.101905423 +0000 UTC m=+7860.630812684" lastFinishedPulling="2025-12-05 09:06:03.850500837 +0000 UTC m=+7864.379408108" observedRunningTime="2025-12-05 09:06:04.814509928 +0000 UTC m=+7865.343417209" watchObservedRunningTime="2025-12-05 09:06:04.822278827 +0000 UTC m=+7865.351186088" Dec 05 09:06:05 crc kubenswrapper[4997]: I1205 09:06:05.775114 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51b4105e-c02f-40ac-a542-c2b86d0c6f64" path="/var/lib/kubelet/pods/51b4105e-c02f-40ac-a542-c2b86d0c6f64/volumes" Dec 05 09:06:05 crc kubenswrapper[4997]: I1205 09:06:05.809643 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8347c13-a145-4306-a8a9-f3f8b617f296","Type":"ContainerStarted","Data":"d24348b52188e11c84b19c6cfdfbfb63dea98de53d6c177b725d62e9d7ed9f7e"} Dec 05 09:06:06 crc kubenswrapper[4997]: I1205 09:06:06.820021 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8347c13-a145-4306-a8a9-f3f8b617f296","Type":"ContainerStarted","Data":"04735f873c32d25d40773cecd9f768d0e9a5fd47916f5f7eca2190758a9242ed"} Dec 05 09:06:07 crc kubenswrapper[4997]: I1205 09:06:07.831485 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8347c13-a145-4306-a8a9-f3f8b617f296","Type":"ContainerStarted","Data":"e4f498aab2f391e7425ca69f4a11c646ae365266d2a061b73155d932a0057242"} Dec 05 09:06:08 crc kubenswrapper[4997]: I1205 09:06:08.846684 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8347c13-a145-4306-a8a9-f3f8b617f296","Type":"ContainerStarted","Data":"b6b88db011a8f01238fadbf8d88ceb77c8c62ac98f34b74fb1ff1efb0539c561"} Dec 05 09:06:08 crc kubenswrapper[4997]: I1205 09:06:08.849362 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 09:06:08 crc kubenswrapper[4997]: I1205 09:06:08.881362 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.086614187 podStartE2EDuration="5.881340995s" podCreationTimestamp="2025-12-05 09:06:03 +0000 UTC" firstStartedPulling="2025-12-05 09:06:04.760112591 +0000 UTC m=+7865.289019892" lastFinishedPulling="2025-12-05 09:06:08.554839439 +0000 UTC m=+7869.083746700" observedRunningTime="2025-12-05 09:06:08.873548006 +0000 UTC m=+7869.402455317" watchObservedRunningTime="2025-12-05 09:06:08.881340995 +0000 UTC m=+7869.410248276" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.031989 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-lsxlp"] Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.034379 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-lsxlp" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.050741 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp845\" (UniqueName: \"kubernetes.io/projected/ae61f102-3277-41ea-8c08-086bd0225419-kube-api-access-sp845\") pod \"manila-db-create-lsxlp\" (UID: \"ae61f102-3277-41ea-8c08-086bd0225419\") " pod="openstack/manila-db-create-lsxlp" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.050804 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae61f102-3277-41ea-8c08-086bd0225419-operator-scripts\") pod \"manila-db-create-lsxlp\" (UID: \"ae61f102-3277-41ea-8c08-086bd0225419\") " pod="openstack/manila-db-create-lsxlp" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.087854 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-lsxlp"] Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.137511 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-9580-account-create-update-6cdhz"] Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.139987 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9580-account-create-update-6cdhz" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.145887 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.147158 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-9580-account-create-update-6cdhz"] Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.152571 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp845\" (UniqueName: \"kubernetes.io/projected/ae61f102-3277-41ea-8c08-086bd0225419-kube-api-access-sp845\") pod \"manila-db-create-lsxlp\" (UID: \"ae61f102-3277-41ea-8c08-086bd0225419\") " pod="openstack/manila-db-create-lsxlp" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.152661 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae61f102-3277-41ea-8c08-086bd0225419-operator-scripts\") pod \"manila-db-create-lsxlp\" (UID: \"ae61f102-3277-41ea-8c08-086bd0225419\") " pod="openstack/manila-db-create-lsxlp" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.153409 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae61f102-3277-41ea-8c08-086bd0225419-operator-scripts\") pod \"manila-db-create-lsxlp\" (UID: \"ae61f102-3277-41ea-8c08-086bd0225419\") " pod="openstack/manila-db-create-lsxlp" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.180826 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp845\" (UniqueName: \"kubernetes.io/projected/ae61f102-3277-41ea-8c08-086bd0225419-kube-api-access-sp845\") pod \"manila-db-create-lsxlp\" (UID: \"ae61f102-3277-41ea-8c08-086bd0225419\") " pod="openstack/manila-db-create-lsxlp" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.254821 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vjbp\" (UniqueName: \"kubernetes.io/projected/535f8966-92df-4881-9771-c5dfc02bf193-kube-api-access-2vjbp\") pod \"manila-9580-account-create-update-6cdhz\" (UID: \"535f8966-92df-4881-9771-c5dfc02bf193\") " pod="openstack/manila-9580-account-create-update-6cdhz" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.254929 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/535f8966-92df-4881-9771-c5dfc02bf193-operator-scripts\") pod \"manila-9580-account-create-update-6cdhz\" (UID: \"535f8966-92df-4881-9771-c5dfc02bf193\") " pod="openstack/manila-9580-account-create-update-6cdhz" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.351436 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-lsxlp" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.357063 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vjbp\" (UniqueName: \"kubernetes.io/projected/535f8966-92df-4881-9771-c5dfc02bf193-kube-api-access-2vjbp\") pod \"manila-9580-account-create-update-6cdhz\" (UID: \"535f8966-92df-4881-9771-c5dfc02bf193\") " pod="openstack/manila-9580-account-create-update-6cdhz" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.357193 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/535f8966-92df-4881-9771-c5dfc02bf193-operator-scripts\") pod \"manila-9580-account-create-update-6cdhz\" (UID: \"535f8966-92df-4881-9771-c5dfc02bf193\") " pod="openstack/manila-9580-account-create-update-6cdhz" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.357992 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/535f8966-92df-4881-9771-c5dfc02bf193-operator-scripts\") pod \"manila-9580-account-create-update-6cdhz\" (UID: \"535f8966-92df-4881-9771-c5dfc02bf193\") " pod="openstack/manila-9580-account-create-update-6cdhz" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.393459 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vjbp\" (UniqueName: \"kubernetes.io/projected/535f8966-92df-4881-9771-c5dfc02bf193-kube-api-access-2vjbp\") pod \"manila-9580-account-create-update-6cdhz\" (UID: \"535f8966-92df-4881-9771-c5dfc02bf193\") " pod="openstack/manila-9580-account-create-update-6cdhz" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.472154 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9580-account-create-update-6cdhz" Dec 05 09:06:10 crc kubenswrapper[4997]: I1205 09:06:10.979088 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-lsxlp"] Dec 05 09:06:10 crc kubenswrapper[4997]: W1205 09:06:10.979200 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae61f102_3277_41ea_8c08_086bd0225419.slice/crio-87e032edde78a8e409de6a02a462e3422f04bf4ff891f8c570763c6651cb35f3 WatchSource:0}: Error finding container 87e032edde78a8e409de6a02a462e3422f04bf4ff891f8c570763c6651cb35f3: Status 404 returned error can't find the container with id 87e032edde78a8e409de6a02a462e3422f04bf4ff891f8c570763c6651cb35f3 Dec 05 09:06:11 crc kubenswrapper[4997]: I1205 09:06:11.095165 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-9580-account-create-update-6cdhz"] Dec 05 09:06:11 crc kubenswrapper[4997]: W1205 09:06:11.102847 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod535f8966_92df_4881_9771_c5dfc02bf193.slice/crio-3bea8c4e25c983ac339c74918bac20c684e5c12241aa3d8898c9e33271ccb720 WatchSource:0}: Error finding container 3bea8c4e25c983ac339c74918bac20c684e5c12241aa3d8898c9e33271ccb720: Status 404 returned error can't find the container with id 3bea8c4e25c983ac339c74918bac20c684e5c12241aa3d8898c9e33271ccb720 Dec 05 09:06:11 crc kubenswrapper[4997]: I1205 09:06:11.895333 4997 generic.go:334] "Generic (PLEG): container finished" podID="535f8966-92df-4881-9771-c5dfc02bf193" containerID="6877c786e03b895e087df0c7a627e2791e8fe2ee547c29e3c80826e467e7b60f" exitCode=0 Dec 05 09:06:11 crc kubenswrapper[4997]: I1205 09:06:11.895442 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-9580-account-create-update-6cdhz" event={"ID":"535f8966-92df-4881-9771-c5dfc02bf193","Type":"ContainerDied","Data":"6877c786e03b895e087df0c7a627e2791e8fe2ee547c29e3c80826e467e7b60f"} Dec 05 09:06:11 crc kubenswrapper[4997]: I1205 09:06:11.895745 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-9580-account-create-update-6cdhz" event={"ID":"535f8966-92df-4881-9771-c5dfc02bf193","Type":"ContainerStarted","Data":"3bea8c4e25c983ac339c74918bac20c684e5c12241aa3d8898c9e33271ccb720"} Dec 05 09:06:11 crc kubenswrapper[4997]: I1205 09:06:11.898085 4997 generic.go:334] "Generic (PLEG): container finished" podID="ae61f102-3277-41ea-8c08-086bd0225419" containerID="e6e03db2d9a8ebc1ddf973faa2a742ae658c2c5b445c85b578b67999c85bce05" exitCode=0 Dec 05 09:06:11 crc kubenswrapper[4997]: I1205 09:06:11.898116 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-lsxlp" event={"ID":"ae61f102-3277-41ea-8c08-086bd0225419","Type":"ContainerDied","Data":"e6e03db2d9a8ebc1ddf973faa2a742ae658c2c5b445c85b578b67999c85bce05"} Dec 05 09:06:11 crc kubenswrapper[4997]: I1205 09:06:11.898131 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-lsxlp" event={"ID":"ae61f102-3277-41ea-8c08-086bd0225419","Type":"ContainerStarted","Data":"87e032edde78a8e409de6a02a462e3422f04bf4ff891f8c570763c6651cb35f3"} Dec 05 09:06:12 crc kubenswrapper[4997]: I1205 09:06:12.044096 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-7959t"] Dec 05 09:06:12 crc kubenswrapper[4997]: I1205 09:06:12.055227 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-7959t"] Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.425902 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9580-account-create-update-6cdhz" Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.433964 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-lsxlp" Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.536206 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/535f8966-92df-4881-9771-c5dfc02bf193-operator-scripts\") pod \"535f8966-92df-4881-9771-c5dfc02bf193\" (UID: \"535f8966-92df-4881-9771-c5dfc02bf193\") " Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.536961 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp845\" (UniqueName: \"kubernetes.io/projected/ae61f102-3277-41ea-8c08-086bd0225419-kube-api-access-sp845\") pod \"ae61f102-3277-41ea-8c08-086bd0225419\" (UID: \"ae61f102-3277-41ea-8c08-086bd0225419\") " Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.537023 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vjbp\" (UniqueName: \"kubernetes.io/projected/535f8966-92df-4881-9771-c5dfc02bf193-kube-api-access-2vjbp\") pod \"535f8966-92df-4881-9771-c5dfc02bf193\" (UID: \"535f8966-92df-4881-9771-c5dfc02bf193\") " Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.536958 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/535f8966-92df-4881-9771-c5dfc02bf193-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "535f8966-92df-4881-9771-c5dfc02bf193" (UID: "535f8966-92df-4881-9771-c5dfc02bf193"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.537133 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae61f102-3277-41ea-8c08-086bd0225419-operator-scripts\") pod \"ae61f102-3277-41ea-8c08-086bd0225419\" (UID: \"ae61f102-3277-41ea-8c08-086bd0225419\") " Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.537476 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/535f8966-92df-4881-9771-c5dfc02bf193-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.537711 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae61f102-3277-41ea-8c08-086bd0225419-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ae61f102-3277-41ea-8c08-086bd0225419" (UID: "ae61f102-3277-41ea-8c08-086bd0225419"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.542525 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/535f8966-92df-4881-9771-c5dfc02bf193-kube-api-access-2vjbp" (OuterVolumeSpecName: "kube-api-access-2vjbp") pod "535f8966-92df-4881-9771-c5dfc02bf193" (UID: "535f8966-92df-4881-9771-c5dfc02bf193"). InnerVolumeSpecName "kube-api-access-2vjbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.545205 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae61f102-3277-41ea-8c08-086bd0225419-kube-api-access-sp845" (OuterVolumeSpecName: "kube-api-access-sp845") pod "ae61f102-3277-41ea-8c08-086bd0225419" (UID: "ae61f102-3277-41ea-8c08-086bd0225419"). InnerVolumeSpecName "kube-api-access-sp845". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.638991 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp845\" (UniqueName: \"kubernetes.io/projected/ae61f102-3277-41ea-8c08-086bd0225419-kube-api-access-sp845\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.639020 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vjbp\" (UniqueName: \"kubernetes.io/projected/535f8966-92df-4881-9771-c5dfc02bf193-kube-api-access-2vjbp\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.639031 4997 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae61f102-3277-41ea-8c08-086bd0225419-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.762502 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924" path="/var/lib/kubelet/pods/0fbdaeae-8ae1-4bea-9c38-3dc4e6fb7924/volumes" Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.919442 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-lsxlp" event={"ID":"ae61f102-3277-41ea-8c08-086bd0225419","Type":"ContainerDied","Data":"87e032edde78a8e409de6a02a462e3422f04bf4ff891f8c570763c6651cb35f3"} Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.919489 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87e032edde78a8e409de6a02a462e3422f04bf4ff891f8c570763c6651cb35f3" Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.919547 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-lsxlp" Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.922365 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-9580-account-create-update-6cdhz" event={"ID":"535f8966-92df-4881-9771-c5dfc02bf193","Type":"ContainerDied","Data":"3bea8c4e25c983ac339c74918bac20c684e5c12241aa3d8898c9e33271ccb720"} Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.922397 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bea8c4e25c983ac339c74918bac20c684e5c12241aa3d8898c9e33271ccb720" Dec 05 09:06:13 crc kubenswrapper[4997]: I1205 09:06:13.922444 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9580-account-create-update-6cdhz" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.554509 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-nlscc"] Dec 05 09:06:15 crc kubenswrapper[4997]: E1205 09:06:15.556317 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="535f8966-92df-4881-9771-c5dfc02bf193" containerName="mariadb-account-create-update" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.556338 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="535f8966-92df-4881-9771-c5dfc02bf193" containerName="mariadb-account-create-update" Dec 05 09:06:15 crc kubenswrapper[4997]: E1205 09:06:15.556377 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae61f102-3277-41ea-8c08-086bd0225419" containerName="mariadb-database-create" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.556386 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae61f102-3277-41ea-8c08-086bd0225419" containerName="mariadb-database-create" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.556658 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae61f102-3277-41ea-8c08-086bd0225419" containerName="mariadb-database-create" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.556679 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="535f8966-92df-4881-9771-c5dfc02bf193" containerName="mariadb-account-create-update" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.557575 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-nlscc" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.560083 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.560093 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-5pqng" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.566221 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-nlscc"] Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.678768 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-job-config-data\") pod \"manila-db-sync-nlscc\" (UID: \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\") " pod="openstack/manila-db-sync-nlscc" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.678845 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk6q7\" (UniqueName: \"kubernetes.io/projected/6c864cae-1e34-46f0-8f85-c83c8963a6bc-kube-api-access-wk6q7\") pod \"manila-db-sync-nlscc\" (UID: \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\") " pod="openstack/manila-db-sync-nlscc" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.679824 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-config-data\") pod \"manila-db-sync-nlscc\" (UID: \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\") " pod="openstack/manila-db-sync-nlscc" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.679958 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-combined-ca-bundle\") pod \"manila-db-sync-nlscc\" (UID: \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\") " pod="openstack/manila-db-sync-nlscc" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.782444 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-config-data\") pod \"manila-db-sync-nlscc\" (UID: \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\") " pod="openstack/manila-db-sync-nlscc" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.782521 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-combined-ca-bundle\") pod \"manila-db-sync-nlscc\" (UID: \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\") " pod="openstack/manila-db-sync-nlscc" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.782717 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-job-config-data\") pod \"manila-db-sync-nlscc\" (UID: \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\") " pod="openstack/manila-db-sync-nlscc" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.782791 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk6q7\" (UniqueName: \"kubernetes.io/projected/6c864cae-1e34-46f0-8f85-c83c8963a6bc-kube-api-access-wk6q7\") pod \"manila-db-sync-nlscc\" (UID: \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\") " pod="openstack/manila-db-sync-nlscc" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.787158 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-job-config-data\") pod \"manila-db-sync-nlscc\" (UID: \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\") " pod="openstack/manila-db-sync-nlscc" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.787892 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-combined-ca-bundle\") pod \"manila-db-sync-nlscc\" (UID: \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\") " pod="openstack/manila-db-sync-nlscc" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.788972 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-config-data\") pod \"manila-db-sync-nlscc\" (UID: \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\") " pod="openstack/manila-db-sync-nlscc" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.799415 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk6q7\" (UniqueName: \"kubernetes.io/projected/6c864cae-1e34-46f0-8f85-c83c8963a6bc-kube-api-access-wk6q7\") pod \"manila-db-sync-nlscc\" (UID: \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\") " pod="openstack/manila-db-sync-nlscc" Dec 05 09:06:15 crc kubenswrapper[4997]: I1205 09:06:15.877255 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-nlscc" Dec 05 09:06:16 crc kubenswrapper[4997]: I1205 09:06:16.575485 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-nlscc"] Dec 05 09:06:16 crc kubenswrapper[4997]: W1205 09:06:16.576461 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c864cae_1e34_46f0_8f85_c83c8963a6bc.slice/crio-fb99cf971f2dbaab56559a06834d3520caa4a3ac1cc9753406235e2722621223 WatchSource:0}: Error finding container fb99cf971f2dbaab56559a06834d3520caa4a3ac1cc9753406235e2722621223: Status 404 returned error can't find the container with id fb99cf971f2dbaab56559a06834d3520caa4a3ac1cc9753406235e2722621223 Dec 05 09:06:16 crc kubenswrapper[4997]: I1205 09:06:16.963309 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-nlscc" event={"ID":"6c864cae-1e34-46f0-8f85-c83c8963a6bc","Type":"ContainerStarted","Data":"fb99cf971f2dbaab56559a06834d3520caa4a3ac1cc9753406235e2722621223"} Dec 05 09:06:17 crc kubenswrapper[4997]: I1205 09:06:17.641625 4997 scope.go:117] "RemoveContainer" containerID="bd5e7d9c0313564c3df38c8b5a3e5f433d878996a240eda09e395ae2b708f35c" Dec 05 09:06:17 crc kubenswrapper[4997]: I1205 09:06:17.682394 4997 scope.go:117] "RemoveContainer" containerID="52186b60dbc9147c4a35f0872c0d85a34dda95d6728765468f540cf913486964" Dec 05 09:06:17 crc kubenswrapper[4997]: I1205 09:06:17.725023 4997 scope.go:117] "RemoveContainer" containerID="dfd4f3305734811a291c363bcfb98fae4e5dc21b8cac4081c8f81957ac03ec08" Dec 05 09:06:17 crc kubenswrapper[4997]: I1205 09:06:17.770440 4997 scope.go:117] "RemoveContainer" containerID="968e6edc631cecf7c42ee56b31588268f8277c6db305828b4e3f368c78a688d2" Dec 05 09:06:19 crc kubenswrapper[4997]: I1205 09:06:19.770074 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:06:19 crc kubenswrapper[4997]: I1205 09:06:19.770582 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:06:22 crc kubenswrapper[4997]: I1205 09:06:22.032603 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-nlscc" event={"ID":"6c864cae-1e34-46f0-8f85-c83c8963a6bc","Type":"ContainerStarted","Data":"4e03ef7591ecc13bc654b364b37b1e8b4da87249ab0b97600dc99b42895c3b90"} Dec 05 09:06:22 crc kubenswrapper[4997]: I1205 09:06:22.053676 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-nlscc" podStartSLOduration=2.787820614 podStartE2EDuration="7.053657479s" podCreationTimestamp="2025-12-05 09:06:15 +0000 UTC" firstStartedPulling="2025-12-05 09:06:16.579204736 +0000 UTC m=+7877.108111997" lastFinishedPulling="2025-12-05 09:06:20.845041561 +0000 UTC m=+7881.373948862" observedRunningTime="2025-12-05 09:06:22.053253738 +0000 UTC m=+7882.582161019" watchObservedRunningTime="2025-12-05 09:06:22.053657479 +0000 UTC m=+7882.582564750" Dec 05 09:06:24 crc kubenswrapper[4997]: I1205 09:06:24.060404 4997 generic.go:334] "Generic (PLEG): container finished" podID="6c864cae-1e34-46f0-8f85-c83c8963a6bc" containerID="4e03ef7591ecc13bc654b364b37b1e8b4da87249ab0b97600dc99b42895c3b90" exitCode=0 Dec 05 09:06:24 crc kubenswrapper[4997]: I1205 09:06:24.061011 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-nlscc" event={"ID":"6c864cae-1e34-46f0-8f85-c83c8963a6bc","Type":"ContainerDied","Data":"4e03ef7591ecc13bc654b364b37b1e8b4da87249ab0b97600dc99b42895c3b90"} Dec 05 09:06:25 crc kubenswrapper[4997]: I1205 09:06:25.614693 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-nlscc" Dec 05 09:06:25 crc kubenswrapper[4997]: I1205 09:06:25.796142 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-config-data\") pod \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\" (UID: \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\") " Dec 05 09:06:25 crc kubenswrapper[4997]: I1205 09:06:25.796257 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wk6q7\" (UniqueName: \"kubernetes.io/projected/6c864cae-1e34-46f0-8f85-c83c8963a6bc-kube-api-access-wk6q7\") pod \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\" (UID: \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\") " Dec 05 09:06:25 crc kubenswrapper[4997]: I1205 09:06:25.796688 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-combined-ca-bundle\") pod \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\" (UID: \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\") " Dec 05 09:06:25 crc kubenswrapper[4997]: I1205 09:06:25.796733 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-job-config-data\") pod \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\" (UID: \"6c864cae-1e34-46f0-8f85-c83c8963a6bc\") " Dec 05 09:06:25 crc kubenswrapper[4997]: I1205 09:06:25.804650 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c864cae-1e34-46f0-8f85-c83c8963a6bc-kube-api-access-wk6q7" (OuterVolumeSpecName: "kube-api-access-wk6q7") pod "6c864cae-1e34-46f0-8f85-c83c8963a6bc" (UID: "6c864cae-1e34-46f0-8f85-c83c8963a6bc"). InnerVolumeSpecName "kube-api-access-wk6q7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:06:25 crc kubenswrapper[4997]: I1205 09:06:25.805821 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "6c864cae-1e34-46f0-8f85-c83c8963a6bc" (UID: "6c864cae-1e34-46f0-8f85-c83c8963a6bc"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:06:25 crc kubenswrapper[4997]: I1205 09:06:25.808925 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-config-data" (OuterVolumeSpecName: "config-data") pod "6c864cae-1e34-46f0-8f85-c83c8963a6bc" (UID: "6c864cae-1e34-46f0-8f85-c83c8963a6bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:06:25 crc kubenswrapper[4997]: I1205 09:06:25.832844 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c864cae-1e34-46f0-8f85-c83c8963a6bc" (UID: "6c864cae-1e34-46f0-8f85-c83c8963a6bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:06:25 crc kubenswrapper[4997]: I1205 09:06:25.898844 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wk6q7\" (UniqueName: \"kubernetes.io/projected/6c864cae-1e34-46f0-8f85-c83c8963a6bc-kube-api-access-wk6q7\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:25 crc kubenswrapper[4997]: I1205 09:06:25.898899 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:25 crc kubenswrapper[4997]: I1205 09:06:25.898911 4997 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-job-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:25 crc kubenswrapper[4997]: I1205 09:06:25.898924 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c864cae-1e34-46f0-8f85-c83c8963a6bc-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.079911 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-nlscc" event={"ID":"6c864cae-1e34-46f0-8f85-c83c8963a6bc","Type":"ContainerDied","Data":"fb99cf971f2dbaab56559a06834d3520caa4a3ac1cc9753406235e2722621223"} Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.079969 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb99cf971f2dbaab56559a06834d3520caa4a3ac1cc9753406235e2722621223" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.080048 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-nlscc" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.496688 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 05 09:06:26 crc kubenswrapper[4997]: E1205 09:06:26.497836 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c864cae-1e34-46f0-8f85-c83c8963a6bc" containerName="manila-db-sync" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.497964 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c864cae-1e34-46f0-8f85-c83c8963a6bc" containerName="manila-db-sync" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.498352 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c864cae-1e34-46f0-8f85-c83c8963a6bc" containerName="manila-db-sync" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.500134 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.505929 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.506194 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.506332 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-5pqng" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.506471 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.509797 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbz2g\" (UniqueName: \"kubernetes.io/projected/ed63e934-a1ac-479c-96cc-0100c0869765-kube-api-access-nbz2g\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.509862 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ed63e934-a1ac-479c-96cc-0100c0869765-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.509922 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed63e934-a1ac-479c-96cc-0100c0869765-scripts\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.509942 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed63e934-a1ac-479c-96cc-0100c0869765-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.509971 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed63e934-a1ac-479c-96cc-0100c0869765-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.510231 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed63e934-a1ac-479c-96cc-0100c0869765-config-data\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.542695 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.545161 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.548823 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.555695 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.578091 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.614280 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbz2g\" (UniqueName: \"kubernetes.io/projected/ed63e934-a1ac-479c-96cc-0100c0869765-kube-api-access-nbz2g\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.614345 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/799f7992-40e0-4fea-8e26-84a596b0500f-config-data\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.614373 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/799f7992-40e0-4fea-8e26-84a596b0500f-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.614409 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/799f7992-40e0-4fea-8e26-84a596b0500f-scripts\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.614429 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ed63e934-a1ac-479c-96cc-0100c0869765-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.614448 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/799f7992-40e0-4fea-8e26-84a596b0500f-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.614485 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed63e934-a1ac-479c-96cc-0100c0869765-scripts\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.614504 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed63e934-a1ac-479c-96cc-0100c0869765-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.614527 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed63e934-a1ac-479c-96cc-0100c0869765-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.614562 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4rhh\" (UniqueName: \"kubernetes.io/projected/799f7992-40e0-4fea-8e26-84a596b0500f-kube-api-access-v4rhh\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.614606 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/799f7992-40e0-4fea-8e26-84a596b0500f-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.614747 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/799f7992-40e0-4fea-8e26-84a596b0500f-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.615670 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed63e934-a1ac-479c-96cc-0100c0869765-config-data\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.615930 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ed63e934-a1ac-479c-96cc-0100c0869765-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.615949 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/799f7992-40e0-4fea-8e26-84a596b0500f-ceph\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.642431 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed63e934-a1ac-479c-96cc-0100c0869765-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.642864 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed63e934-a1ac-479c-96cc-0100c0869765-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.648126 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbz2g\" (UniqueName: \"kubernetes.io/projected/ed63e934-a1ac-479c-96cc-0100c0869765-kube-api-access-nbz2g\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.651463 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed63e934-a1ac-479c-96cc-0100c0869765-scripts\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.654133 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed63e934-a1ac-479c-96cc-0100c0869765-config-data\") pod \"manila-scheduler-0\" (UID: \"ed63e934-a1ac-479c-96cc-0100c0869765\") " pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.684558 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c5dcbcb5-vz7hb"] Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.688677 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.718901 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/799f7992-40e0-4fea-8e26-84a596b0500f-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.718971 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/799f7992-40e0-4fea-8e26-84a596b0500f-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.719168 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/799f7992-40e0-4fea-8e26-84a596b0500f-ceph\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.719212 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/799f7992-40e0-4fea-8e26-84a596b0500f-config-data\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.719239 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/799f7992-40e0-4fea-8e26-84a596b0500f-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.719278 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/799f7992-40e0-4fea-8e26-84a596b0500f-scripts\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.719306 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/799f7992-40e0-4fea-8e26-84a596b0500f-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.719371 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4rhh\" (UniqueName: \"kubernetes.io/projected/799f7992-40e0-4fea-8e26-84a596b0500f-kube-api-access-v4rhh\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.719828 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/799f7992-40e0-4fea-8e26-84a596b0500f-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.720873 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/799f7992-40e0-4fea-8e26-84a596b0500f-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.724060 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/799f7992-40e0-4fea-8e26-84a596b0500f-ceph\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.724142 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/799f7992-40e0-4fea-8e26-84a596b0500f-scripts\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.728426 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/799f7992-40e0-4fea-8e26-84a596b0500f-config-data\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.733705 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/799f7992-40e0-4fea-8e26-84a596b0500f-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.735493 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/799f7992-40e0-4fea-8e26-84a596b0500f-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.739443 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c5dcbcb5-vz7hb"] Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.760713 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4rhh\" (UniqueName: \"kubernetes.io/projected/799f7992-40e0-4fea-8e26-84a596b0500f-kube-api-access-v4rhh\") pod \"manila-share-share1-0\" (UID: \"799f7992-40e0-4fea-8e26-84a596b0500f\") " pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.821125 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-ovsdbserver-nb\") pod \"dnsmasq-dns-7c5dcbcb5-vz7hb\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.821166 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-config\") pod \"dnsmasq-dns-7c5dcbcb5-vz7hb\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.821227 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-dns-svc\") pod \"dnsmasq-dns-7c5dcbcb5-vz7hb\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.821262 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-ovsdbserver-sb\") pod \"dnsmasq-dns-7c5dcbcb5-vz7hb\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.821299 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jt54\" (UniqueName: \"kubernetes.io/projected/4913c005-1117-40c8-9916-aecc8625d1ed-kube-api-access-9jt54\") pod \"dnsmasq-dns-7c5dcbcb5-vz7hb\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.834469 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.836698 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.838795 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.841893 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.855978 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.881843 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.923506 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-ovsdbserver-nb\") pod \"dnsmasq-dns-7c5dcbcb5-vz7hb\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.923555 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-config\") pod \"dnsmasq-dns-7c5dcbcb5-vz7hb\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.923724 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-dns-svc\") pod \"dnsmasq-dns-7c5dcbcb5-vz7hb\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.923765 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-ovsdbserver-sb\") pod \"dnsmasq-dns-7c5dcbcb5-vz7hb\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.923805 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jt54\" (UniqueName: \"kubernetes.io/projected/4913c005-1117-40c8-9916-aecc8625d1ed-kube-api-access-9jt54\") pod \"dnsmasq-dns-7c5dcbcb5-vz7hb\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.932922 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-ovsdbserver-nb\") pod \"dnsmasq-dns-7c5dcbcb5-vz7hb\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.933862 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-config\") pod \"dnsmasq-dns-7c5dcbcb5-vz7hb\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.933953 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-dns-svc\") pod \"dnsmasq-dns-7c5dcbcb5-vz7hb\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.935829 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-ovsdbserver-sb\") pod \"dnsmasq-dns-7c5dcbcb5-vz7hb\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:26 crc kubenswrapper[4997]: I1205 09:06:26.947158 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jt54\" (UniqueName: \"kubernetes.io/projected/4913c005-1117-40c8-9916-aecc8625d1ed-kube-api-access-9jt54\") pod \"dnsmasq-dns-7c5dcbcb5-vz7hb\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.029796 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bcc15490-e838-4503-8318-f79aafc8c961-etc-machine-id\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.029840 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bcc15490-e838-4503-8318-f79aafc8c961-scripts\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.029864 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4kbw\" (UniqueName: \"kubernetes.io/projected/bcc15490-e838-4503-8318-f79aafc8c961-kube-api-access-p4kbw\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.029900 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcc15490-e838-4503-8318-f79aafc8c961-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.029943 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcc15490-e838-4503-8318-f79aafc8c961-config-data\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.029958 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bcc15490-e838-4503-8318-f79aafc8c961-config-data-custom\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.030023 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcc15490-e838-4503-8318-f79aafc8c961-logs\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.131606 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bcc15490-e838-4503-8318-f79aafc8c961-etc-machine-id\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.131666 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bcc15490-e838-4503-8318-f79aafc8c961-scripts\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.131692 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4kbw\" (UniqueName: \"kubernetes.io/projected/bcc15490-e838-4503-8318-f79aafc8c961-kube-api-access-p4kbw\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.131728 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcc15490-e838-4503-8318-f79aafc8c961-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.131773 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcc15490-e838-4503-8318-f79aafc8c961-config-data\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.131792 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bcc15490-e838-4503-8318-f79aafc8c961-config-data-custom\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.131864 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcc15490-e838-4503-8318-f79aafc8c961-logs\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.132294 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcc15490-e838-4503-8318-f79aafc8c961-logs\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.132337 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bcc15490-e838-4503-8318-f79aafc8c961-etc-machine-id\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.134150 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.137089 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcc15490-e838-4503-8318-f79aafc8c961-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.141162 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcc15490-e838-4503-8318-f79aafc8c961-config-data\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.145128 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bcc15490-e838-4503-8318-f79aafc8c961-config-data-custom\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.157046 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bcc15490-e838-4503-8318-f79aafc8c961-scripts\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.163893 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4kbw\" (UniqueName: \"kubernetes.io/projected/bcc15490-e838-4503-8318-f79aafc8c961-kube-api-access-p4kbw\") pod \"manila-api-0\" (UID: \"bcc15490-e838-4503-8318-f79aafc8c961\") " pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.457104 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.548663 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 05 09:06:27 crc kubenswrapper[4997]: W1205 09:06:27.679732 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded63e934_a1ac_479c_96cc_0100c0869765.slice/crio-638039202e71b7df864a30121f797312e35dc9e58cc6f68c6ee50294ca4df7a1 WatchSource:0}: Error finding container 638039202e71b7df864a30121f797312e35dc9e58cc6f68c6ee50294ca4df7a1: Status 404 returned error can't find the container with id 638039202e71b7df864a30121f797312e35dc9e58cc6f68c6ee50294ca4df7a1 Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.889506 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 05 09:06:27 crc kubenswrapper[4997]: I1205 09:06:27.962953 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c5dcbcb5-vz7hb"] Dec 05 09:06:28 crc kubenswrapper[4997]: I1205 09:06:28.120652 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" event={"ID":"4913c005-1117-40c8-9916-aecc8625d1ed","Type":"ContainerStarted","Data":"0687d407f5b92b6d235cc5dae7ae8a6d0e36cdaaf77d73b2c52a65a751f72d41"} Dec 05 09:06:28 crc kubenswrapper[4997]: I1205 09:06:28.125490 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"799f7992-40e0-4fea-8e26-84a596b0500f","Type":"ContainerStarted","Data":"33a84e2032cfea3755695bc4739b610e9a91731570f00dbce9d191c289a4ce1a"} Dec 05 09:06:28 crc kubenswrapper[4997]: I1205 09:06:28.130380 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"ed63e934-a1ac-479c-96cc-0100c0869765","Type":"ContainerStarted","Data":"638039202e71b7df864a30121f797312e35dc9e58cc6f68c6ee50294ca4df7a1"} Dec 05 09:06:28 crc kubenswrapper[4997]: W1205 09:06:28.194833 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcc15490_e838_4503_8318_f79aafc8c961.slice/crio-a5f0c6b7d4a26e2d3c967f416ba2a91b8c22c91f073723968529ba3ac0e6f74c WatchSource:0}: Error finding container a5f0c6b7d4a26e2d3c967f416ba2a91b8c22c91f073723968529ba3ac0e6f74c: Status 404 returned error can't find the container with id a5f0c6b7d4a26e2d3c967f416ba2a91b8c22c91f073723968529ba3ac0e6f74c Dec 05 09:06:28 crc kubenswrapper[4997]: I1205 09:06:28.198147 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 05 09:06:28 crc kubenswrapper[4997]: E1205 09:06:28.667868 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4913c005_1117_40c8_9916_aecc8625d1ed.slice/crio-conmon-cd33f76d1fe4114fe360da1f7e8ea95c04569d18649fea06839c27dac260552c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4913c005_1117_40c8_9916_aecc8625d1ed.slice/crio-cd33f76d1fe4114fe360da1f7e8ea95c04569d18649fea06839c27dac260552c.scope\": RecentStats: unable to find data in memory cache]" Dec 05 09:06:29 crc kubenswrapper[4997]: I1205 09:06:29.146636 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"bcc15490-e838-4503-8318-f79aafc8c961","Type":"ContainerStarted","Data":"7ca1f35584215fae8fd7abff39887eeffd05580ebedddbc5de02a24412528b25"} Dec 05 09:06:29 crc kubenswrapper[4997]: I1205 09:06:29.147015 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"bcc15490-e838-4503-8318-f79aafc8c961","Type":"ContainerStarted","Data":"a5f0c6b7d4a26e2d3c967f416ba2a91b8c22c91f073723968529ba3ac0e6f74c"} Dec 05 09:06:29 crc kubenswrapper[4997]: I1205 09:06:29.149703 4997 generic.go:334] "Generic (PLEG): container finished" podID="4913c005-1117-40c8-9916-aecc8625d1ed" containerID="cd33f76d1fe4114fe360da1f7e8ea95c04569d18649fea06839c27dac260552c" exitCode=0 Dec 05 09:06:29 crc kubenswrapper[4997]: I1205 09:06:29.149795 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" event={"ID":"4913c005-1117-40c8-9916-aecc8625d1ed","Type":"ContainerDied","Data":"cd33f76d1fe4114fe360da1f7e8ea95c04569d18649fea06839c27dac260552c"} Dec 05 09:06:29 crc kubenswrapper[4997]: I1205 09:06:29.165288 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"ed63e934-a1ac-479c-96cc-0100c0869765","Type":"ContainerStarted","Data":"551b62e6b998e168b57df5b2dfdd6719cbfc227f675559f10cf1950fea5472a5"} Dec 05 09:06:30 crc kubenswrapper[4997]: I1205 09:06:30.191869 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" event={"ID":"4913c005-1117-40c8-9916-aecc8625d1ed","Type":"ContainerStarted","Data":"fec63ae921673bae9d114a02c2f3d6e257ed483ba0168ccf7b0ad21b73a3254c"} Dec 05 09:06:30 crc kubenswrapper[4997]: I1205 09:06:30.192269 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:30 crc kubenswrapper[4997]: I1205 09:06:30.198388 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"ed63e934-a1ac-479c-96cc-0100c0869765","Type":"ContainerStarted","Data":"e488b8479fa45fb45ee39238618200e4e8309da260d313374df3f894a0a5fb06"} Dec 05 09:06:30 crc kubenswrapper[4997]: I1205 09:06:30.205736 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"bcc15490-e838-4503-8318-f79aafc8c961","Type":"ContainerStarted","Data":"4f3be169e5b5114ced7240ccd82967debf0e34980150e5c9f67b4e6abbc4f12a"} Dec 05 09:06:30 crc kubenswrapper[4997]: I1205 09:06:30.206731 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 05 09:06:30 crc kubenswrapper[4997]: I1205 09:06:30.229247 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" podStartSLOduration=4.229229384 podStartE2EDuration="4.229229384s" podCreationTimestamp="2025-12-05 09:06:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:06:30.222589585 +0000 UTC m=+7890.751496856" watchObservedRunningTime="2025-12-05 09:06:30.229229384 +0000 UTC m=+7890.758136645" Dec 05 09:06:30 crc kubenswrapper[4997]: I1205 09:06:30.261534 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=4.261516475 podStartE2EDuration="4.261516475s" podCreationTimestamp="2025-12-05 09:06:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:06:30.260831867 +0000 UTC m=+7890.789739158" watchObservedRunningTime="2025-12-05 09:06:30.261516475 +0000 UTC m=+7890.790423736" Dec 05 09:06:34 crc kubenswrapper[4997]: I1205 09:06:34.481236 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 09:06:34 crc kubenswrapper[4997]: I1205 09:06:34.526007 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=8.17031635 podStartE2EDuration="8.525972093s" podCreationTimestamp="2025-12-05 09:06:26 +0000 UTC" firstStartedPulling="2025-12-05 09:06:27.685046894 +0000 UTC m=+7888.213954155" lastFinishedPulling="2025-12-05 09:06:28.040702647 +0000 UTC m=+7888.569609898" observedRunningTime="2025-12-05 09:06:30.301886474 +0000 UTC m=+7890.830793745" watchObservedRunningTime="2025-12-05 09:06:34.525972093 +0000 UTC m=+7895.054879354" Dec 05 09:06:36 crc kubenswrapper[4997]: I1205 09:06:36.842967 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 05 09:06:37 crc kubenswrapper[4997]: I1205 09:06:37.135635 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:06:37 crc kubenswrapper[4997]: I1205 09:06:37.239248 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d69fcbd89-g6w6k"] Dec 05 09:06:37 crc kubenswrapper[4997]: I1205 09:06:37.326061 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" podUID="127f17b4-1636-4e11-932b-a15383b8132d" containerName="dnsmasq-dns" containerID="cri-o://44ccca432d52c528cdc51a7e3b204d1fa8b317d36bc087e12bdf15ab4057b21d" gracePeriod=10 Dec 05 09:06:37 crc kubenswrapper[4997]: I1205 09:06:37.327568 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"799f7992-40e0-4fea-8e26-84a596b0500f","Type":"ContainerStarted","Data":"9f5ebbf083003ad0f96f3653d445fe2857c98e4f6cb1e4b464e52fccf474624a"} Dec 05 09:06:37 crc kubenswrapper[4997]: I1205 09:06:37.327609 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"799f7992-40e0-4fea-8e26-84a596b0500f","Type":"ContainerStarted","Data":"932b29b196be9faa6ad3d99e32775e596ba7aac34fa36d37f786130074fe1295"} Dec 05 09:06:37 crc kubenswrapper[4997]: I1205 09:06:37.369316 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.436239018 podStartE2EDuration="11.369292912s" podCreationTimestamp="2025-12-05 09:06:26 +0000 UTC" firstStartedPulling="2025-12-05 09:06:28.037030948 +0000 UTC m=+7888.565938209" lastFinishedPulling="2025-12-05 09:06:35.970084822 +0000 UTC m=+7896.498992103" observedRunningTime="2025-12-05 09:06:37.353954309 +0000 UTC m=+7897.882861600" watchObservedRunningTime="2025-12-05 09:06:37.369292912 +0000 UTC m=+7897.898200183" Dec 05 09:06:37 crc kubenswrapper[4997]: I1205 09:06:37.991341 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.104810 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-dns-svc\") pod \"127f17b4-1636-4e11-932b-a15383b8132d\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.104877 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-config\") pod \"127f17b4-1636-4e11-932b-a15383b8132d\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.105146 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7t4j\" (UniqueName: \"kubernetes.io/projected/127f17b4-1636-4e11-932b-a15383b8132d-kube-api-access-h7t4j\") pod \"127f17b4-1636-4e11-932b-a15383b8132d\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.105238 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-ovsdbserver-sb\") pod \"127f17b4-1636-4e11-932b-a15383b8132d\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.105269 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-ovsdbserver-nb\") pod \"127f17b4-1636-4e11-932b-a15383b8132d\" (UID: \"127f17b4-1636-4e11-932b-a15383b8132d\") " Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.138150 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/127f17b4-1636-4e11-932b-a15383b8132d-kube-api-access-h7t4j" (OuterVolumeSpecName: "kube-api-access-h7t4j") pod "127f17b4-1636-4e11-932b-a15383b8132d" (UID: "127f17b4-1636-4e11-932b-a15383b8132d"). InnerVolumeSpecName "kube-api-access-h7t4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.222580 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7t4j\" (UniqueName: \"kubernetes.io/projected/127f17b4-1636-4e11-932b-a15383b8132d-kube-api-access-h7t4j\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.251814 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "127f17b4-1636-4e11-932b-a15383b8132d" (UID: "127f17b4-1636-4e11-932b-a15383b8132d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.252000 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "127f17b4-1636-4e11-932b-a15383b8132d" (UID: "127f17b4-1636-4e11-932b-a15383b8132d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.256191 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "127f17b4-1636-4e11-932b-a15383b8132d" (UID: "127f17b4-1636-4e11-932b-a15383b8132d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.273118 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-config" (OuterVolumeSpecName: "config") pod "127f17b4-1636-4e11-932b-a15383b8132d" (UID: "127f17b4-1636-4e11-932b-a15383b8132d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.325083 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.325125 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.325142 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.325153 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/127f17b4-1636-4e11-932b-a15383b8132d-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.337379 4997 generic.go:334] "Generic (PLEG): container finished" podID="127f17b4-1636-4e11-932b-a15383b8132d" containerID="44ccca432d52c528cdc51a7e3b204d1fa8b317d36bc087e12bdf15ab4057b21d" exitCode=0 Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.337442 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.337436 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" event={"ID":"127f17b4-1636-4e11-932b-a15383b8132d","Type":"ContainerDied","Data":"44ccca432d52c528cdc51a7e3b204d1fa8b317d36bc087e12bdf15ab4057b21d"} Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.337591 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d69fcbd89-g6w6k" event={"ID":"127f17b4-1636-4e11-932b-a15383b8132d","Type":"ContainerDied","Data":"c930ef5a7239f2f29d8f16ba58def8958d612c8e860a71eb0e7c5cf6bedeac34"} Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.337635 4997 scope.go:117] "RemoveContainer" containerID="44ccca432d52c528cdc51a7e3b204d1fa8b317d36bc087e12bdf15ab4057b21d" Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.361844 4997 scope.go:117] "RemoveContainer" containerID="584351b11a83e68616611079acabefdbefdeb7b783697ecd7d8b26b8d5cc0c7a" Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.377895 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d69fcbd89-g6w6k"] Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.395551 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d69fcbd89-g6w6k"] Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.399251 4997 scope.go:117] "RemoveContainer" containerID="44ccca432d52c528cdc51a7e3b204d1fa8b317d36bc087e12bdf15ab4057b21d" Dec 05 09:06:38 crc kubenswrapper[4997]: E1205 09:06:38.401032 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44ccca432d52c528cdc51a7e3b204d1fa8b317d36bc087e12bdf15ab4057b21d\": container with ID starting with 44ccca432d52c528cdc51a7e3b204d1fa8b317d36bc087e12bdf15ab4057b21d not found: ID does not exist" containerID="44ccca432d52c528cdc51a7e3b204d1fa8b317d36bc087e12bdf15ab4057b21d" Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.401063 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44ccca432d52c528cdc51a7e3b204d1fa8b317d36bc087e12bdf15ab4057b21d"} err="failed to get container status \"44ccca432d52c528cdc51a7e3b204d1fa8b317d36bc087e12bdf15ab4057b21d\": rpc error: code = NotFound desc = could not find container \"44ccca432d52c528cdc51a7e3b204d1fa8b317d36bc087e12bdf15ab4057b21d\": container with ID starting with 44ccca432d52c528cdc51a7e3b204d1fa8b317d36bc087e12bdf15ab4057b21d not found: ID does not exist" Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.401085 4997 scope.go:117] "RemoveContainer" containerID="584351b11a83e68616611079acabefdbefdeb7b783697ecd7d8b26b8d5cc0c7a" Dec 05 09:06:38 crc kubenswrapper[4997]: E1205 09:06:38.401778 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"584351b11a83e68616611079acabefdbefdeb7b783697ecd7d8b26b8d5cc0c7a\": container with ID starting with 584351b11a83e68616611079acabefdbefdeb7b783697ecd7d8b26b8d5cc0c7a not found: ID does not exist" containerID="584351b11a83e68616611079acabefdbefdeb7b783697ecd7d8b26b8d5cc0c7a" Dec 05 09:06:38 crc kubenswrapper[4997]: I1205 09:06:38.401801 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"584351b11a83e68616611079acabefdbefdeb7b783697ecd7d8b26b8d5cc0c7a"} err="failed to get container status \"584351b11a83e68616611079acabefdbefdeb7b783697ecd7d8b26b8d5cc0c7a\": rpc error: code = NotFound desc = could not find container \"584351b11a83e68616611079acabefdbefdeb7b783697ecd7d8b26b8d5cc0c7a\": container with ID starting with 584351b11a83e68616611079acabefdbefdeb7b783697ecd7d8b26b8d5cc0c7a not found: ID does not exist" Dec 05 09:06:39 crc kubenswrapper[4997]: I1205 09:06:39.763186 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="127f17b4-1636-4e11-932b-a15383b8132d" path="/var/lib/kubelet/pods/127f17b4-1636-4e11-932b-a15383b8132d/volumes" Dec 05 09:06:40 crc kubenswrapper[4997]: I1205 09:06:40.253540 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:06:40 crc kubenswrapper[4997]: I1205 09:06:40.254164 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerName="ceilometer-central-agent" containerID="cri-o://d24348b52188e11c84b19c6cfdfbfb63dea98de53d6c177b725d62e9d7ed9f7e" gracePeriod=30 Dec 05 09:06:40 crc kubenswrapper[4997]: I1205 09:06:40.254242 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerName="sg-core" containerID="cri-o://e4f498aab2f391e7425ca69f4a11c646ae365266d2a061b73155d932a0057242" gracePeriod=30 Dec 05 09:06:40 crc kubenswrapper[4997]: I1205 09:06:40.254281 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerName="ceilometer-notification-agent" containerID="cri-o://04735f873c32d25d40773cecd9f768d0e9a5fd47916f5f7eca2190758a9242ed" gracePeriod=30 Dec 05 09:06:40 crc kubenswrapper[4997]: I1205 09:06:40.254588 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerName="proxy-httpd" containerID="cri-o://b6b88db011a8f01238fadbf8d88ceb77c8c62ac98f34b74fb1ff1efb0539c561" gracePeriod=30 Dec 05 09:06:41 crc kubenswrapper[4997]: I1205 09:06:41.377332 4997 generic.go:334] "Generic (PLEG): container finished" podID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerID="b6b88db011a8f01238fadbf8d88ceb77c8c62ac98f34b74fb1ff1efb0539c561" exitCode=0 Dec 05 09:06:41 crc kubenswrapper[4997]: I1205 09:06:41.377973 4997 generic.go:334] "Generic (PLEG): container finished" podID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerID="e4f498aab2f391e7425ca69f4a11c646ae365266d2a061b73155d932a0057242" exitCode=2 Dec 05 09:06:41 crc kubenswrapper[4997]: I1205 09:06:41.377989 4997 generic.go:334] "Generic (PLEG): container finished" podID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerID="d24348b52188e11c84b19c6cfdfbfb63dea98de53d6c177b725d62e9d7ed9f7e" exitCode=0 Dec 05 09:06:41 crc kubenswrapper[4997]: I1205 09:06:41.377387 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8347c13-a145-4306-a8a9-f3f8b617f296","Type":"ContainerDied","Data":"b6b88db011a8f01238fadbf8d88ceb77c8c62ac98f34b74fb1ff1efb0539c561"} Dec 05 09:06:41 crc kubenswrapper[4997]: I1205 09:06:41.378063 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8347c13-a145-4306-a8a9-f3f8b617f296","Type":"ContainerDied","Data":"e4f498aab2f391e7425ca69f4a11c646ae365266d2a061b73155d932a0057242"} Dec 05 09:06:41 crc kubenswrapper[4997]: I1205 09:06:41.378098 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8347c13-a145-4306-a8a9-f3f8b617f296","Type":"ContainerDied","Data":"d24348b52188e11c84b19c6cfdfbfb63dea98de53d6c177b725d62e9d7ed9f7e"} Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.217580 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.337326 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-sg-core-conf-yaml\") pod \"e8347c13-a145-4306-a8a9-f3f8b617f296\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.337430 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-scripts\") pod \"e8347c13-a145-4306-a8a9-f3f8b617f296\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.337606 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-config-data\") pod \"e8347c13-a145-4306-a8a9-f3f8b617f296\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.338947 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8347c13-a145-4306-a8a9-f3f8b617f296-log-httpd\") pod \"e8347c13-a145-4306-a8a9-f3f8b617f296\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.339013 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hrk7\" (UniqueName: \"kubernetes.io/projected/e8347c13-a145-4306-a8a9-f3f8b617f296-kube-api-access-5hrk7\") pod \"e8347c13-a145-4306-a8a9-f3f8b617f296\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.339547 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8347c13-a145-4306-a8a9-f3f8b617f296-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e8347c13-a145-4306-a8a9-f3f8b617f296" (UID: "e8347c13-a145-4306-a8a9-f3f8b617f296"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.339639 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8347c13-a145-4306-a8a9-f3f8b617f296-run-httpd\") pod \"e8347c13-a145-4306-a8a9-f3f8b617f296\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.339793 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-combined-ca-bundle\") pod \"e8347c13-a145-4306-a8a9-f3f8b617f296\" (UID: \"e8347c13-a145-4306-a8a9-f3f8b617f296\") " Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.340128 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8347c13-a145-4306-a8a9-f3f8b617f296-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e8347c13-a145-4306-a8a9-f3f8b617f296" (UID: "e8347c13-a145-4306-a8a9-f3f8b617f296"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.341432 4997 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8347c13-a145-4306-a8a9-f3f8b617f296-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.341458 4997 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8347c13-a145-4306-a8a9-f3f8b617f296-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.345904 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8347c13-a145-4306-a8a9-f3f8b617f296-kube-api-access-5hrk7" (OuterVolumeSpecName: "kube-api-access-5hrk7") pod "e8347c13-a145-4306-a8a9-f3f8b617f296" (UID: "e8347c13-a145-4306-a8a9-f3f8b617f296"). InnerVolumeSpecName "kube-api-access-5hrk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.346629 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-scripts" (OuterVolumeSpecName: "scripts") pod "e8347c13-a145-4306-a8a9-f3f8b617f296" (UID: "e8347c13-a145-4306-a8a9-f3f8b617f296"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.375216 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e8347c13-a145-4306-a8a9-f3f8b617f296" (UID: "e8347c13-a145-4306-a8a9-f3f8b617f296"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.411035 4997 generic.go:334] "Generic (PLEG): container finished" podID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerID="04735f873c32d25d40773cecd9f768d0e9a5fd47916f5f7eca2190758a9242ed" exitCode=0 Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.411086 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8347c13-a145-4306-a8a9-f3f8b617f296","Type":"ContainerDied","Data":"04735f873c32d25d40773cecd9f768d0e9a5fd47916f5f7eca2190758a9242ed"} Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.411120 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8347c13-a145-4306-a8a9-f3f8b617f296","Type":"ContainerDied","Data":"1812c75822b87a74d8547661db0fd405e03871b8baa42ef4f6b0d176880a014e"} Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.411142 4997 scope.go:117] "RemoveContainer" containerID="b6b88db011a8f01238fadbf8d88ceb77c8c62ac98f34b74fb1ff1efb0539c561" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.411160 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.427288 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8347c13-a145-4306-a8a9-f3f8b617f296" (UID: "e8347c13-a145-4306-a8a9-f3f8b617f296"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.443764 4997 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.443802 4997 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.443815 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hrk7\" (UniqueName: \"kubernetes.io/projected/e8347c13-a145-4306-a8a9-f3f8b617f296-kube-api-access-5hrk7\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.443824 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.453296 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-config-data" (OuterVolumeSpecName: "config-data") pod "e8347c13-a145-4306-a8a9-f3f8b617f296" (UID: "e8347c13-a145-4306-a8a9-f3f8b617f296"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.474490 4997 scope.go:117] "RemoveContainer" containerID="e4f498aab2f391e7425ca69f4a11c646ae365266d2a061b73155d932a0057242" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.499846 4997 scope.go:117] "RemoveContainer" containerID="04735f873c32d25d40773cecd9f768d0e9a5fd47916f5f7eca2190758a9242ed" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.524486 4997 scope.go:117] "RemoveContainer" containerID="d24348b52188e11c84b19c6cfdfbfb63dea98de53d6c177b725d62e9d7ed9f7e" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.545334 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8347c13-a145-4306-a8a9-f3f8b617f296-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.549214 4997 scope.go:117] "RemoveContainer" containerID="b6b88db011a8f01238fadbf8d88ceb77c8c62ac98f34b74fb1ff1efb0539c561" Dec 05 09:06:43 crc kubenswrapper[4997]: E1205 09:06:43.551506 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6b88db011a8f01238fadbf8d88ceb77c8c62ac98f34b74fb1ff1efb0539c561\": container with ID starting with b6b88db011a8f01238fadbf8d88ceb77c8c62ac98f34b74fb1ff1efb0539c561 not found: ID does not exist" containerID="b6b88db011a8f01238fadbf8d88ceb77c8c62ac98f34b74fb1ff1efb0539c561" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.551561 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6b88db011a8f01238fadbf8d88ceb77c8c62ac98f34b74fb1ff1efb0539c561"} err="failed to get container status \"b6b88db011a8f01238fadbf8d88ceb77c8c62ac98f34b74fb1ff1efb0539c561\": rpc error: code = NotFound desc = could not find container \"b6b88db011a8f01238fadbf8d88ceb77c8c62ac98f34b74fb1ff1efb0539c561\": container with ID starting with b6b88db011a8f01238fadbf8d88ceb77c8c62ac98f34b74fb1ff1efb0539c561 not found: ID does not exist" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.551593 4997 scope.go:117] "RemoveContainer" containerID="e4f498aab2f391e7425ca69f4a11c646ae365266d2a061b73155d932a0057242" Dec 05 09:06:43 crc kubenswrapper[4997]: E1205 09:06:43.552132 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4f498aab2f391e7425ca69f4a11c646ae365266d2a061b73155d932a0057242\": container with ID starting with e4f498aab2f391e7425ca69f4a11c646ae365266d2a061b73155d932a0057242 not found: ID does not exist" containerID="e4f498aab2f391e7425ca69f4a11c646ae365266d2a061b73155d932a0057242" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.552182 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4f498aab2f391e7425ca69f4a11c646ae365266d2a061b73155d932a0057242"} err="failed to get container status \"e4f498aab2f391e7425ca69f4a11c646ae365266d2a061b73155d932a0057242\": rpc error: code = NotFound desc = could not find container \"e4f498aab2f391e7425ca69f4a11c646ae365266d2a061b73155d932a0057242\": container with ID starting with e4f498aab2f391e7425ca69f4a11c646ae365266d2a061b73155d932a0057242 not found: ID does not exist" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.552212 4997 scope.go:117] "RemoveContainer" containerID="04735f873c32d25d40773cecd9f768d0e9a5fd47916f5f7eca2190758a9242ed" Dec 05 09:06:43 crc kubenswrapper[4997]: E1205 09:06:43.553089 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04735f873c32d25d40773cecd9f768d0e9a5fd47916f5f7eca2190758a9242ed\": container with ID starting with 04735f873c32d25d40773cecd9f768d0e9a5fd47916f5f7eca2190758a9242ed not found: ID does not exist" containerID="04735f873c32d25d40773cecd9f768d0e9a5fd47916f5f7eca2190758a9242ed" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.553122 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04735f873c32d25d40773cecd9f768d0e9a5fd47916f5f7eca2190758a9242ed"} err="failed to get container status \"04735f873c32d25d40773cecd9f768d0e9a5fd47916f5f7eca2190758a9242ed\": rpc error: code = NotFound desc = could not find container \"04735f873c32d25d40773cecd9f768d0e9a5fd47916f5f7eca2190758a9242ed\": container with ID starting with 04735f873c32d25d40773cecd9f768d0e9a5fd47916f5f7eca2190758a9242ed not found: ID does not exist" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.553138 4997 scope.go:117] "RemoveContainer" containerID="d24348b52188e11c84b19c6cfdfbfb63dea98de53d6c177b725d62e9d7ed9f7e" Dec 05 09:06:43 crc kubenswrapper[4997]: E1205 09:06:43.554946 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d24348b52188e11c84b19c6cfdfbfb63dea98de53d6c177b725d62e9d7ed9f7e\": container with ID starting with d24348b52188e11c84b19c6cfdfbfb63dea98de53d6c177b725d62e9d7ed9f7e not found: ID does not exist" containerID="d24348b52188e11c84b19c6cfdfbfb63dea98de53d6c177b725d62e9d7ed9f7e" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.554970 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d24348b52188e11c84b19c6cfdfbfb63dea98de53d6c177b725d62e9d7ed9f7e"} err="failed to get container status \"d24348b52188e11c84b19c6cfdfbfb63dea98de53d6c177b725d62e9d7ed9f7e\": rpc error: code = NotFound desc = could not find container \"d24348b52188e11c84b19c6cfdfbfb63dea98de53d6c177b725d62e9d7ed9f7e\": container with ID starting with d24348b52188e11c84b19c6cfdfbfb63dea98de53d6c177b725d62e9d7ed9f7e not found: ID does not exist" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.750179 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.764022 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.784473 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:06:43 crc kubenswrapper[4997]: E1205 09:06:43.785338 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerName="ceilometer-central-agent" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.785448 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerName="ceilometer-central-agent" Dec 05 09:06:43 crc kubenswrapper[4997]: E1205 09:06:43.786464 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="127f17b4-1636-4e11-932b-a15383b8132d" containerName="init" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.786497 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="127f17b4-1636-4e11-932b-a15383b8132d" containerName="init" Dec 05 09:06:43 crc kubenswrapper[4997]: E1205 09:06:43.786554 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerName="sg-core" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.786565 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerName="sg-core" Dec 05 09:06:43 crc kubenswrapper[4997]: E1205 09:06:43.786608 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerName="proxy-httpd" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.786644 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerName="proxy-httpd" Dec 05 09:06:43 crc kubenswrapper[4997]: E1205 09:06:43.786666 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="127f17b4-1636-4e11-932b-a15383b8132d" containerName="dnsmasq-dns" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.786674 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="127f17b4-1636-4e11-932b-a15383b8132d" containerName="dnsmasq-dns" Dec 05 09:06:43 crc kubenswrapper[4997]: E1205 09:06:43.786692 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerName="ceilometer-notification-agent" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.786701 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerName="ceilometer-notification-agent" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.787168 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="127f17b4-1636-4e11-932b-a15383b8132d" containerName="dnsmasq-dns" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.787198 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerName="sg-core" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.787243 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerName="proxy-httpd" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.787261 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerName="ceilometer-central-agent" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.787273 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8347c13-a145-4306-a8a9-f3f8b617f296" containerName="ceilometer-notification-agent" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.789741 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.791893 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.792169 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.804016 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.959879 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvzqn\" (UniqueName: \"kubernetes.io/projected/2a2df928-f920-4e13-87ae-a412539dc001-kube-api-access-hvzqn\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.959930 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a2df928-f920-4e13-87ae-a412539dc001-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.959980 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a2df928-f920-4e13-87ae-a412539dc001-config-data\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.960077 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a2df928-f920-4e13-87ae-a412539dc001-log-httpd\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.960158 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a2df928-f920-4e13-87ae-a412539dc001-run-httpd\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.960211 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a2df928-f920-4e13-87ae-a412539dc001-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:43 crc kubenswrapper[4997]: I1205 09:06:43.960284 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a2df928-f920-4e13-87ae-a412539dc001-scripts\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:44 crc kubenswrapper[4997]: I1205 09:06:44.061979 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a2df928-f920-4e13-87ae-a412539dc001-run-httpd\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:44 crc kubenswrapper[4997]: I1205 09:06:44.062040 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a2df928-f920-4e13-87ae-a412539dc001-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:44 crc kubenswrapper[4997]: I1205 09:06:44.062107 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a2df928-f920-4e13-87ae-a412539dc001-scripts\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:44 crc kubenswrapper[4997]: I1205 09:06:44.062205 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvzqn\" (UniqueName: \"kubernetes.io/projected/2a2df928-f920-4e13-87ae-a412539dc001-kube-api-access-hvzqn\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:44 crc kubenswrapper[4997]: I1205 09:06:44.062232 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a2df928-f920-4e13-87ae-a412539dc001-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:44 crc kubenswrapper[4997]: I1205 09:06:44.062269 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a2df928-f920-4e13-87ae-a412539dc001-config-data\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:44 crc kubenswrapper[4997]: I1205 09:06:44.062306 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a2df928-f920-4e13-87ae-a412539dc001-log-httpd\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:44 crc kubenswrapper[4997]: I1205 09:06:44.062710 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a2df928-f920-4e13-87ae-a412539dc001-run-httpd\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:44 crc kubenswrapper[4997]: I1205 09:06:44.062803 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a2df928-f920-4e13-87ae-a412539dc001-log-httpd\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:44 crc kubenswrapper[4997]: I1205 09:06:44.066304 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a2df928-f920-4e13-87ae-a412539dc001-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:44 crc kubenswrapper[4997]: I1205 09:06:44.068399 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a2df928-f920-4e13-87ae-a412539dc001-config-data\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:44 crc kubenswrapper[4997]: I1205 09:06:44.069165 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a2df928-f920-4e13-87ae-a412539dc001-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:44 crc kubenswrapper[4997]: I1205 09:06:44.079097 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a2df928-f920-4e13-87ae-a412539dc001-scripts\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:44 crc kubenswrapper[4997]: I1205 09:06:44.081355 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvzqn\" (UniqueName: \"kubernetes.io/projected/2a2df928-f920-4e13-87ae-a412539dc001-kube-api-access-hvzqn\") pod \"ceilometer-0\" (UID: \"2a2df928-f920-4e13-87ae-a412539dc001\") " pod="openstack/ceilometer-0" Dec 05 09:06:44 crc kubenswrapper[4997]: I1205 09:06:44.106868 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:06:44 crc kubenswrapper[4997]: W1205 09:06:44.608252 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a2df928_f920_4e13_87ae_a412539dc001.slice/crio-e561fd9c3975d5dc7dd7a3c3ee4b08bb391869bc02d9419d1fa874bee423d3de WatchSource:0}: Error finding container e561fd9c3975d5dc7dd7a3c3ee4b08bb391869bc02d9419d1fa874bee423d3de: Status 404 returned error can't find the container with id e561fd9c3975d5dc7dd7a3c3ee4b08bb391869bc02d9419d1fa874bee423d3de Dec 05 09:06:44 crc kubenswrapper[4997]: I1205 09:06:44.612059 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:06:45 crc kubenswrapper[4997]: I1205 09:06:45.449663 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a2df928-f920-4e13-87ae-a412539dc001","Type":"ContainerStarted","Data":"5c29397ef88a0f032666241990b355b804da3972431a5176d69ad97de3c0a57a"} Dec 05 09:06:45 crc kubenswrapper[4997]: I1205 09:06:45.449979 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a2df928-f920-4e13-87ae-a412539dc001","Type":"ContainerStarted","Data":"60db9016fb6f40a2ff6b5d3bdd4b790aa713b06fff0fc66201e5a52b5ca6d124"} Dec 05 09:06:45 crc kubenswrapper[4997]: I1205 09:06:45.449989 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a2df928-f920-4e13-87ae-a412539dc001","Type":"ContainerStarted","Data":"e561fd9c3975d5dc7dd7a3c3ee4b08bb391869bc02d9419d1fa874bee423d3de"} Dec 05 09:06:45 crc kubenswrapper[4997]: I1205 09:06:45.761659 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8347c13-a145-4306-a8a9-f3f8b617f296" path="/var/lib/kubelet/pods/e8347c13-a145-4306-a8a9-f3f8b617f296/volumes" Dec 05 09:06:46 crc kubenswrapper[4997]: I1205 09:06:46.461005 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a2df928-f920-4e13-87ae-a412539dc001","Type":"ContainerStarted","Data":"c7cc3ddc0e6218c7b77c6bbbb725379966babff79e4e1c89c0112c982b543841"} Dec 05 09:06:46 crc kubenswrapper[4997]: I1205 09:06:46.885206 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 05 09:06:47 crc kubenswrapper[4997]: I1205 09:06:47.474064 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a2df928-f920-4e13-87ae-a412539dc001","Type":"ContainerStarted","Data":"74341a034f8dfe42768750c243cd7903ab7d7713c7bb58373c91ab61ad4db7d1"} Dec 05 09:06:47 crc kubenswrapper[4997]: I1205 09:06:47.474518 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 09:06:47 crc kubenswrapper[4997]: I1205 09:06:47.506649 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.998717557 podStartE2EDuration="4.506610368s" podCreationTimestamp="2025-12-05 09:06:43 +0000 UTC" firstStartedPulling="2025-12-05 09:06:44.610538058 +0000 UTC m=+7905.139445319" lastFinishedPulling="2025-12-05 09:06:47.118430859 +0000 UTC m=+7907.647338130" observedRunningTime="2025-12-05 09:06:47.500864373 +0000 UTC m=+7908.029771644" watchObservedRunningTime="2025-12-05 09:06:47.506610368 +0000 UTC m=+7908.035517629" Dec 05 09:06:48 crc kubenswrapper[4997]: I1205 09:06:48.780526 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 05 09:06:48 crc kubenswrapper[4997]: I1205 09:06:48.825895 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 05 09:06:48 crc kubenswrapper[4997]: I1205 09:06:48.997397 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Dec 05 09:06:49 crc kubenswrapper[4997]: I1205 09:06:49.772378 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:06:49 crc kubenswrapper[4997]: I1205 09:06:49.773028 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:07:11 crc kubenswrapper[4997]: I1205 09:07:11.046662 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-a76d-account-create-update-rn7tm"] Dec 05 09:07:11 crc kubenswrapper[4997]: I1205 09:07:11.062271 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-5tctv"] Dec 05 09:07:11 crc kubenswrapper[4997]: I1205 09:07:11.070933 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-750c-account-create-update-jd4sj"] Dec 05 09:07:11 crc kubenswrapper[4997]: I1205 09:07:11.079084 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-5tctv"] Dec 05 09:07:11 crc kubenswrapper[4997]: I1205 09:07:11.095795 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-mkz6j"] Dec 05 09:07:11 crc kubenswrapper[4997]: I1205 09:07:11.107077 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-w6nhg"] Dec 05 09:07:11 crc kubenswrapper[4997]: I1205 09:07:11.116658 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-750c-account-create-update-jd4sj"] Dec 05 09:07:11 crc kubenswrapper[4997]: I1205 09:07:11.127109 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-a76d-account-create-update-rn7tm"] Dec 05 09:07:11 crc kubenswrapper[4997]: I1205 09:07:11.137405 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-mkz6j"] Dec 05 09:07:11 crc kubenswrapper[4997]: I1205 09:07:11.146481 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-w6nhg"] Dec 05 09:07:11 crc kubenswrapper[4997]: I1205 09:07:11.762175 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3987fcc4-0e92-4006-a1f5-ad2c937f057e" path="/var/lib/kubelet/pods/3987fcc4-0e92-4006-a1f5-ad2c937f057e/volumes" Dec 05 09:07:11 crc kubenswrapper[4997]: I1205 09:07:11.762969 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ec1f8f1-691f-4750-ad3b-8da18a706133" path="/var/lib/kubelet/pods/3ec1f8f1-691f-4750-ad3b-8da18a706133/volumes" Dec 05 09:07:11 crc kubenswrapper[4997]: I1205 09:07:11.763683 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3ea1a78-5866-44a2-b8ce-5203e199f4f9" path="/var/lib/kubelet/pods/c3ea1a78-5866-44a2-b8ce-5203e199f4f9/volumes" Dec 05 09:07:11 crc kubenswrapper[4997]: I1205 09:07:11.764288 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f" path="/var/lib/kubelet/pods/ccfb9b2b-0899-4f23-a7a2-601fd5a6a66f/volumes" Dec 05 09:07:11 crc kubenswrapper[4997]: I1205 09:07:11.765479 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0fb2077-9180-483e-94ce-c73bee41dd1b" path="/var/lib/kubelet/pods/d0fb2077-9180-483e-94ce-c73bee41dd1b/volumes" Dec 05 09:07:12 crc kubenswrapper[4997]: I1205 09:07:12.031886 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-fbe8-account-create-update-p8p6z"] Dec 05 09:07:12 crc kubenswrapper[4997]: I1205 09:07:12.045046 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-fbe8-account-create-update-p8p6z"] Dec 05 09:07:13 crc kubenswrapper[4997]: I1205 09:07:13.760896 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e23dfad8-111d-415a-bff0-2b8796cc951d" path="/var/lib/kubelet/pods/e23dfad8-111d-415a-bff0-2b8796cc951d/volumes" Dec 05 09:07:14 crc kubenswrapper[4997]: I1205 09:07:14.113744 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 09:07:17 crc kubenswrapper[4997]: I1205 09:07:17.923744 4997 scope.go:117] "RemoveContainer" containerID="7832ca93ad1b7e818faff5771a9058448de46d6009bdc24638f6f51ca37d74a1" Dec 05 09:07:17 crc kubenswrapper[4997]: I1205 09:07:17.946806 4997 scope.go:117] "RemoveContainer" containerID="3ee65aa68794c7532078e07e4859042a969598949ca778893474cbf07e1236e9" Dec 05 09:07:17 crc kubenswrapper[4997]: I1205 09:07:17.991571 4997 scope.go:117] "RemoveContainer" containerID="b25f9149e65b3ad1ff2e3498fea6cd8ea1e00287214d430556e09c7021437fa4" Dec 05 09:07:18 crc kubenswrapper[4997]: I1205 09:07:18.034534 4997 scope.go:117] "RemoveContainer" containerID="a782bd326b4eb5a7bc1dcf43d39a57cb83d62c4c33ec62697bc83ad8d3e9fe23" Dec 05 09:07:18 crc kubenswrapper[4997]: I1205 09:07:18.079342 4997 scope.go:117] "RemoveContainer" containerID="afbdaf466a44206a95ee91f0f74bdc6cb1a2fcff639c2147026e6faf413d2157" Dec 05 09:07:18 crc kubenswrapper[4997]: I1205 09:07:18.137529 4997 scope.go:117] "RemoveContainer" containerID="5d80064eba66e65f9b836496de33afded37525a73aa1631ca6a871a1fc465fd2" Dec 05 09:07:19 crc kubenswrapper[4997]: I1205 09:07:19.770169 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:07:19 crc kubenswrapper[4997]: I1205 09:07:19.770469 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:07:19 crc kubenswrapper[4997]: I1205 09:07:19.770508 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 09:07:19 crc kubenswrapper[4997]: I1205 09:07:19.771244 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a4c28f2d7c6c0a58e95d5724c2f234da68b209b35a39f917c1bd66336a29eaeb"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:07:19 crc kubenswrapper[4997]: I1205 09:07:19.771294 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://a4c28f2d7c6c0a58e95d5724c2f234da68b209b35a39f917c1bd66336a29eaeb" gracePeriod=600 Dec 05 09:07:20 crc kubenswrapper[4997]: I1205 09:07:20.801046 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="a4c28f2d7c6c0a58e95d5724c2f234da68b209b35a39f917c1bd66336a29eaeb" exitCode=0 Dec 05 09:07:20 crc kubenswrapper[4997]: I1205 09:07:20.801116 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"a4c28f2d7c6c0a58e95d5724c2f234da68b209b35a39f917c1bd66336a29eaeb"} Dec 05 09:07:20 crc kubenswrapper[4997]: I1205 09:07:20.801459 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9"} Dec 05 09:07:20 crc kubenswrapper[4997]: I1205 09:07:20.801476 4997 scope.go:117] "RemoveContainer" containerID="0daa369ccd66d629dd5e19ad5f5c73c8b226a4ada06535e3aca8f5adaba1034b" Dec 05 09:07:32 crc kubenswrapper[4997]: I1205 09:07:32.437947 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bpbv8"] Dec 05 09:07:32 crc kubenswrapper[4997]: I1205 09:07:32.450971 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bpbv8"] Dec 05 09:07:33 crc kubenswrapper[4997]: I1205 09:07:33.765384 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66bea143-bb89-4a9c-82cb-d0cff893d8ad" path="/var/lib/kubelet/pods/66bea143-bb89-4a9c-82cb-d0cff893d8ad/volumes" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.639714 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-ff67498f5-cvdxb"] Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.641601 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.643420 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.655566 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-ff67498f5-cvdxb"] Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.838571 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-ovsdbserver-nb\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.838940 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-config\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.839041 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-openstack-cell1\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.839308 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-ovsdbserver-sb\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.839400 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jz56\" (UniqueName: \"kubernetes.io/projected/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-kube-api-access-8jz56\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.839506 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-dns-svc\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.941063 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-ovsdbserver-nb\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.941116 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-config\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.941157 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-openstack-cell1\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.941220 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-ovsdbserver-sb\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.941260 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jz56\" (UniqueName: \"kubernetes.io/projected/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-kube-api-access-8jz56\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.941304 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-dns-svc\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.942278 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-dns-svc\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.942315 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-ovsdbserver-nb\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.942358 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-config\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.942359 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-openstack-cell1\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.942468 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-ovsdbserver-sb\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.962556 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jz56\" (UniqueName: \"kubernetes.io/projected/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-kube-api-access-8jz56\") pod \"dnsmasq-dns-ff67498f5-cvdxb\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:34 crc kubenswrapper[4997]: I1205 09:07:34.972147 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:35 crc kubenswrapper[4997]: I1205 09:07:35.487342 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-ff67498f5-cvdxb"] Dec 05 09:07:35 crc kubenswrapper[4997]: I1205 09:07:35.962815 4997 generic.go:334] "Generic (PLEG): container finished" podID="c8ef7156-24ee-4b0a-b6ad-662bc4c30243" containerID="1d657bd903cedb7de10b0cc1640b6914a19b9a6629ed1efc6c7184162484beda" exitCode=0 Dec 05 09:07:35 crc kubenswrapper[4997]: I1205 09:07:35.962853 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" event={"ID":"c8ef7156-24ee-4b0a-b6ad-662bc4c30243","Type":"ContainerDied","Data":"1d657bd903cedb7de10b0cc1640b6914a19b9a6629ed1efc6c7184162484beda"} Dec 05 09:07:35 crc kubenswrapper[4997]: I1205 09:07:35.963225 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" event={"ID":"c8ef7156-24ee-4b0a-b6ad-662bc4c30243","Type":"ContainerStarted","Data":"12e0c7594832dc667149b0b7171e7613a8d3dc5ab972fde63ada0807cbe5512b"} Dec 05 09:07:36 crc kubenswrapper[4997]: I1205 09:07:36.974510 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" event={"ID":"c8ef7156-24ee-4b0a-b6ad-662bc4c30243","Type":"ContainerStarted","Data":"e6144019c8587fd424380aaaf0b68b82c95b1e804c934f8432dffe76843abe39"} Dec 05 09:07:36 crc kubenswrapper[4997]: I1205 09:07:36.975164 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:36 crc kubenswrapper[4997]: I1205 09:07:36.992811 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" podStartSLOduration=2.992788804 podStartE2EDuration="2.992788804s" podCreationTimestamp="2025-12-05 09:07:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:07:36.991269242 +0000 UTC m=+7957.520176513" watchObservedRunningTime="2025-12-05 09:07:36.992788804 +0000 UTC m=+7957.521696065" Dec 05 09:07:44 crc kubenswrapper[4997]: I1205 09:07:44.973821 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.044690 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c5dcbcb5-vz7hb"] Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.044954 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" podUID="4913c005-1117-40c8-9916-aecc8625d1ed" containerName="dnsmasq-dns" containerID="cri-o://fec63ae921673bae9d114a02c2f3d6e257ed483ba0168ccf7b0ad21b73a3254c" gracePeriod=10 Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.228563 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q"] Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.230300 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.234298 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-networker" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.245193 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q"] Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.332248 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q"] Dec 05 09:07:45 crc kubenswrapper[4997]: E1205 09:07:45.333162 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-xplbw openstack-cell1 openstack-networker ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" podUID="316f0f37-2cc5-49f5-b4c2-1f6980b2c593" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.360304 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-ovsdbserver-sb\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.360384 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xplbw\" (UniqueName: \"kubernetes.io/projected/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-kube-api-access-xplbw\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.360421 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-openstack-networker\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.360469 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-config\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.360638 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-openstack-cell1\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.360685 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-ovsdbserver-nb\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.360769 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-dns-svc\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.380828 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dc8dd569-nzqx8"] Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.382979 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.401909 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dc8dd569-nzqx8"] Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.464023 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-openstack-networker\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.464133 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-dns-svc\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.464166 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-openstack-cell1\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.464189 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-config\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.464228 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-769ts\" (UniqueName: \"kubernetes.io/projected/3b87119b-5ad1-435b-9289-5a5a71765f8a-kube-api-access-769ts\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.464289 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-ovsdbserver-sb\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.464323 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-ovsdbserver-nb\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.464354 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xplbw\" (UniqueName: \"kubernetes.io/projected/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-kube-api-access-xplbw\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.464372 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-ovsdbserver-sb\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.464390 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-openstack-networker\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.464430 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-config\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.464451 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-openstack-cell1\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.464490 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-ovsdbserver-nb\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.464515 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-dns-svc\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.465456 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-dns-svc\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.466362 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-ovsdbserver-sb\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.466926 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-config\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.467467 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-openstack-networker\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.468019 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-openstack-cell1\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.468528 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-ovsdbserver-nb\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.492354 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xplbw\" (UniqueName: \"kubernetes.io/projected/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-kube-api-access-xplbw\") pod \"dnsmasq-dns-7fb6c4b5c5-g2q2q\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.565931 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-ovsdbserver-nb\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.566519 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-ovsdbserver-sb\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.566673 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-dns-svc\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.566805 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-openstack-networker\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.566956 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-openstack-cell1\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.567029 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-config\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.567113 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-769ts\" (UniqueName: \"kubernetes.io/projected/3b87119b-5ad1-435b-9289-5a5a71765f8a-kube-api-access-769ts\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.568819 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-ovsdbserver-nb\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.569987 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-openstack-cell1\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.570042 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-openstack-networker\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.570862 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-config\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.571079 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-dns-svc\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.571240 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b87119b-5ad1-435b-9289-5a5a71765f8a-ovsdbserver-sb\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.602343 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-769ts\" (UniqueName: \"kubernetes.io/projected/3b87119b-5ad1-435b-9289-5a5a71765f8a-kube-api-access-769ts\") pod \"dnsmasq-dns-78dc8dd569-nzqx8\" (UID: \"3b87119b-5ad1-435b-9289-5a5a71765f8a\") " pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.732532 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.735393 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.770336 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-config\") pod \"4913c005-1117-40c8-9916-aecc8625d1ed\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.771807 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-ovsdbserver-nb\") pod \"4913c005-1117-40c8-9916-aecc8625d1ed\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.771877 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-dns-svc\") pod \"4913c005-1117-40c8-9916-aecc8625d1ed\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.771962 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-ovsdbserver-sb\") pod \"4913c005-1117-40c8-9916-aecc8625d1ed\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.771988 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jt54\" (UniqueName: \"kubernetes.io/projected/4913c005-1117-40c8-9916-aecc8625d1ed-kube-api-access-9jt54\") pod \"4913c005-1117-40c8-9916-aecc8625d1ed\" (UID: \"4913c005-1117-40c8-9916-aecc8625d1ed\") " Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.775779 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4913c005-1117-40c8-9916-aecc8625d1ed-kube-api-access-9jt54" (OuterVolumeSpecName: "kube-api-access-9jt54") pod "4913c005-1117-40c8-9916-aecc8625d1ed" (UID: "4913c005-1117-40c8-9916-aecc8625d1ed"). InnerVolumeSpecName "kube-api-access-9jt54". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.855948 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4913c005-1117-40c8-9916-aecc8625d1ed" (UID: "4913c005-1117-40c8-9916-aecc8625d1ed"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.859243 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-config" (OuterVolumeSpecName: "config") pod "4913c005-1117-40c8-9916-aecc8625d1ed" (UID: "4913c005-1117-40c8-9916-aecc8625d1ed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.874339 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4913c005-1117-40c8-9916-aecc8625d1ed" (UID: "4913c005-1117-40c8-9916-aecc8625d1ed"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.875416 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.875441 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.875454 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.875465 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jt54\" (UniqueName: \"kubernetes.io/projected/4913c005-1117-40c8-9916-aecc8625d1ed-kube-api-access-9jt54\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.879168 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4913c005-1117-40c8-9916-aecc8625d1ed" (UID: "4913c005-1117-40c8-9916-aecc8625d1ed"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:45 crc kubenswrapper[4997]: I1205 09:07:45.977043 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4913c005-1117-40c8-9916-aecc8625d1ed-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.090730 4997 generic.go:334] "Generic (PLEG): container finished" podID="4913c005-1117-40c8-9916-aecc8625d1ed" containerID="fec63ae921673bae9d114a02c2f3d6e257ed483ba0168ccf7b0ad21b73a3254c" exitCode=0 Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.090797 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.091416 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.091919 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" event={"ID":"4913c005-1117-40c8-9916-aecc8625d1ed","Type":"ContainerDied","Data":"fec63ae921673bae9d114a02c2f3d6e257ed483ba0168ccf7b0ad21b73a3254c"} Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.091969 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c5dcbcb5-vz7hb" event={"ID":"4913c005-1117-40c8-9916-aecc8625d1ed","Type":"ContainerDied","Data":"0687d407f5b92b6d235cc5dae7ae8a6d0e36cdaaf77d73b2c52a65a751f72d41"} Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.091990 4997 scope.go:117] "RemoveContainer" containerID="fec63ae921673bae9d114a02c2f3d6e257ed483ba0168ccf7b0ad21b73a3254c" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.102098 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.119523 4997 scope.go:117] "RemoveContainer" containerID="cd33f76d1fe4114fe360da1f7e8ea95c04569d18649fea06839c27dac260552c" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.128737 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c5dcbcb5-vz7hb"] Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.140140 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c5dcbcb5-vz7hb"] Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.160160 4997 scope.go:117] "RemoveContainer" containerID="fec63ae921673bae9d114a02c2f3d6e257ed483ba0168ccf7b0ad21b73a3254c" Dec 05 09:07:46 crc kubenswrapper[4997]: E1205 09:07:46.160753 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fec63ae921673bae9d114a02c2f3d6e257ed483ba0168ccf7b0ad21b73a3254c\": container with ID starting with fec63ae921673bae9d114a02c2f3d6e257ed483ba0168ccf7b0ad21b73a3254c not found: ID does not exist" containerID="fec63ae921673bae9d114a02c2f3d6e257ed483ba0168ccf7b0ad21b73a3254c" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.160795 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fec63ae921673bae9d114a02c2f3d6e257ed483ba0168ccf7b0ad21b73a3254c"} err="failed to get container status \"fec63ae921673bae9d114a02c2f3d6e257ed483ba0168ccf7b0ad21b73a3254c\": rpc error: code = NotFound desc = could not find container \"fec63ae921673bae9d114a02c2f3d6e257ed483ba0168ccf7b0ad21b73a3254c\": container with ID starting with fec63ae921673bae9d114a02c2f3d6e257ed483ba0168ccf7b0ad21b73a3254c not found: ID does not exist" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.160820 4997 scope.go:117] "RemoveContainer" containerID="cd33f76d1fe4114fe360da1f7e8ea95c04569d18649fea06839c27dac260552c" Dec 05 09:07:46 crc kubenswrapper[4997]: E1205 09:07:46.161306 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd33f76d1fe4114fe360da1f7e8ea95c04569d18649fea06839c27dac260552c\": container with ID starting with cd33f76d1fe4114fe360da1f7e8ea95c04569d18649fea06839c27dac260552c not found: ID does not exist" containerID="cd33f76d1fe4114fe360da1f7e8ea95c04569d18649fea06839c27dac260552c" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.161337 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd33f76d1fe4114fe360da1f7e8ea95c04569d18649fea06839c27dac260552c"} err="failed to get container status \"cd33f76d1fe4114fe360da1f7e8ea95c04569d18649fea06839c27dac260552c\": rpc error: code = NotFound desc = could not find container \"cd33f76d1fe4114fe360da1f7e8ea95c04569d18649fea06839c27dac260552c\": container with ID starting with cd33f76d1fe4114fe360da1f7e8ea95c04569d18649fea06839c27dac260552c not found: ID does not exist" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.179653 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-config\") pod \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.179706 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-openstack-cell1\") pod \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.179824 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-openstack-networker\") pod \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.179961 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-ovsdbserver-sb\") pod \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.180263 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-ovsdbserver-nb\") pod \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.180336 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xplbw\" (UniqueName: \"kubernetes.io/projected/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-kube-api-access-xplbw\") pod \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.180400 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-dns-svc\") pod \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\" (UID: \"316f0f37-2cc5-49f5-b4c2-1f6980b2c593\") " Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.181749 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-openstack-networker" (OuterVolumeSpecName: "openstack-networker") pod "316f0f37-2cc5-49f5-b4c2-1f6980b2c593" (UID: "316f0f37-2cc5-49f5-b4c2-1f6980b2c593"). InnerVolumeSpecName "openstack-networker". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.182142 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "316f0f37-2cc5-49f5-b4c2-1f6980b2c593" (UID: "316f0f37-2cc5-49f5-b4c2-1f6980b2c593"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.182289 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "316f0f37-2cc5-49f5-b4c2-1f6980b2c593" (UID: "316f0f37-2cc5-49f5-b4c2-1f6980b2c593"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.182299 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "316f0f37-2cc5-49f5-b4c2-1f6980b2c593" (UID: "316f0f37-2cc5-49f5-b4c2-1f6980b2c593"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.182681 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "316f0f37-2cc5-49f5-b4c2-1f6980b2c593" (UID: "316f0f37-2cc5-49f5-b4c2-1f6980b2c593"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.183518 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-config" (OuterVolumeSpecName: "config") pod "316f0f37-2cc5-49f5-b4c2-1f6980b2c593" (UID: "316f0f37-2cc5-49f5-b4c2-1f6980b2c593"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.185489 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-kube-api-access-xplbw" (OuterVolumeSpecName: "kube-api-access-xplbw") pod "316f0f37-2cc5-49f5-b4c2-1f6980b2c593" (UID: "316f0f37-2cc5-49f5-b4c2-1f6980b2c593"). InnerVolumeSpecName "kube-api-access-xplbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.228012 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dc8dd569-nzqx8"] Dec 05 09:07:46 crc kubenswrapper[4997]: W1205 09:07:46.235805 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b87119b_5ad1_435b_9289_5a5a71765f8a.slice/crio-9eb803fc4dd1bddf71f587aad6ccdc7b6f416c549ce3a59053ce4777fa0c0a82 WatchSource:0}: Error finding container 9eb803fc4dd1bddf71f587aad6ccdc7b6f416c549ce3a59053ce4777fa0c0a82: Status 404 returned error can't find the container with id 9eb803fc4dd1bddf71f587aad6ccdc7b6f416c549ce3a59053ce4777fa0c0a82 Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.282744 4997 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-openstack-cell1\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.282781 4997 reconciler_common.go:293] "Volume detached for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-openstack-networker\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.282793 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.282802 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.282812 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xplbw\" (UniqueName: \"kubernetes.io/projected/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-kube-api-access-xplbw\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.282824 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:46 crc kubenswrapper[4997]: I1205 09:07:46.282833 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/316f0f37-2cc5-49f5-b4c2-1f6980b2c593-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:47 crc kubenswrapper[4997]: I1205 09:07:47.101422 4997 generic.go:334] "Generic (PLEG): container finished" podID="3b87119b-5ad1-435b-9289-5a5a71765f8a" containerID="8cc5088769bdfeda928d82dc85f2a324d07adde2f44ad012d773e90e7d2b2855" exitCode=0 Dec 05 09:07:47 crc kubenswrapper[4997]: I1205 09:07:47.101587 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" event={"ID":"3b87119b-5ad1-435b-9289-5a5a71765f8a","Type":"ContainerDied","Data":"8cc5088769bdfeda928d82dc85f2a324d07adde2f44ad012d773e90e7d2b2855"} Dec 05 09:07:47 crc kubenswrapper[4997]: I1205 09:07:47.101871 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" event={"ID":"3b87119b-5ad1-435b-9289-5a5a71765f8a","Type":"ContainerStarted","Data":"9eb803fc4dd1bddf71f587aad6ccdc7b6f416c549ce3a59053ce4777fa0c0a82"} Dec 05 09:07:47 crc kubenswrapper[4997]: I1205 09:07:47.108816 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q" Dec 05 09:07:47 crc kubenswrapper[4997]: I1205 09:07:47.186931 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q"] Dec 05 09:07:47 crc kubenswrapper[4997]: I1205 09:07:47.200928 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fb6c4b5c5-g2q2q"] Dec 05 09:07:47 crc kubenswrapper[4997]: I1205 09:07:47.762350 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="316f0f37-2cc5-49f5-b4c2-1f6980b2c593" path="/var/lib/kubelet/pods/316f0f37-2cc5-49f5-b4c2-1f6980b2c593/volumes" Dec 05 09:07:47 crc kubenswrapper[4997]: I1205 09:07:47.763029 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4913c005-1117-40c8-9916-aecc8625d1ed" path="/var/lib/kubelet/pods/4913c005-1117-40c8-9916-aecc8625d1ed/volumes" Dec 05 09:07:48 crc kubenswrapper[4997]: I1205 09:07:48.118889 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" event={"ID":"3b87119b-5ad1-435b-9289-5a5a71765f8a","Type":"ContainerStarted","Data":"313c17328399349bff099e8ce44a12bfc674d1212f605b60c5012d829096690d"} Dec 05 09:07:48 crc kubenswrapper[4997]: I1205 09:07:48.119266 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:48 crc kubenswrapper[4997]: I1205 09:07:48.151788 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" podStartSLOduration=3.151765334 podStartE2EDuration="3.151765334s" podCreationTimestamp="2025-12-05 09:07:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:07:48.1438093 +0000 UTC m=+7968.672716581" watchObservedRunningTime="2025-12-05 09:07:48.151765334 +0000 UTC m=+7968.680672595" Dec 05 09:07:50 crc kubenswrapper[4997]: I1205 09:07:50.055604 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6vfwt"] Dec 05 09:07:50 crc kubenswrapper[4997]: I1205 09:07:50.068902 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6vfwt"] Dec 05 09:07:51 crc kubenswrapper[4997]: I1205 09:07:51.047436 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-n4vrb"] Dec 05 09:07:51 crc kubenswrapper[4997]: I1205 09:07:51.057271 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-n4vrb"] Dec 05 09:07:51 crc kubenswrapper[4997]: I1205 09:07:51.760174 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3" path="/var/lib/kubelet/pods/6cd23ff3-7420-4d8b-99ca-8a4eaadcf9b3/volumes" Dec 05 09:07:51 crc kubenswrapper[4997]: I1205 09:07:51.760787 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b195cd20-21ad-4b84-b5dd-911265d4a823" path="/var/lib/kubelet/pods/b195cd20-21ad-4b84-b5dd-911265d4a823/volumes" Dec 05 09:07:55 crc kubenswrapper[4997]: I1205 09:07:55.734831 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78dc8dd569-nzqx8" Dec 05 09:07:55 crc kubenswrapper[4997]: I1205 09:07:55.832352 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-ff67498f5-cvdxb"] Dec 05 09:07:55 crc kubenswrapper[4997]: I1205 09:07:55.832742 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" podUID="c8ef7156-24ee-4b0a-b6ad-662bc4c30243" containerName="dnsmasq-dns" containerID="cri-o://e6144019c8587fd424380aaaf0b68b82c95b1e804c934f8432dffe76843abe39" gracePeriod=10 Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.215968 4997 generic.go:334] "Generic (PLEG): container finished" podID="c8ef7156-24ee-4b0a-b6ad-662bc4c30243" containerID="e6144019c8587fd424380aaaf0b68b82c95b1e804c934f8432dffe76843abe39" exitCode=0 Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.216120 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" event={"ID":"c8ef7156-24ee-4b0a-b6ad-662bc4c30243","Type":"ContainerDied","Data":"e6144019c8587fd424380aaaf0b68b82c95b1e804c934f8432dffe76843abe39"} Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.372024 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.517773 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-config\") pod \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.518137 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-ovsdbserver-sb\") pod \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.518275 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-openstack-cell1\") pod \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.518385 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jz56\" (UniqueName: \"kubernetes.io/projected/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-kube-api-access-8jz56\") pod \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.518511 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-ovsdbserver-nb\") pod \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.518588 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-dns-svc\") pod \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\" (UID: \"c8ef7156-24ee-4b0a-b6ad-662bc4c30243\") " Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.577949 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-kube-api-access-8jz56" (OuterVolumeSpecName: "kube-api-access-8jz56") pod "c8ef7156-24ee-4b0a-b6ad-662bc4c30243" (UID: "c8ef7156-24ee-4b0a-b6ad-662bc4c30243"). InnerVolumeSpecName "kube-api-access-8jz56". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.623270 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jz56\" (UniqueName: \"kubernetes.io/projected/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-kube-api-access-8jz56\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.661808 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c8ef7156-24ee-4b0a-b6ad-662bc4c30243" (UID: "c8ef7156-24ee-4b0a-b6ad-662bc4c30243"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.677090 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c8ef7156-24ee-4b0a-b6ad-662bc4c30243" (UID: "c8ef7156-24ee-4b0a-b6ad-662bc4c30243"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.684964 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-config" (OuterVolumeSpecName: "config") pod "c8ef7156-24ee-4b0a-b6ad-662bc4c30243" (UID: "c8ef7156-24ee-4b0a-b6ad-662bc4c30243"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.692460 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "c8ef7156-24ee-4b0a-b6ad-662bc4c30243" (UID: "c8ef7156-24ee-4b0a-b6ad-662bc4c30243"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.695036 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c8ef7156-24ee-4b0a-b6ad-662bc4c30243" (UID: "c8ef7156-24ee-4b0a-b6ad-662bc4c30243"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.725414 4997 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-openstack-cell1\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.725454 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.725465 4997 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.725473 4997 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:56 crc kubenswrapper[4997]: I1205 09:07:56.725484 4997 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8ef7156-24ee-4b0a-b6ad-662bc4c30243-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:57 crc kubenswrapper[4997]: I1205 09:07:57.228188 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" event={"ID":"c8ef7156-24ee-4b0a-b6ad-662bc4c30243","Type":"ContainerDied","Data":"12e0c7594832dc667149b0b7171e7613a8d3dc5ab972fde63ada0807cbe5512b"} Dec 05 09:07:57 crc kubenswrapper[4997]: I1205 09:07:57.228253 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-ff67498f5-cvdxb" Dec 05 09:07:57 crc kubenswrapper[4997]: I1205 09:07:57.228534 4997 scope.go:117] "RemoveContainer" containerID="e6144019c8587fd424380aaaf0b68b82c95b1e804c934f8432dffe76843abe39" Dec 05 09:07:57 crc kubenswrapper[4997]: I1205 09:07:57.252350 4997 scope.go:117] "RemoveContainer" containerID="1d657bd903cedb7de10b0cc1640b6914a19b9a6629ed1efc6c7184162484beda" Dec 05 09:07:57 crc kubenswrapper[4997]: I1205 09:07:57.269185 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-ff67498f5-cvdxb"] Dec 05 09:07:57 crc kubenswrapper[4997]: I1205 09:07:57.339060 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-ff67498f5-cvdxb"] Dec 05 09:07:57 crc kubenswrapper[4997]: I1205 09:07:57.760108 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8ef7156-24ee-4b0a-b6ad-662bc4c30243" path="/var/lib/kubelet/pods/c8ef7156-24ee-4b0a-b6ad-662bc4c30243/volumes" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.700319 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m"] Dec 05 09:08:06 crc kubenswrapper[4997]: E1205 09:08:06.701346 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ef7156-24ee-4b0a-b6ad-662bc4c30243" containerName="dnsmasq-dns" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.701360 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ef7156-24ee-4b0a-b6ad-662bc4c30243" containerName="dnsmasq-dns" Dec 05 09:08:06 crc kubenswrapper[4997]: E1205 09:08:06.701377 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4913c005-1117-40c8-9916-aecc8625d1ed" containerName="init" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.701383 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4913c005-1117-40c8-9916-aecc8625d1ed" containerName="init" Dec 05 09:08:06 crc kubenswrapper[4997]: E1205 09:08:06.701398 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4913c005-1117-40c8-9916-aecc8625d1ed" containerName="dnsmasq-dns" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.701404 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4913c005-1117-40c8-9916-aecc8625d1ed" containerName="dnsmasq-dns" Dec 05 09:08:06 crc kubenswrapper[4997]: E1205 09:08:06.701425 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ef7156-24ee-4b0a-b6ad-662bc4c30243" containerName="init" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.701431 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ef7156-24ee-4b0a-b6ad-662bc4c30243" containerName="init" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.701655 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8ef7156-24ee-4b0a-b6ad-662bc4c30243" containerName="dnsmasq-dns" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.701671 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="4913c005-1117-40c8-9916-aecc8625d1ed" containerName="dnsmasq-dns" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.702453 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.705746 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.705878 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.705882 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.705981 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.710429 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n"] Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.712360 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.716051 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-4542r" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.716514 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.717783 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n"] Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.728336 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m"] Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.855900 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.856307 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n\" (UID: \"a364f0ce-49a9-4284-982f-d303bddb5685\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.856375 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.856587 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.857606 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n\" (UID: \"a364f0ce-49a9-4284-982f-d303bddb5685\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.857693 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.857726 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxdw4\" (UniqueName: \"kubernetes.io/projected/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-kube-api-access-wxdw4\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.857816 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n\" (UID: \"a364f0ce-49a9-4284-982f-d303bddb5685\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.857956 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxqwm\" (UniqueName: \"kubernetes.io/projected/a364f0ce-49a9-4284-982f-d303bddb5685-kube-api-access-lxqwm\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n\" (UID: \"a364f0ce-49a9-4284-982f-d303bddb5685\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.959775 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n\" (UID: \"a364f0ce-49a9-4284-982f-d303bddb5685\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.959815 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.959839 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxdw4\" (UniqueName: \"kubernetes.io/projected/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-kube-api-access-wxdw4\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.959869 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n\" (UID: \"a364f0ce-49a9-4284-982f-d303bddb5685\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.959919 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxqwm\" (UniqueName: \"kubernetes.io/projected/a364f0ce-49a9-4284-982f-d303bddb5685-kube-api-access-lxqwm\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n\" (UID: \"a364f0ce-49a9-4284-982f-d303bddb5685\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.959943 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.959981 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n\" (UID: \"a364f0ce-49a9-4284-982f-d303bddb5685\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.960043 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.960101 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.965965 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.966573 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n\" (UID: \"a364f0ce-49a9-4284-982f-d303bddb5685\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.967426 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.968028 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.970650 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.977679 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n\" (UID: \"a364f0ce-49a9-4284-982f-d303bddb5685\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.978069 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n\" (UID: \"a364f0ce-49a9-4284-982f-d303bddb5685\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.980519 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxdw4\" (UniqueName: \"kubernetes.io/projected/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-kube-api-access-wxdw4\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:06 crc kubenswrapper[4997]: I1205 09:08:06.992399 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxqwm\" (UniqueName: \"kubernetes.io/projected/a364f0ce-49a9-4284-982f-d303bddb5685-kube-api-access-lxqwm\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n\" (UID: \"a364f0ce-49a9-4284-982f-d303bddb5685\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" Dec 05 09:08:07 crc kubenswrapper[4997]: I1205 09:08:07.038805 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:07 crc kubenswrapper[4997]: I1205 09:08:07.060464 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" Dec 05 09:08:07 crc kubenswrapper[4997]: I1205 09:08:07.832486 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n"] Dec 05 09:08:07 crc kubenswrapper[4997]: W1205 09:08:07.840747 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda364f0ce_49a9_4284_982f_d303bddb5685.slice/crio-9be0c38deca9faec6617e1e46867c94cd05a44c86e05fd286a761d6ea718acfa WatchSource:0}: Error finding container 9be0c38deca9faec6617e1e46867c94cd05a44c86e05fd286a761d6ea718acfa: Status 404 returned error can't find the container with id 9be0c38deca9faec6617e1e46867c94cd05a44c86e05fd286a761d6ea718acfa Dec 05 09:08:08 crc kubenswrapper[4997]: I1205 09:08:08.371717 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" event={"ID":"a364f0ce-49a9-4284-982f-d303bddb5685","Type":"ContainerStarted","Data":"9be0c38deca9faec6617e1e46867c94cd05a44c86e05fd286a761d6ea718acfa"} Dec 05 09:08:08 crc kubenswrapper[4997]: I1205 09:08:08.457363 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m"] Dec 05 09:08:09 crc kubenswrapper[4997]: I1205 09:08:09.050413 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-qb474"] Dec 05 09:08:09 crc kubenswrapper[4997]: I1205 09:08:09.063049 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-qb474"] Dec 05 09:08:09 crc kubenswrapper[4997]: I1205 09:08:09.383560 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" event={"ID":"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf","Type":"ContainerStarted","Data":"5f0015835f8ae98efc7d90a83645e435dd4253be37bba57f0c91e0011e5d8462"} Dec 05 09:08:09 crc kubenswrapper[4997]: I1205 09:08:09.764223 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b8eda8f-e194-47b6-a429-07fbd87596ae" path="/var/lib/kubelet/pods/6b8eda8f-e194-47b6-a429-07fbd87596ae/volumes" Dec 05 09:08:17 crc kubenswrapper[4997]: I1205 09:08:17.471838 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" event={"ID":"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf","Type":"ContainerStarted","Data":"910d70106f4a2b169f2e3ccbcaf975cf37745981a181401f06571784e086ea91"} Dec 05 09:08:17 crc kubenswrapper[4997]: I1205 09:08:17.476164 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" event={"ID":"a364f0ce-49a9-4284-982f-d303bddb5685","Type":"ContainerStarted","Data":"36233af8ad06d0827e2b2e4006b6bdf94c2ea461046feabc89d663310801627b"} Dec 05 09:08:17 crc kubenswrapper[4997]: I1205 09:08:17.497911 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" podStartSLOduration=3.10939657 podStartE2EDuration="11.497877537s" podCreationTimestamp="2025-12-05 09:08:06 +0000 UTC" firstStartedPulling="2025-12-05 09:08:08.463537739 +0000 UTC m=+7988.992445000" lastFinishedPulling="2025-12-05 09:08:16.852018686 +0000 UTC m=+7997.380925967" observedRunningTime="2025-12-05 09:08:17.492574883 +0000 UTC m=+7998.021482144" watchObservedRunningTime="2025-12-05 09:08:17.497877537 +0000 UTC m=+7998.026784798" Dec 05 09:08:17 crc kubenswrapper[4997]: I1205 09:08:17.516475 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" podStartSLOduration=2.527102174 podStartE2EDuration="11.516447577s" podCreationTimestamp="2025-12-05 09:08:06 +0000 UTC" firstStartedPulling="2025-12-05 09:08:07.847044512 +0000 UTC m=+7988.375951773" lastFinishedPulling="2025-12-05 09:08:16.836389915 +0000 UTC m=+7997.365297176" observedRunningTime="2025-12-05 09:08:17.510499337 +0000 UTC m=+7998.039406638" watchObservedRunningTime="2025-12-05 09:08:17.516447577 +0000 UTC m=+7998.045354878" Dec 05 09:08:18 crc kubenswrapper[4997]: I1205 09:08:18.347728 4997 scope.go:117] "RemoveContainer" containerID="bb93c1917706fc42241924f8d7bd1c0dcd0b6e200158acbeadac262575c2fcef" Dec 05 09:08:18 crc kubenswrapper[4997]: I1205 09:08:18.400018 4997 scope.go:117] "RemoveContainer" containerID="0fdc373de62acc63a374a7697504a0e6168c81577a14a035e7b3a0b2c3613768" Dec 05 09:08:18 crc kubenswrapper[4997]: I1205 09:08:18.439904 4997 scope.go:117] "RemoveContainer" containerID="971f750d4d3626d8b4a1c5eb133f6f45e6bae2d49b9d31f1f3b26d326de5403f" Dec 05 09:08:18 crc kubenswrapper[4997]: I1205 09:08:18.486168 4997 scope.go:117] "RemoveContainer" containerID="b978475c065f58bcfa921adf4ea2d5e99a71fadae7eccc3992831e6da1066e04" Dec 05 09:08:28 crc kubenswrapper[4997]: I1205 09:08:28.598321 4997 generic.go:334] "Generic (PLEG): container finished" podID="a364f0ce-49a9-4284-982f-d303bddb5685" containerID="36233af8ad06d0827e2b2e4006b6bdf94c2ea461046feabc89d663310801627b" exitCode=0 Dec 05 09:08:28 crc kubenswrapper[4997]: I1205 09:08:28.598418 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" event={"ID":"a364f0ce-49a9-4284-982f-d303bddb5685","Type":"ContainerDied","Data":"36233af8ad06d0827e2b2e4006b6bdf94c2ea461046feabc89d663310801627b"} Dec 05 09:08:29 crc kubenswrapper[4997]: I1205 09:08:29.610496 4997 generic.go:334] "Generic (PLEG): container finished" podID="10a290ee-4d1b-4e9d-b747-9a4e9698a0cf" containerID="910d70106f4a2b169f2e3ccbcaf975cf37745981a181401f06571784e086ea91" exitCode=0 Dec 05 09:08:29 crc kubenswrapper[4997]: I1205 09:08:29.610595 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" event={"ID":"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf","Type":"ContainerDied","Data":"910d70106f4a2b169f2e3ccbcaf975cf37745981a181401f06571784e086ea91"} Dec 05 09:08:30 crc kubenswrapper[4997]: I1205 09:08:30.106224 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" Dec 05 09:08:30 crc kubenswrapper[4997]: I1205 09:08:30.178969 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxqwm\" (UniqueName: \"kubernetes.io/projected/a364f0ce-49a9-4284-982f-d303bddb5685-kube-api-access-lxqwm\") pod \"a364f0ce-49a9-4284-982f-d303bddb5685\" (UID: \"a364f0ce-49a9-4284-982f-d303bddb5685\") " Dec 05 09:08:30 crc kubenswrapper[4997]: I1205 09:08:30.179064 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-pre-adoption-validation-combined-ca-bundle\") pod \"a364f0ce-49a9-4284-982f-d303bddb5685\" (UID: \"a364f0ce-49a9-4284-982f-d303bddb5685\") " Dec 05 09:08:30 crc kubenswrapper[4997]: I1205 09:08:30.179377 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-inventory\") pod \"a364f0ce-49a9-4284-982f-d303bddb5685\" (UID: \"a364f0ce-49a9-4284-982f-d303bddb5685\") " Dec 05 09:08:30 crc kubenswrapper[4997]: I1205 09:08:30.179519 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-ssh-key\") pod \"a364f0ce-49a9-4284-982f-d303bddb5685\" (UID: \"a364f0ce-49a9-4284-982f-d303bddb5685\") " Dec 05 09:08:30 crc kubenswrapper[4997]: I1205 09:08:30.185046 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "a364f0ce-49a9-4284-982f-d303bddb5685" (UID: "a364f0ce-49a9-4284-982f-d303bddb5685"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:08:30 crc kubenswrapper[4997]: I1205 09:08:30.185209 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a364f0ce-49a9-4284-982f-d303bddb5685-kube-api-access-lxqwm" (OuterVolumeSpecName: "kube-api-access-lxqwm") pod "a364f0ce-49a9-4284-982f-d303bddb5685" (UID: "a364f0ce-49a9-4284-982f-d303bddb5685"). InnerVolumeSpecName "kube-api-access-lxqwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:08:30 crc kubenswrapper[4997]: I1205 09:08:30.213655 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-inventory" (OuterVolumeSpecName: "inventory") pod "a364f0ce-49a9-4284-982f-d303bddb5685" (UID: "a364f0ce-49a9-4284-982f-d303bddb5685"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:08:30 crc kubenswrapper[4997]: I1205 09:08:30.216461 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a364f0ce-49a9-4284-982f-d303bddb5685" (UID: "a364f0ce-49a9-4284-982f-d303bddb5685"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:08:30 crc kubenswrapper[4997]: I1205 09:08:30.282027 4997 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:08:30 crc kubenswrapper[4997]: I1205 09:08:30.282071 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:08:30 crc kubenswrapper[4997]: I1205 09:08:30.282084 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a364f0ce-49a9-4284-982f-d303bddb5685-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:08:30 crc kubenswrapper[4997]: I1205 09:08:30.282155 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxqwm\" (UniqueName: \"kubernetes.io/projected/a364f0ce-49a9-4284-982f-d303bddb5685-kube-api-access-lxqwm\") on node \"crc\" DevicePath \"\"" Dec 05 09:08:30 crc kubenswrapper[4997]: I1205 09:08:30.627292 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" Dec 05 09:08:30 crc kubenswrapper[4997]: I1205 09:08:30.627294 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n" event={"ID":"a364f0ce-49a9-4284-982f-d303bddb5685","Type":"ContainerDied","Data":"9be0c38deca9faec6617e1e46867c94cd05a44c86e05fd286a761d6ea718acfa"} Dec 05 09:08:30 crc kubenswrapper[4997]: I1205 09:08:30.627354 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9be0c38deca9faec6617e1e46867c94cd05a44c86e05fd286a761d6ea718acfa" Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.096800 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.203587 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-pre-adoption-validation-combined-ca-bundle\") pod \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.203747 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-inventory\") pod \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.203779 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-ceph\") pod \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.203891 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxdw4\" (UniqueName: \"kubernetes.io/projected/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-kube-api-access-wxdw4\") pod \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.203990 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-ssh-key\") pod \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\" (UID: \"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf\") " Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.208821 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-kube-api-access-wxdw4" (OuterVolumeSpecName: "kube-api-access-wxdw4") pod "10a290ee-4d1b-4e9d-b747-9a4e9698a0cf" (UID: "10a290ee-4d1b-4e9d-b747-9a4e9698a0cf"). InnerVolumeSpecName "kube-api-access-wxdw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.209264 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-ceph" (OuterVolumeSpecName: "ceph") pod "10a290ee-4d1b-4e9d-b747-9a4e9698a0cf" (UID: "10a290ee-4d1b-4e9d-b747-9a4e9698a0cf"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.211888 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "10a290ee-4d1b-4e9d-b747-9a4e9698a0cf" (UID: "10a290ee-4d1b-4e9d-b747-9a4e9698a0cf"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.236037 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "10a290ee-4d1b-4e9d-b747-9a4e9698a0cf" (UID: "10a290ee-4d1b-4e9d-b747-9a4e9698a0cf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.242636 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-inventory" (OuterVolumeSpecName: "inventory") pod "10a290ee-4d1b-4e9d-b747-9a4e9698a0cf" (UID: "10a290ee-4d1b-4e9d-b747-9a4e9698a0cf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.306578 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxdw4\" (UniqueName: \"kubernetes.io/projected/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-kube-api-access-wxdw4\") on node \"crc\" DevicePath \"\"" Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.306633 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.306650 4997 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.306662 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.306674 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/10a290ee-4d1b-4e9d-b747-9a4e9698a0cf-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.637785 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" event={"ID":"10a290ee-4d1b-4e9d-b747-9a4e9698a0cf","Type":"ContainerDied","Data":"5f0015835f8ae98efc7d90a83645e435dd4253be37bba57f0c91e0011e5d8462"} Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.638431 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f0015835f8ae98efc7d90a83645e435dd4253be37bba57f0c91e0011e5d8462" Dec 05 09:08:31 crc kubenswrapper[4997]: I1205 09:08:31.637816 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.534886 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx"] Dec 05 09:08:39 crc kubenswrapper[4997]: E1205 09:08:39.535977 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a290ee-4d1b-4e9d-b747-9a4e9698a0cf" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.536005 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a290ee-4d1b-4e9d-b747-9a4e9698a0cf" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 05 09:08:39 crc kubenswrapper[4997]: E1205 09:08:39.536036 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a364f0ce-49a9-4284-982f-d303bddb5685" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-networ" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.536046 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a364f0ce-49a9-4284-982f-d303bddb5685" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-networ" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.536348 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a290ee-4d1b-4e9d-b747-9a4e9698a0cf" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.536386 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a364f0ce-49a9-4284-982f-d303bddb5685" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-networ" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.537392 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.539424 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.539603 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.539852 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.542132 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.546048 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2"] Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.548327 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.551555 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-4542r" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.551837 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.558075 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2"] Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.568970 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx"] Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.663101 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.663174 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.663201 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2\" (UID: \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.663241 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dppw8\" (UniqueName: \"kubernetes.io/projected/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-kube-api-access-dppw8\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2\" (UID: \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.663276 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2\" (UID: \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.663362 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2\" (UID: \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.663416 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.663437 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmfsw\" (UniqueName: \"kubernetes.io/projected/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-kube-api-access-tmfsw\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.663503 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.765332 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.765372 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2\" (UID: \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.765417 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dppw8\" (UniqueName: \"kubernetes.io/projected/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-kube-api-access-dppw8\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2\" (UID: \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.765450 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2\" (UID: \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.765478 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2\" (UID: \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.765526 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.765555 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmfsw\" (UniqueName: \"kubernetes.io/projected/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-kube-api-access-tmfsw\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.766365 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.766422 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.771600 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.771661 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2\" (UID: \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.771724 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2\" (UID: \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.772185 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2\" (UID: \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.772361 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.772601 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.772641 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.783234 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmfsw\" (UniqueName: \"kubernetes.io/projected/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-kube-api-access-tmfsw\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.783751 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dppw8\" (UniqueName: \"kubernetes.io/projected/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-kube-api-access-dppw8\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2\" (UID: \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.866795 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:08:39 crc kubenswrapper[4997]: I1205 09:08:39.882148 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" Dec 05 09:08:40 crc kubenswrapper[4997]: I1205 09:08:40.428092 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx"] Dec 05 09:08:40 crc kubenswrapper[4997]: I1205 09:08:40.530101 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2"] Dec 05 09:08:40 crc kubenswrapper[4997]: W1205 09:08:40.531101 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d475848_9ccb_482d_aa85_f6c9e3ef4ef8.slice/crio-7c8ab7034aaa3716351c07b2404b76c938fd4188add4026ef69d14740d271299 WatchSource:0}: Error finding container 7c8ab7034aaa3716351c07b2404b76c938fd4188add4026ef69d14740d271299: Status 404 returned error can't find the container with id 7c8ab7034aaa3716351c07b2404b76c938fd4188add4026ef69d14740d271299 Dec 05 09:08:40 crc kubenswrapper[4997]: I1205 09:08:40.738989 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" event={"ID":"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8","Type":"ContainerStarted","Data":"7c8ab7034aaa3716351c07b2404b76c938fd4188add4026ef69d14740d271299"} Dec 05 09:08:40 crc kubenswrapper[4997]: I1205 09:08:40.741068 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" event={"ID":"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d","Type":"ContainerStarted","Data":"34eb374ce6f81e95f6fa93ebfed5616d5420e98c8a1f2c3c8e8027af1a6958d3"} Dec 05 09:08:41 crc kubenswrapper[4997]: I1205 09:08:41.770268 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" event={"ID":"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d","Type":"ContainerStarted","Data":"0b6c53e049496e78722cd4d19b8036b4bd04265cb5d20c6bf52f43652381f5fa"} Dec 05 09:08:41 crc kubenswrapper[4997]: I1205 09:08:41.770888 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" event={"ID":"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8","Type":"ContainerStarted","Data":"caafb1fe21eae472a1b02c1bb8ef06b4c5c4338a838ea8ddbd2e39c99c05c212"} Dec 05 09:08:41 crc kubenswrapper[4997]: I1205 09:08:41.784747 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" podStartSLOduration=2.355817316 podStartE2EDuration="2.784718594s" podCreationTimestamp="2025-12-05 09:08:39 +0000 UTC" firstStartedPulling="2025-12-05 09:08:40.441046303 +0000 UTC m=+8020.969953574" lastFinishedPulling="2025-12-05 09:08:40.869947581 +0000 UTC m=+8021.398854852" observedRunningTime="2025-12-05 09:08:41.773360708 +0000 UTC m=+8022.302268019" watchObservedRunningTime="2025-12-05 09:08:41.784718594 +0000 UTC m=+8022.313625885" Dec 05 09:08:41 crc kubenswrapper[4997]: I1205 09:08:41.826125 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" podStartSLOduration=2.368332183 podStartE2EDuration="2.82610106s" podCreationTimestamp="2025-12-05 09:08:39 +0000 UTC" firstStartedPulling="2025-12-05 09:08:40.535877451 +0000 UTC m=+8021.064784712" lastFinishedPulling="2025-12-05 09:08:40.993646328 +0000 UTC m=+8021.522553589" observedRunningTime="2025-12-05 09:08:41.802922905 +0000 UTC m=+8022.331830176" watchObservedRunningTime="2025-12-05 09:08:41.82610106 +0000 UTC m=+8022.355008321" Dec 05 09:08:51 crc kubenswrapper[4997]: I1205 09:08:51.048045 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-0f52-account-create-update-glgr7"] Dec 05 09:08:51 crc kubenswrapper[4997]: I1205 09:08:51.059799 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-wczzn"] Dec 05 09:08:51 crc kubenswrapper[4997]: I1205 09:08:51.078740 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-0f52-account-create-update-glgr7"] Dec 05 09:08:51 crc kubenswrapper[4997]: I1205 09:08:51.091231 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-wczzn"] Dec 05 09:08:51 crc kubenswrapper[4997]: I1205 09:08:51.763228 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db566dd6-c449-499f-9bf3-457a8657e5da" path="/var/lib/kubelet/pods/db566dd6-c449-499f-9bf3-457a8657e5da/volumes" Dec 05 09:08:51 crc kubenswrapper[4997]: I1205 09:08:51.764205 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc7d296f-ae2b-4178-abdb-135f3e239e72" path="/var/lib/kubelet/pods/fc7d296f-ae2b-4178-abdb-135f3e239e72/volumes" Dec 05 09:09:16 crc kubenswrapper[4997]: I1205 09:09:16.059665 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-8vfdh"] Dec 05 09:09:16 crc kubenswrapper[4997]: I1205 09:09:16.069159 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-8vfdh"] Dec 05 09:09:17 crc kubenswrapper[4997]: I1205 09:09:17.760323 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdc0b8fc-c988-410a-9b97-d80cb48754a2" path="/var/lib/kubelet/pods/bdc0b8fc-c988-410a-9b97-d80cb48754a2/volumes" Dec 05 09:09:18 crc kubenswrapper[4997]: I1205 09:09:18.683025 4997 scope.go:117] "RemoveContainer" containerID="f322072eb980292b4c462cf9b6fe234392255c65125ec6d0c4ea92760528fbe2" Dec 05 09:09:18 crc kubenswrapper[4997]: I1205 09:09:18.713012 4997 scope.go:117] "RemoveContainer" containerID="2f2cb21e9120ff0d3ccb1e72be0184d31016379bca916dd03c6f31891722c88b" Dec 05 09:09:18 crc kubenswrapper[4997]: I1205 09:09:18.787847 4997 scope.go:117] "RemoveContainer" containerID="76f07e30bd6fec0c5ecea375ce071e42b3293940713fe10e1d1753e1585ba432" Dec 05 09:09:49 crc kubenswrapper[4997]: I1205 09:09:49.770370 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:09:49 crc kubenswrapper[4997]: I1205 09:09:49.771157 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:10:03 crc kubenswrapper[4997]: I1205 09:10:03.426688 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nn5jp"] Dec 05 09:10:03 crc kubenswrapper[4997]: I1205 09:10:03.464219 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nn5jp" Dec 05 09:10:03 crc kubenswrapper[4997]: I1205 09:10:03.467794 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nn5jp"] Dec 05 09:10:03 crc kubenswrapper[4997]: I1205 09:10:03.553105 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgkdb\" (UniqueName: \"kubernetes.io/projected/d228c4a4-4656-4624-af35-94bf3db4c79e-kube-api-access-qgkdb\") pod \"community-operators-nn5jp\" (UID: \"d228c4a4-4656-4624-af35-94bf3db4c79e\") " pod="openshift-marketplace/community-operators-nn5jp" Dec 05 09:10:03 crc kubenswrapper[4997]: I1205 09:10:03.553293 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d228c4a4-4656-4624-af35-94bf3db4c79e-utilities\") pod \"community-operators-nn5jp\" (UID: \"d228c4a4-4656-4624-af35-94bf3db4c79e\") " pod="openshift-marketplace/community-operators-nn5jp" Dec 05 09:10:03 crc kubenswrapper[4997]: I1205 09:10:03.553381 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d228c4a4-4656-4624-af35-94bf3db4c79e-catalog-content\") pod \"community-operators-nn5jp\" (UID: \"d228c4a4-4656-4624-af35-94bf3db4c79e\") " pod="openshift-marketplace/community-operators-nn5jp" Dec 05 09:10:03 crc kubenswrapper[4997]: I1205 09:10:03.655356 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d228c4a4-4656-4624-af35-94bf3db4c79e-utilities\") pod \"community-operators-nn5jp\" (UID: \"d228c4a4-4656-4624-af35-94bf3db4c79e\") " pod="openshift-marketplace/community-operators-nn5jp" Dec 05 09:10:03 crc kubenswrapper[4997]: I1205 09:10:03.655452 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d228c4a4-4656-4624-af35-94bf3db4c79e-catalog-content\") pod \"community-operators-nn5jp\" (UID: \"d228c4a4-4656-4624-af35-94bf3db4c79e\") " pod="openshift-marketplace/community-operators-nn5jp" Dec 05 09:10:03 crc kubenswrapper[4997]: I1205 09:10:03.655519 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgkdb\" (UniqueName: \"kubernetes.io/projected/d228c4a4-4656-4624-af35-94bf3db4c79e-kube-api-access-qgkdb\") pod \"community-operators-nn5jp\" (UID: \"d228c4a4-4656-4624-af35-94bf3db4c79e\") " pod="openshift-marketplace/community-operators-nn5jp" Dec 05 09:10:03 crc kubenswrapper[4997]: I1205 09:10:03.656094 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d228c4a4-4656-4624-af35-94bf3db4c79e-utilities\") pod \"community-operators-nn5jp\" (UID: \"d228c4a4-4656-4624-af35-94bf3db4c79e\") " pod="openshift-marketplace/community-operators-nn5jp" Dec 05 09:10:03 crc kubenswrapper[4997]: I1205 09:10:03.656193 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d228c4a4-4656-4624-af35-94bf3db4c79e-catalog-content\") pod \"community-operators-nn5jp\" (UID: \"d228c4a4-4656-4624-af35-94bf3db4c79e\") " pod="openshift-marketplace/community-operators-nn5jp" Dec 05 09:10:03 crc kubenswrapper[4997]: I1205 09:10:03.678737 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgkdb\" (UniqueName: \"kubernetes.io/projected/d228c4a4-4656-4624-af35-94bf3db4c79e-kube-api-access-qgkdb\") pod \"community-operators-nn5jp\" (UID: \"d228c4a4-4656-4624-af35-94bf3db4c79e\") " pod="openshift-marketplace/community-operators-nn5jp" Dec 05 09:10:03 crc kubenswrapper[4997]: I1205 09:10:03.793125 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nn5jp" Dec 05 09:10:04 crc kubenswrapper[4997]: I1205 09:10:04.372686 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nn5jp"] Dec 05 09:10:04 crc kubenswrapper[4997]: I1205 09:10:04.629826 4997 generic.go:334] "Generic (PLEG): container finished" podID="d228c4a4-4656-4624-af35-94bf3db4c79e" containerID="5bf81c2a914c7bef8709c0ba9ef00d74fa4d0c0a2189f2b64d949558d178d169" exitCode=0 Dec 05 09:10:04 crc kubenswrapper[4997]: I1205 09:10:04.629873 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nn5jp" event={"ID":"d228c4a4-4656-4624-af35-94bf3db4c79e","Type":"ContainerDied","Data":"5bf81c2a914c7bef8709c0ba9ef00d74fa4d0c0a2189f2b64d949558d178d169"} Dec 05 09:10:04 crc kubenswrapper[4997]: I1205 09:10:04.629901 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nn5jp" event={"ID":"d228c4a4-4656-4624-af35-94bf3db4c79e","Type":"ContainerStarted","Data":"c1c2f105a2d11ae4eaf38d137272bfb9054c3b972554e8e568d630cb7c3b9e64"} Dec 05 09:10:04 crc kubenswrapper[4997]: I1205 09:10:04.632926 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:10:05 crc kubenswrapper[4997]: I1205 09:10:05.217837 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nq77q"] Dec 05 09:10:05 crc kubenswrapper[4997]: I1205 09:10:05.220445 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nq77q" Dec 05 09:10:05 crc kubenswrapper[4997]: I1205 09:10:05.239799 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nq77q"] Dec 05 09:10:05 crc kubenswrapper[4997]: I1205 09:10:05.391383 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5bfa50-e54b-490a-a09d-9632b63d9362-catalog-content\") pod \"redhat-marketplace-nq77q\" (UID: \"3c5bfa50-e54b-490a-a09d-9632b63d9362\") " pod="openshift-marketplace/redhat-marketplace-nq77q" Dec 05 09:10:05 crc kubenswrapper[4997]: I1205 09:10:05.391583 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgbns\" (UniqueName: \"kubernetes.io/projected/3c5bfa50-e54b-490a-a09d-9632b63d9362-kube-api-access-cgbns\") pod \"redhat-marketplace-nq77q\" (UID: \"3c5bfa50-e54b-490a-a09d-9632b63d9362\") " pod="openshift-marketplace/redhat-marketplace-nq77q" Dec 05 09:10:05 crc kubenswrapper[4997]: I1205 09:10:05.391664 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5bfa50-e54b-490a-a09d-9632b63d9362-utilities\") pod \"redhat-marketplace-nq77q\" (UID: \"3c5bfa50-e54b-490a-a09d-9632b63d9362\") " pod="openshift-marketplace/redhat-marketplace-nq77q" Dec 05 09:10:05 crc kubenswrapper[4997]: I1205 09:10:05.493951 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgbns\" (UniqueName: \"kubernetes.io/projected/3c5bfa50-e54b-490a-a09d-9632b63d9362-kube-api-access-cgbns\") pod \"redhat-marketplace-nq77q\" (UID: \"3c5bfa50-e54b-490a-a09d-9632b63d9362\") " pod="openshift-marketplace/redhat-marketplace-nq77q" Dec 05 09:10:05 crc kubenswrapper[4997]: I1205 09:10:05.494048 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5bfa50-e54b-490a-a09d-9632b63d9362-utilities\") pod \"redhat-marketplace-nq77q\" (UID: \"3c5bfa50-e54b-490a-a09d-9632b63d9362\") " pod="openshift-marketplace/redhat-marketplace-nq77q" Dec 05 09:10:05 crc kubenswrapper[4997]: I1205 09:10:05.494110 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5bfa50-e54b-490a-a09d-9632b63d9362-catalog-content\") pod \"redhat-marketplace-nq77q\" (UID: \"3c5bfa50-e54b-490a-a09d-9632b63d9362\") " pod="openshift-marketplace/redhat-marketplace-nq77q" Dec 05 09:10:05 crc kubenswrapper[4997]: I1205 09:10:05.494683 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5bfa50-e54b-490a-a09d-9632b63d9362-catalog-content\") pod \"redhat-marketplace-nq77q\" (UID: \"3c5bfa50-e54b-490a-a09d-9632b63d9362\") " pod="openshift-marketplace/redhat-marketplace-nq77q" Dec 05 09:10:05 crc kubenswrapper[4997]: I1205 09:10:05.494865 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5bfa50-e54b-490a-a09d-9632b63d9362-utilities\") pod \"redhat-marketplace-nq77q\" (UID: \"3c5bfa50-e54b-490a-a09d-9632b63d9362\") " pod="openshift-marketplace/redhat-marketplace-nq77q" Dec 05 09:10:05 crc kubenswrapper[4997]: I1205 09:10:05.517491 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgbns\" (UniqueName: \"kubernetes.io/projected/3c5bfa50-e54b-490a-a09d-9632b63d9362-kube-api-access-cgbns\") pod \"redhat-marketplace-nq77q\" (UID: \"3c5bfa50-e54b-490a-a09d-9632b63d9362\") " pod="openshift-marketplace/redhat-marketplace-nq77q" Dec 05 09:10:05 crc kubenswrapper[4997]: I1205 09:10:05.560713 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nq77q" Dec 05 09:10:06 crc kubenswrapper[4997]: I1205 09:10:06.049251 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nq77q"] Dec 05 09:10:06 crc kubenswrapper[4997]: W1205 09:10:06.049871 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c5bfa50_e54b_490a_a09d_9632b63d9362.slice/crio-6e19e990c1f670b64a9fb606fb85ffc9133071df7e4d4aadb13a167229b0c3d0 WatchSource:0}: Error finding container 6e19e990c1f670b64a9fb606fb85ffc9133071df7e4d4aadb13a167229b0c3d0: Status 404 returned error can't find the container with id 6e19e990c1f670b64a9fb606fb85ffc9133071df7e4d4aadb13a167229b0c3d0 Dec 05 09:10:06 crc kubenswrapper[4997]: I1205 09:10:06.680211 4997 generic.go:334] "Generic (PLEG): container finished" podID="3c5bfa50-e54b-490a-a09d-9632b63d9362" containerID="4aa190137da47dace203d8b24e7d72c94d5d3d4f49a7d83a7e3d6d42210e544c" exitCode=0 Dec 05 09:10:06 crc kubenswrapper[4997]: I1205 09:10:06.680325 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nq77q" event={"ID":"3c5bfa50-e54b-490a-a09d-9632b63d9362","Type":"ContainerDied","Data":"4aa190137da47dace203d8b24e7d72c94d5d3d4f49a7d83a7e3d6d42210e544c"} Dec 05 09:10:06 crc kubenswrapper[4997]: I1205 09:10:06.680470 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nq77q" event={"ID":"3c5bfa50-e54b-490a-a09d-9632b63d9362","Type":"ContainerStarted","Data":"6e19e990c1f670b64a9fb606fb85ffc9133071df7e4d4aadb13a167229b0c3d0"} Dec 05 09:10:08 crc kubenswrapper[4997]: I1205 09:10:08.705761 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nn5jp" event={"ID":"d228c4a4-4656-4624-af35-94bf3db4c79e","Type":"ContainerStarted","Data":"d34d8c5f09b67a211664e41c6e59f7e8deb2360cb12880f98721f365fd43d391"} Dec 05 09:10:08 crc kubenswrapper[4997]: I1205 09:10:08.708687 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nq77q" event={"ID":"3c5bfa50-e54b-490a-a09d-9632b63d9362","Type":"ContainerStarted","Data":"b48ffda1358838e310bb18dc53bbf02a6d9af00638dc6ba0a47c6f95dcedf01b"} Dec 05 09:10:09 crc kubenswrapper[4997]: I1205 09:10:09.721100 4997 generic.go:334] "Generic (PLEG): container finished" podID="3c5bfa50-e54b-490a-a09d-9632b63d9362" containerID="b48ffda1358838e310bb18dc53bbf02a6d9af00638dc6ba0a47c6f95dcedf01b" exitCode=0 Dec 05 09:10:09 crc kubenswrapper[4997]: I1205 09:10:09.721205 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nq77q" event={"ID":"3c5bfa50-e54b-490a-a09d-9632b63d9362","Type":"ContainerDied","Data":"b48ffda1358838e310bb18dc53bbf02a6d9af00638dc6ba0a47c6f95dcedf01b"} Dec 05 09:10:09 crc kubenswrapper[4997]: I1205 09:10:09.723368 4997 generic.go:334] "Generic (PLEG): container finished" podID="d228c4a4-4656-4624-af35-94bf3db4c79e" containerID="d34d8c5f09b67a211664e41c6e59f7e8deb2360cb12880f98721f365fd43d391" exitCode=0 Dec 05 09:10:09 crc kubenswrapper[4997]: I1205 09:10:09.723394 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nn5jp" event={"ID":"d228c4a4-4656-4624-af35-94bf3db4c79e","Type":"ContainerDied","Data":"d34d8c5f09b67a211664e41c6e59f7e8deb2360cb12880f98721f365fd43d391"} Dec 05 09:10:10 crc kubenswrapper[4997]: I1205 09:10:10.736738 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nn5jp" event={"ID":"d228c4a4-4656-4624-af35-94bf3db4c79e","Type":"ContainerStarted","Data":"4ceb7b5daf74f25b3e33e393a852ea1c3229a743abc02a7c72843ad3eb82d5d0"} Dec 05 09:10:10 crc kubenswrapper[4997]: I1205 09:10:10.741393 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nq77q" event={"ID":"3c5bfa50-e54b-490a-a09d-9632b63d9362","Type":"ContainerStarted","Data":"14f8b6aa5681d6218a6f0c779dc1b4758cc08f370a813349fd6a619775fa43fa"} Dec 05 09:10:10 crc kubenswrapper[4997]: I1205 09:10:10.759168 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nn5jp" podStartSLOduration=2.269783556 podStartE2EDuration="7.75914838s" podCreationTimestamp="2025-12-05 09:10:03 +0000 UTC" firstStartedPulling="2025-12-05 09:10:04.63259646 +0000 UTC m=+8105.161503721" lastFinishedPulling="2025-12-05 09:10:10.121961264 +0000 UTC m=+8110.650868545" observedRunningTime="2025-12-05 09:10:10.755897262 +0000 UTC m=+8111.284804543" watchObservedRunningTime="2025-12-05 09:10:10.75914838 +0000 UTC m=+8111.288055641" Dec 05 09:10:10 crc kubenswrapper[4997]: I1205 09:10:10.784533 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nq77q" podStartSLOduration=2.331939004 podStartE2EDuration="5.784502454s" podCreationTimestamp="2025-12-05 09:10:05 +0000 UTC" firstStartedPulling="2025-12-05 09:10:06.682773835 +0000 UTC m=+8107.211681096" lastFinishedPulling="2025-12-05 09:10:10.135337285 +0000 UTC m=+8110.664244546" observedRunningTime="2025-12-05 09:10:10.781483072 +0000 UTC m=+8111.310390353" watchObservedRunningTime="2025-12-05 09:10:10.784502454 +0000 UTC m=+8111.313409715" Dec 05 09:10:13 crc kubenswrapper[4997]: I1205 09:10:13.793940 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nn5jp" Dec 05 09:10:13 crc kubenswrapper[4997]: I1205 09:10:13.794319 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nn5jp" Dec 05 09:10:13 crc kubenswrapper[4997]: I1205 09:10:13.848133 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nn5jp" Dec 05 09:10:15 crc kubenswrapper[4997]: I1205 09:10:15.561511 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nq77q" Dec 05 09:10:15 crc kubenswrapper[4997]: I1205 09:10:15.562078 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nq77q" Dec 05 09:10:15 crc kubenswrapper[4997]: I1205 09:10:15.643275 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nq77q" Dec 05 09:10:15 crc kubenswrapper[4997]: I1205 09:10:15.851135 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nq77q" Dec 05 09:10:16 crc kubenswrapper[4997]: I1205 09:10:16.803841 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nq77q"] Dec 05 09:10:17 crc kubenswrapper[4997]: I1205 09:10:17.818019 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nq77q" podUID="3c5bfa50-e54b-490a-a09d-9632b63d9362" containerName="registry-server" containerID="cri-o://14f8b6aa5681d6218a6f0c779dc1b4758cc08f370a813349fd6a619775fa43fa" gracePeriod=2 Dec 05 09:10:18 crc kubenswrapper[4997]: I1205 09:10:18.830125 4997 generic.go:334] "Generic (PLEG): container finished" podID="3c5bfa50-e54b-490a-a09d-9632b63d9362" containerID="14f8b6aa5681d6218a6f0c779dc1b4758cc08f370a813349fd6a619775fa43fa" exitCode=0 Dec 05 09:10:18 crc kubenswrapper[4997]: I1205 09:10:18.830200 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nq77q" event={"ID":"3c5bfa50-e54b-490a-a09d-9632b63d9362","Type":"ContainerDied","Data":"14f8b6aa5681d6218a6f0c779dc1b4758cc08f370a813349fd6a619775fa43fa"} Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.413999 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nq77q" Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.499951 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5bfa50-e54b-490a-a09d-9632b63d9362-catalog-content\") pod \"3c5bfa50-e54b-490a-a09d-9632b63d9362\" (UID: \"3c5bfa50-e54b-490a-a09d-9632b63d9362\") " Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.500070 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5bfa50-e54b-490a-a09d-9632b63d9362-utilities\") pod \"3c5bfa50-e54b-490a-a09d-9632b63d9362\" (UID: \"3c5bfa50-e54b-490a-a09d-9632b63d9362\") " Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.501073 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c5bfa50-e54b-490a-a09d-9632b63d9362-utilities" (OuterVolumeSpecName: "utilities") pod "3c5bfa50-e54b-490a-a09d-9632b63d9362" (UID: "3c5bfa50-e54b-490a-a09d-9632b63d9362"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.501341 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgbns\" (UniqueName: \"kubernetes.io/projected/3c5bfa50-e54b-490a-a09d-9632b63d9362-kube-api-access-cgbns\") pod \"3c5bfa50-e54b-490a-a09d-9632b63d9362\" (UID: \"3c5bfa50-e54b-490a-a09d-9632b63d9362\") " Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.502553 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5bfa50-e54b-490a-a09d-9632b63d9362-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.507323 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c5bfa50-e54b-490a-a09d-9632b63d9362-kube-api-access-cgbns" (OuterVolumeSpecName: "kube-api-access-cgbns") pod "3c5bfa50-e54b-490a-a09d-9632b63d9362" (UID: "3c5bfa50-e54b-490a-a09d-9632b63d9362"). InnerVolumeSpecName "kube-api-access-cgbns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.519696 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c5bfa50-e54b-490a-a09d-9632b63d9362-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c5bfa50-e54b-490a-a09d-9632b63d9362" (UID: "3c5bfa50-e54b-490a-a09d-9632b63d9362"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.604976 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgbns\" (UniqueName: \"kubernetes.io/projected/3c5bfa50-e54b-490a-a09d-9632b63d9362-kube-api-access-cgbns\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.605011 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5bfa50-e54b-490a-a09d-9632b63d9362-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.773008 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.773107 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.839590 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nq77q" event={"ID":"3c5bfa50-e54b-490a-a09d-9632b63d9362","Type":"ContainerDied","Data":"6e19e990c1f670b64a9fb606fb85ffc9133071df7e4d4aadb13a167229b0c3d0"} Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.839684 4997 scope.go:117] "RemoveContainer" containerID="14f8b6aa5681d6218a6f0c779dc1b4758cc08f370a813349fd6a619775fa43fa" Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.839745 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nq77q" Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.865808 4997 scope.go:117] "RemoveContainer" containerID="b48ffda1358838e310bb18dc53bbf02a6d9af00638dc6ba0a47c6f95dcedf01b" Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.870939 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nq77q"] Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.882286 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nq77q"] Dec 05 09:10:19 crc kubenswrapper[4997]: I1205 09:10:19.886802 4997 scope.go:117] "RemoveContainer" containerID="4aa190137da47dace203d8b24e7d72c94d5d3d4f49a7d83a7e3d6d42210e544c" Dec 05 09:10:21 crc kubenswrapper[4997]: I1205 09:10:21.771808 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c5bfa50-e54b-490a-a09d-9632b63d9362" path="/var/lib/kubelet/pods/3c5bfa50-e54b-490a-a09d-9632b63d9362/volumes" Dec 05 09:10:23 crc kubenswrapper[4997]: I1205 09:10:23.865514 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nn5jp" Dec 05 09:10:23 crc kubenswrapper[4997]: I1205 09:10:23.945158 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nn5jp"] Dec 05 09:10:23 crc kubenswrapper[4997]: I1205 09:10:23.974333 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tdf5t"] Dec 05 09:10:23 crc kubenswrapper[4997]: I1205 09:10:23.974592 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tdf5t" podUID="370b26bd-ba35-4485-a022-613df8a920f9" containerName="registry-server" containerID="cri-o://e53254371f35aed52cd92c866695ce57e9363ca5fb3be536f5c33be9a92b3457" gracePeriod=2 Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.537121 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdf5t" Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.621111 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kw4ts\" (UniqueName: \"kubernetes.io/projected/370b26bd-ba35-4485-a022-613df8a920f9-kube-api-access-kw4ts\") pod \"370b26bd-ba35-4485-a022-613df8a920f9\" (UID: \"370b26bd-ba35-4485-a022-613df8a920f9\") " Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.621216 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/370b26bd-ba35-4485-a022-613df8a920f9-catalog-content\") pod \"370b26bd-ba35-4485-a022-613df8a920f9\" (UID: \"370b26bd-ba35-4485-a022-613df8a920f9\") " Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.621319 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/370b26bd-ba35-4485-a022-613df8a920f9-utilities\") pod \"370b26bd-ba35-4485-a022-613df8a920f9\" (UID: \"370b26bd-ba35-4485-a022-613df8a920f9\") " Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.622180 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/370b26bd-ba35-4485-a022-613df8a920f9-utilities" (OuterVolumeSpecName: "utilities") pod "370b26bd-ba35-4485-a022-613df8a920f9" (UID: "370b26bd-ba35-4485-a022-613df8a920f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.640955 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/370b26bd-ba35-4485-a022-613df8a920f9-kube-api-access-kw4ts" (OuterVolumeSpecName: "kube-api-access-kw4ts") pod "370b26bd-ba35-4485-a022-613df8a920f9" (UID: "370b26bd-ba35-4485-a022-613df8a920f9"). InnerVolumeSpecName "kube-api-access-kw4ts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.664291 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/370b26bd-ba35-4485-a022-613df8a920f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "370b26bd-ba35-4485-a022-613df8a920f9" (UID: "370b26bd-ba35-4485-a022-613df8a920f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.723568 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kw4ts\" (UniqueName: \"kubernetes.io/projected/370b26bd-ba35-4485-a022-613df8a920f9-kube-api-access-kw4ts\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.723628 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/370b26bd-ba35-4485-a022-613df8a920f9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.723640 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/370b26bd-ba35-4485-a022-613df8a920f9-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.908786 4997 generic.go:334] "Generic (PLEG): container finished" podID="370b26bd-ba35-4485-a022-613df8a920f9" containerID="e53254371f35aed52cd92c866695ce57e9363ca5fb3be536f5c33be9a92b3457" exitCode=0 Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.908848 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdf5t" event={"ID":"370b26bd-ba35-4485-a022-613df8a920f9","Type":"ContainerDied","Data":"e53254371f35aed52cd92c866695ce57e9363ca5fb3be536f5c33be9a92b3457"} Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.908908 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdf5t" event={"ID":"370b26bd-ba35-4485-a022-613df8a920f9","Type":"ContainerDied","Data":"d47f8885cb4a9d73348004bfc5fd6eda1beb5adb6d7f2449a60daaf2b478542d"} Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.908927 4997 scope.go:117] "RemoveContainer" containerID="e53254371f35aed52cd92c866695ce57e9363ca5fb3be536f5c33be9a92b3457" Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.908934 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdf5t" Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.937299 4997 scope.go:117] "RemoveContainer" containerID="916c435a120557eb40452b62dde4f95b3fe9469ac55cc394c154cf5746df8ac6" Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.948456 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tdf5t"] Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.959665 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tdf5t"] Dec 05 09:10:24 crc kubenswrapper[4997]: I1205 09:10:24.963932 4997 scope.go:117] "RemoveContainer" containerID="1b0e4c587dd686a6a66d69afde2560006d6cea7e7c5fd970a06a0cdc446cdf37" Dec 05 09:10:25 crc kubenswrapper[4997]: I1205 09:10:25.005310 4997 scope.go:117] "RemoveContainer" containerID="e53254371f35aed52cd92c866695ce57e9363ca5fb3be536f5c33be9a92b3457" Dec 05 09:10:25 crc kubenswrapper[4997]: E1205 09:10:25.005882 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e53254371f35aed52cd92c866695ce57e9363ca5fb3be536f5c33be9a92b3457\": container with ID starting with e53254371f35aed52cd92c866695ce57e9363ca5fb3be536f5c33be9a92b3457 not found: ID does not exist" containerID="e53254371f35aed52cd92c866695ce57e9363ca5fb3be536f5c33be9a92b3457" Dec 05 09:10:25 crc kubenswrapper[4997]: I1205 09:10:25.005932 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e53254371f35aed52cd92c866695ce57e9363ca5fb3be536f5c33be9a92b3457"} err="failed to get container status \"e53254371f35aed52cd92c866695ce57e9363ca5fb3be536f5c33be9a92b3457\": rpc error: code = NotFound desc = could not find container \"e53254371f35aed52cd92c866695ce57e9363ca5fb3be536f5c33be9a92b3457\": container with ID starting with e53254371f35aed52cd92c866695ce57e9363ca5fb3be536f5c33be9a92b3457 not found: ID does not exist" Dec 05 09:10:25 crc kubenswrapper[4997]: I1205 09:10:25.005990 4997 scope.go:117] "RemoveContainer" containerID="916c435a120557eb40452b62dde4f95b3fe9469ac55cc394c154cf5746df8ac6" Dec 05 09:10:25 crc kubenswrapper[4997]: E1205 09:10:25.006481 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"916c435a120557eb40452b62dde4f95b3fe9469ac55cc394c154cf5746df8ac6\": container with ID starting with 916c435a120557eb40452b62dde4f95b3fe9469ac55cc394c154cf5746df8ac6 not found: ID does not exist" containerID="916c435a120557eb40452b62dde4f95b3fe9469ac55cc394c154cf5746df8ac6" Dec 05 09:10:25 crc kubenswrapper[4997]: I1205 09:10:25.006543 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"916c435a120557eb40452b62dde4f95b3fe9469ac55cc394c154cf5746df8ac6"} err="failed to get container status \"916c435a120557eb40452b62dde4f95b3fe9469ac55cc394c154cf5746df8ac6\": rpc error: code = NotFound desc = could not find container \"916c435a120557eb40452b62dde4f95b3fe9469ac55cc394c154cf5746df8ac6\": container with ID starting with 916c435a120557eb40452b62dde4f95b3fe9469ac55cc394c154cf5746df8ac6 not found: ID does not exist" Dec 05 09:10:25 crc kubenswrapper[4997]: I1205 09:10:25.006570 4997 scope.go:117] "RemoveContainer" containerID="1b0e4c587dd686a6a66d69afde2560006d6cea7e7c5fd970a06a0cdc446cdf37" Dec 05 09:10:25 crc kubenswrapper[4997]: E1205 09:10:25.007008 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b0e4c587dd686a6a66d69afde2560006d6cea7e7c5fd970a06a0cdc446cdf37\": container with ID starting with 1b0e4c587dd686a6a66d69afde2560006d6cea7e7c5fd970a06a0cdc446cdf37 not found: ID does not exist" containerID="1b0e4c587dd686a6a66d69afde2560006d6cea7e7c5fd970a06a0cdc446cdf37" Dec 05 09:10:25 crc kubenswrapper[4997]: I1205 09:10:25.007033 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b0e4c587dd686a6a66d69afde2560006d6cea7e7c5fd970a06a0cdc446cdf37"} err="failed to get container status \"1b0e4c587dd686a6a66d69afde2560006d6cea7e7c5fd970a06a0cdc446cdf37\": rpc error: code = NotFound desc = could not find container \"1b0e4c587dd686a6a66d69afde2560006d6cea7e7c5fd970a06a0cdc446cdf37\": container with ID starting with 1b0e4c587dd686a6a66d69afde2560006d6cea7e7c5fd970a06a0cdc446cdf37 not found: ID does not exist" Dec 05 09:10:25 crc kubenswrapper[4997]: I1205 09:10:25.762796 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="370b26bd-ba35-4485-a022-613df8a920f9" path="/var/lib/kubelet/pods/370b26bd-ba35-4485-a022-613df8a920f9/volumes" Dec 05 09:10:49 crc kubenswrapper[4997]: I1205 09:10:49.769809 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:10:49 crc kubenswrapper[4997]: I1205 09:10:49.770537 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:10:49 crc kubenswrapper[4997]: I1205 09:10:49.770603 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 09:10:49 crc kubenswrapper[4997]: I1205 09:10:49.771409 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:10:49 crc kubenswrapper[4997]: I1205 09:10:49.771461 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" gracePeriod=600 Dec 05 09:10:49 crc kubenswrapper[4997]: E1205 09:10:49.898028 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:10:50 crc kubenswrapper[4997]: I1205 09:10:50.155414 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" exitCode=0 Dec 05 09:10:50 crc kubenswrapper[4997]: I1205 09:10:50.155459 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9"} Dec 05 09:10:50 crc kubenswrapper[4997]: I1205 09:10:50.155506 4997 scope.go:117] "RemoveContainer" containerID="a4c28f2d7c6c0a58e95d5724c2f234da68b209b35a39f917c1bd66336a29eaeb" Dec 05 09:10:50 crc kubenswrapper[4997]: I1205 09:10:50.155946 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:10:50 crc kubenswrapper[4997]: E1205 09:10:50.156231 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:11:00 crc kubenswrapper[4997]: I1205 09:11:00.750371 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:11:00 crc kubenswrapper[4997]: E1205 09:11:00.751152 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:11:11 crc kubenswrapper[4997]: I1205 09:11:11.749023 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:11:11 crc kubenswrapper[4997]: E1205 09:11:11.749933 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:11:23 crc kubenswrapper[4997]: I1205 09:11:23.749405 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:11:23 crc kubenswrapper[4997]: E1205 09:11:23.750150 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:11:35 crc kubenswrapper[4997]: I1205 09:11:35.749952 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:11:35 crc kubenswrapper[4997]: E1205 09:11:35.750936 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:11:50 crc kubenswrapper[4997]: I1205 09:11:50.749335 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:11:50 crc kubenswrapper[4997]: E1205 09:11:50.750205 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:12:02 crc kubenswrapper[4997]: I1205 09:12:02.749740 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:12:02 crc kubenswrapper[4997]: E1205 09:12:02.750709 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:12:13 crc kubenswrapper[4997]: I1205 09:12:13.749197 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:12:13 crc kubenswrapper[4997]: E1205 09:12:13.750062 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:12:18 crc kubenswrapper[4997]: I1205 09:12:18.963790 4997 scope.go:117] "RemoveContainer" containerID="dc3a9ec7865fd5c8d34148c68fdd3c653709a6669deb39b13eebe6e530600fe8" Dec 05 09:12:18 crc kubenswrapper[4997]: I1205 09:12:18.987509 4997 scope.go:117] "RemoveContainer" containerID="d0fe7082efc594a2fca09dacffe9a414f09967b43053e8c39fe5cc45ae6b7fa5" Dec 05 09:12:19 crc kubenswrapper[4997]: I1205 09:12:19.017339 4997 scope.go:117] "RemoveContainer" containerID="a2a444682897824fb132feaa47224a0610221db6e4e23eef4a3050cd7bc654a9" Dec 05 09:12:19 crc kubenswrapper[4997]: E1205 09:12:19.017511 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0fe7082efc594a2fca09dacffe9a414f09967b43053e8c39fe5cc45ae6b7fa5\": container with ID starting with d0fe7082efc594a2fca09dacffe9a414f09967b43053e8c39fe5cc45ae6b7fa5 not found: ID does not exist" containerID="d0fe7082efc594a2fca09dacffe9a414f09967b43053e8c39fe5cc45ae6b7fa5" Dec 05 09:12:19 crc kubenswrapper[4997]: I1205 09:12:19.036131 4997 scope.go:117] "RemoveContainer" containerID="be7869d770ed2907f9c09b160fc08e7e9b16d2c13a7cb6d244461029f02fa7c9" Dec 05 09:12:27 crc kubenswrapper[4997]: I1205 09:12:27.749762 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:12:27 crc kubenswrapper[4997]: E1205 09:12:27.751043 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:12:38 crc kubenswrapper[4997]: I1205 09:12:38.749129 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:12:38 crc kubenswrapper[4997]: E1205 09:12:38.750041 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:12:51 crc kubenswrapper[4997]: I1205 09:12:51.749676 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:12:51 crc kubenswrapper[4997]: E1205 09:12:51.750519 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:12:56 crc kubenswrapper[4997]: I1205 09:12:56.610414 4997 generic.go:334] "Generic (PLEG): container finished" podID="6d475848-9ccb-482d-aa85-f6c9e3ef4ef8" containerID="caafb1fe21eae472a1b02c1bb8ef06b4c5c4338a838ea8ddbd2e39c99c05c212" exitCode=0 Dec 05 09:12:56 crc kubenswrapper[4997]: I1205 09:12:56.610501 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" event={"ID":"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8","Type":"ContainerDied","Data":"caafb1fe21eae472a1b02c1bb8ef06b4c5c4338a838ea8ddbd2e39c99c05c212"} Dec 05 09:12:58 crc kubenswrapper[4997]: I1205 09:12:58.040798 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" Dec 05 09:12:58 crc kubenswrapper[4997]: I1205 09:12:58.113751 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-inventory\") pod \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\" (UID: \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\") " Dec 05 09:12:58 crc kubenswrapper[4997]: I1205 09:12:58.113845 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dppw8\" (UniqueName: \"kubernetes.io/projected/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-kube-api-access-dppw8\") pod \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\" (UID: \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\") " Dec 05 09:12:58 crc kubenswrapper[4997]: I1205 09:12:58.113921 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-ssh-key\") pod \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\" (UID: \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\") " Dec 05 09:12:58 crc kubenswrapper[4997]: I1205 09:12:58.113993 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-tripleo-cleanup-combined-ca-bundle\") pod \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\" (UID: \"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8\") " Dec 05 09:12:58 crc kubenswrapper[4997]: I1205 09:12:58.129899 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "6d475848-9ccb-482d-aa85-f6c9e3ef4ef8" (UID: "6d475848-9ccb-482d-aa85-f6c9e3ef4ef8"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:12:58 crc kubenswrapper[4997]: I1205 09:12:58.153041 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-inventory" (OuterVolumeSpecName: "inventory") pod "6d475848-9ccb-482d-aa85-f6c9e3ef4ef8" (UID: "6d475848-9ccb-482d-aa85-f6c9e3ef4ef8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:12:58 crc kubenswrapper[4997]: I1205 09:12:58.153421 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6d475848-9ccb-482d-aa85-f6c9e3ef4ef8" (UID: "6d475848-9ccb-482d-aa85-f6c9e3ef4ef8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:12:58 crc kubenswrapper[4997]: I1205 09:12:58.158741 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-kube-api-access-dppw8" (OuterVolumeSpecName: "kube-api-access-dppw8") pod "6d475848-9ccb-482d-aa85-f6c9e3ef4ef8" (UID: "6d475848-9ccb-482d-aa85-f6c9e3ef4ef8"). InnerVolumeSpecName "kube-api-access-dppw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:12:58 crc kubenswrapper[4997]: I1205 09:12:58.228353 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:58 crc kubenswrapper[4997]: I1205 09:12:58.228396 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dppw8\" (UniqueName: \"kubernetes.io/projected/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-kube-api-access-dppw8\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:58 crc kubenswrapper[4997]: I1205 09:12:58.228410 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:58 crc kubenswrapper[4997]: I1205 09:12:58.228423 4997 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d475848-9ccb-482d-aa85-f6c9e3ef4ef8-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:58 crc kubenswrapper[4997]: I1205 09:12:58.630719 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" event={"ID":"6d475848-9ccb-482d-aa85-f6c9e3ef4ef8","Type":"ContainerDied","Data":"7c8ab7034aaa3716351c07b2404b76c938fd4188add4026ef69d14740d271299"} Dec 05 09:12:58 crc kubenswrapper[4997]: I1205 09:12:58.630763 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c8ab7034aaa3716351c07b2404b76c938fd4188add4026ef69d14740d271299" Dec 05 09:12:58 crc kubenswrapper[4997]: I1205 09:12:58.630793 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2" Dec 05 09:13:04 crc kubenswrapper[4997]: I1205 09:13:04.749662 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:13:04 crc kubenswrapper[4997]: E1205 09:13:04.750495 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:13:11 crc kubenswrapper[4997]: I1205 09:13:11.040052 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-5bmw7"] Dec 05 09:13:11 crc kubenswrapper[4997]: I1205 09:13:11.051797 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-139c-account-create-update-5wbwg"] Dec 05 09:13:11 crc kubenswrapper[4997]: I1205 09:13:11.063186 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-139c-account-create-update-5wbwg"] Dec 05 09:13:11 crc kubenswrapper[4997]: I1205 09:13:11.072601 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-5bmw7"] Dec 05 09:13:11 crc kubenswrapper[4997]: I1205 09:13:11.761034 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="608681d2-0ae1-4fb0-ba87-7db90a81e82a" path="/var/lib/kubelet/pods/608681d2-0ae1-4fb0-ba87-7db90a81e82a/volumes" Dec 05 09:13:11 crc kubenswrapper[4997]: I1205 09:13:11.761829 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="787316ba-2693-4628-a18a-076ddbf78183" path="/var/lib/kubelet/pods/787316ba-2693-4628-a18a-076ddbf78183/volumes" Dec 05 09:13:19 crc kubenswrapper[4997]: I1205 09:13:19.084557 4997 scope.go:117] "RemoveContainer" containerID="483475421a032ea533f681073ac0ddcd26c08b5c3807f97e64018163865d929c" Dec 05 09:13:19 crc kubenswrapper[4997]: I1205 09:13:19.113632 4997 scope.go:117] "RemoveContainer" containerID="5ffe0c3c8085a13be53ec7e83f755705efe1bd8ad3ea194984140d28d9e95363" Dec 05 09:13:19 crc kubenswrapper[4997]: I1205 09:13:19.757226 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:13:19 crc kubenswrapper[4997]: E1205 09:13:19.757817 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:13:26 crc kubenswrapper[4997]: I1205 09:13:26.032729 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-fmsj4"] Dec 05 09:13:26 crc kubenswrapper[4997]: I1205 09:13:26.051504 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-fmsj4"] Dec 05 09:13:27 crc kubenswrapper[4997]: I1205 09:13:27.761038 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddf43729-c369-4311-90b3-57ef233b9c68" path="/var/lib/kubelet/pods/ddf43729-c369-4311-90b3-57ef233b9c68/volumes" Dec 05 09:13:32 crc kubenswrapper[4997]: I1205 09:13:32.749999 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:13:32 crc kubenswrapper[4997]: E1205 09:13:32.750854 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:13:46 crc kubenswrapper[4997]: I1205 09:13:46.748859 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:13:46 crc kubenswrapper[4997]: E1205 09:13:46.749765 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:13:59 crc kubenswrapper[4997]: I1205 09:13:59.761194 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:13:59 crc kubenswrapper[4997]: E1205 09:13:59.762644 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:14:12 crc kubenswrapper[4997]: I1205 09:14:12.749237 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:14:12 crc kubenswrapper[4997]: E1205 09:14:12.749948 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:14:19 crc kubenswrapper[4997]: I1205 09:14:19.214554 4997 scope.go:117] "RemoveContainer" containerID="666c39a9ab896fa3a3d555514d76123cd68f76c17b32bb1eebb272d4037b75c2" Dec 05 09:14:26 crc kubenswrapper[4997]: I1205 09:14:26.749473 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:14:26 crc kubenswrapper[4997]: E1205 09:14:26.750283 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:14:40 crc kubenswrapper[4997]: I1205 09:14:40.751770 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:14:40 crc kubenswrapper[4997]: E1205 09:14:40.752919 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.522825 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kbttp"] Dec 05 09:14:47 crc kubenswrapper[4997]: E1205 09:14:47.523949 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="370b26bd-ba35-4485-a022-613df8a920f9" containerName="extract-utilities" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.523971 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="370b26bd-ba35-4485-a022-613df8a920f9" containerName="extract-utilities" Dec 05 09:14:47 crc kubenswrapper[4997]: E1205 09:14:47.523997 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5bfa50-e54b-490a-a09d-9632b63d9362" containerName="extract-utilities" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.524005 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5bfa50-e54b-490a-a09d-9632b63d9362" containerName="extract-utilities" Dec 05 09:14:47 crc kubenswrapper[4997]: E1205 09:14:47.524025 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5bfa50-e54b-490a-a09d-9632b63d9362" containerName="extract-content" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.524035 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5bfa50-e54b-490a-a09d-9632b63d9362" containerName="extract-content" Dec 05 09:14:47 crc kubenswrapper[4997]: E1205 09:14:47.524061 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d475848-9ccb-482d-aa85-f6c9e3ef4ef8" containerName="tripleo-cleanup-tripleo-cleanup-openstack-networker" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.524071 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d475848-9ccb-482d-aa85-f6c9e3ef4ef8" containerName="tripleo-cleanup-tripleo-cleanup-openstack-networker" Dec 05 09:14:47 crc kubenswrapper[4997]: E1205 09:14:47.524086 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="370b26bd-ba35-4485-a022-613df8a920f9" containerName="extract-content" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.524097 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="370b26bd-ba35-4485-a022-613df8a920f9" containerName="extract-content" Dec 05 09:14:47 crc kubenswrapper[4997]: E1205 09:14:47.524107 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5bfa50-e54b-490a-a09d-9632b63d9362" containerName="registry-server" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.524115 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5bfa50-e54b-490a-a09d-9632b63d9362" containerName="registry-server" Dec 05 09:14:47 crc kubenswrapper[4997]: E1205 09:14:47.524157 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="370b26bd-ba35-4485-a022-613df8a920f9" containerName="registry-server" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.524166 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="370b26bd-ba35-4485-a022-613df8a920f9" containerName="registry-server" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.524516 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d475848-9ccb-482d-aa85-f6c9e3ef4ef8" containerName="tripleo-cleanup-tripleo-cleanup-openstack-networker" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.524543 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="370b26bd-ba35-4485-a022-613df8a920f9" containerName="registry-server" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.524564 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c5bfa50-e54b-490a-a09d-9632b63d9362" containerName="registry-server" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.526513 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kbttp" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.563370 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kbttp"] Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.621722 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87ad049a-9d68-4136-a547-469dd10d10d2-catalog-content\") pod \"redhat-operators-kbttp\" (UID: \"87ad049a-9d68-4136-a547-469dd10d10d2\") " pod="openshift-marketplace/redhat-operators-kbttp" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.621773 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87ad049a-9d68-4136-a547-469dd10d10d2-utilities\") pod \"redhat-operators-kbttp\" (UID: \"87ad049a-9d68-4136-a547-469dd10d10d2\") " pod="openshift-marketplace/redhat-operators-kbttp" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.621844 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xddv\" (UniqueName: \"kubernetes.io/projected/87ad049a-9d68-4136-a547-469dd10d10d2-kube-api-access-6xddv\") pod \"redhat-operators-kbttp\" (UID: \"87ad049a-9d68-4136-a547-469dd10d10d2\") " pod="openshift-marketplace/redhat-operators-kbttp" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.723422 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xddv\" (UniqueName: \"kubernetes.io/projected/87ad049a-9d68-4136-a547-469dd10d10d2-kube-api-access-6xddv\") pod \"redhat-operators-kbttp\" (UID: \"87ad049a-9d68-4136-a547-469dd10d10d2\") " pod="openshift-marketplace/redhat-operators-kbttp" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.723601 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87ad049a-9d68-4136-a547-469dd10d10d2-catalog-content\") pod \"redhat-operators-kbttp\" (UID: \"87ad049a-9d68-4136-a547-469dd10d10d2\") " pod="openshift-marketplace/redhat-operators-kbttp" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.723642 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87ad049a-9d68-4136-a547-469dd10d10d2-utilities\") pod \"redhat-operators-kbttp\" (UID: \"87ad049a-9d68-4136-a547-469dd10d10d2\") " pod="openshift-marketplace/redhat-operators-kbttp" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.724569 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87ad049a-9d68-4136-a547-469dd10d10d2-utilities\") pod \"redhat-operators-kbttp\" (UID: \"87ad049a-9d68-4136-a547-469dd10d10d2\") " pod="openshift-marketplace/redhat-operators-kbttp" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.724728 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87ad049a-9d68-4136-a547-469dd10d10d2-catalog-content\") pod \"redhat-operators-kbttp\" (UID: \"87ad049a-9d68-4136-a547-469dd10d10d2\") " pod="openshift-marketplace/redhat-operators-kbttp" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.747645 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xddv\" (UniqueName: \"kubernetes.io/projected/87ad049a-9d68-4136-a547-469dd10d10d2-kube-api-access-6xddv\") pod \"redhat-operators-kbttp\" (UID: \"87ad049a-9d68-4136-a547-469dd10d10d2\") " pod="openshift-marketplace/redhat-operators-kbttp" Dec 05 09:14:47 crc kubenswrapper[4997]: I1205 09:14:47.853026 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kbttp" Dec 05 09:14:48 crc kubenswrapper[4997]: I1205 09:14:48.431798 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kbttp"] Dec 05 09:14:48 crc kubenswrapper[4997]: I1205 09:14:48.755290 4997 generic.go:334] "Generic (PLEG): container finished" podID="87ad049a-9d68-4136-a547-469dd10d10d2" containerID="2a81acdb1361211f1cbf17e92f2f0d733a0a94dd722c2310b1cd522f983dcc35" exitCode=0 Dec 05 09:14:48 crc kubenswrapper[4997]: I1205 09:14:48.755340 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kbttp" event={"ID":"87ad049a-9d68-4136-a547-469dd10d10d2","Type":"ContainerDied","Data":"2a81acdb1361211f1cbf17e92f2f0d733a0a94dd722c2310b1cd522f983dcc35"} Dec 05 09:14:48 crc kubenswrapper[4997]: I1205 09:14:48.755374 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kbttp" event={"ID":"87ad049a-9d68-4136-a547-469dd10d10d2","Type":"ContainerStarted","Data":"c1a23be416db7eead414cb6cd4e2f1efed89644a451b21c71e25e75b9f739fd7"} Dec 05 09:14:49 crc kubenswrapper[4997]: I1205 09:14:49.777183 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kbttp" event={"ID":"87ad049a-9d68-4136-a547-469dd10d10d2","Type":"ContainerStarted","Data":"8de6a82cb0321e633871d7123584e2912d774046af731e3e2706f7d702aa733d"} Dec 05 09:14:52 crc kubenswrapper[4997]: I1205 09:14:52.749061 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:14:52 crc kubenswrapper[4997]: E1205 09:14:52.750079 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:14:53 crc kubenswrapper[4997]: I1205 09:14:53.842900 4997 generic.go:334] "Generic (PLEG): container finished" podID="87ad049a-9d68-4136-a547-469dd10d10d2" containerID="8de6a82cb0321e633871d7123584e2912d774046af731e3e2706f7d702aa733d" exitCode=0 Dec 05 09:14:53 crc kubenswrapper[4997]: I1205 09:14:53.843102 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kbttp" event={"ID":"87ad049a-9d68-4136-a547-469dd10d10d2","Type":"ContainerDied","Data":"8de6a82cb0321e633871d7123584e2912d774046af731e3e2706f7d702aa733d"} Dec 05 09:14:54 crc kubenswrapper[4997]: I1205 09:14:54.855529 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kbttp" event={"ID":"87ad049a-9d68-4136-a547-469dd10d10d2","Type":"ContainerStarted","Data":"9f4feb62e7359eeb86ff6a37d3c5d5715b80247d5326de18b6b62192f65255be"} Dec 05 09:14:54 crc kubenswrapper[4997]: I1205 09:14:54.883923 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kbttp" podStartSLOduration=2.396779757 podStartE2EDuration="7.883903392s" podCreationTimestamp="2025-12-05 09:14:47 +0000 UTC" firstStartedPulling="2025-12-05 09:14:48.757048363 +0000 UTC m=+8389.285955624" lastFinishedPulling="2025-12-05 09:14:54.244171998 +0000 UTC m=+8394.773079259" observedRunningTime="2025-12-05 09:14:54.872734721 +0000 UTC m=+8395.401642022" watchObservedRunningTime="2025-12-05 09:14:54.883903392 +0000 UTC m=+8395.412810653" Dec 05 09:14:57 crc kubenswrapper[4997]: I1205 09:14:57.853511 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kbttp" Dec 05 09:14:57 crc kubenswrapper[4997]: I1205 09:14:57.854007 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kbttp" Dec 05 09:14:58 crc kubenswrapper[4997]: I1205 09:14:58.921196 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kbttp" podUID="87ad049a-9d68-4136-a547-469dd10d10d2" containerName="registry-server" probeResult="failure" output=< Dec 05 09:14:58 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 09:14:58 crc kubenswrapper[4997]: > Dec 05 09:15:00 crc kubenswrapper[4997]: I1205 09:15:00.162006 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr"] Dec 05 09:15:00 crc kubenswrapper[4997]: I1205 09:15:00.164117 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr" Dec 05 09:15:00 crc kubenswrapper[4997]: I1205 09:15:00.167840 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 09:15:00 crc kubenswrapper[4997]: I1205 09:15:00.168313 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 09:15:00 crc kubenswrapper[4997]: I1205 09:15:00.177153 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr"] Dec 05 09:15:00 crc kubenswrapper[4997]: I1205 09:15:00.301682 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zqhz\" (UniqueName: \"kubernetes.io/projected/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-kube-api-access-6zqhz\") pod \"collect-profiles-29415435-5nqrr\" (UID: \"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr" Dec 05 09:15:00 crc kubenswrapper[4997]: I1205 09:15:00.302093 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-secret-volume\") pod \"collect-profiles-29415435-5nqrr\" (UID: \"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr" Dec 05 09:15:00 crc kubenswrapper[4997]: I1205 09:15:00.302225 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-config-volume\") pod \"collect-profiles-29415435-5nqrr\" (UID: \"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr" Dec 05 09:15:00 crc kubenswrapper[4997]: I1205 09:15:00.404667 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zqhz\" (UniqueName: \"kubernetes.io/projected/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-kube-api-access-6zqhz\") pod \"collect-profiles-29415435-5nqrr\" (UID: \"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr" Dec 05 09:15:00 crc kubenswrapper[4997]: I1205 09:15:00.404875 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-secret-volume\") pod \"collect-profiles-29415435-5nqrr\" (UID: \"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr" Dec 05 09:15:00 crc kubenswrapper[4997]: I1205 09:15:00.404947 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-config-volume\") pod \"collect-profiles-29415435-5nqrr\" (UID: \"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr" Dec 05 09:15:00 crc kubenswrapper[4997]: I1205 09:15:00.406316 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-config-volume\") pod \"collect-profiles-29415435-5nqrr\" (UID: \"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr" Dec 05 09:15:00 crc kubenswrapper[4997]: I1205 09:15:00.420970 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-secret-volume\") pod \"collect-profiles-29415435-5nqrr\" (UID: \"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr" Dec 05 09:15:00 crc kubenswrapper[4997]: I1205 09:15:00.423398 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zqhz\" (UniqueName: \"kubernetes.io/projected/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-kube-api-access-6zqhz\") pod \"collect-profiles-29415435-5nqrr\" (UID: \"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr" Dec 05 09:15:00 crc kubenswrapper[4997]: I1205 09:15:00.488935 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr" Dec 05 09:15:00 crc kubenswrapper[4997]: I1205 09:15:00.959584 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr"] Dec 05 09:15:01 crc kubenswrapper[4997]: I1205 09:15:01.919421 4997 generic.go:334] "Generic (PLEG): container finished" podID="f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1" containerID="46f6bdd664b048dd2d1854cdcdbc93d413ce0f5a3f1e90c9f0de703378393c59" exitCode=0 Dec 05 09:15:01 crc kubenswrapper[4997]: I1205 09:15:01.919541 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr" event={"ID":"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1","Type":"ContainerDied","Data":"46f6bdd664b048dd2d1854cdcdbc93d413ce0f5a3f1e90c9f0de703378393c59"} Dec 05 09:15:01 crc kubenswrapper[4997]: I1205 09:15:01.919892 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr" event={"ID":"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1","Type":"ContainerStarted","Data":"02e0d9a547c716734ea39713b94bf2f3984f749271dcbc86c2991fc366081610"} Dec 05 09:15:03 crc kubenswrapper[4997]: I1205 09:15:03.332952 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr" Dec 05 09:15:03 crc kubenswrapper[4997]: I1205 09:15:03.478389 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zqhz\" (UniqueName: \"kubernetes.io/projected/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-kube-api-access-6zqhz\") pod \"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1\" (UID: \"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1\") " Dec 05 09:15:03 crc kubenswrapper[4997]: I1205 09:15:03.478608 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-config-volume\") pod \"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1\" (UID: \"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1\") " Dec 05 09:15:03 crc kubenswrapper[4997]: I1205 09:15:03.478803 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-secret-volume\") pod \"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1\" (UID: \"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1\") " Dec 05 09:15:03 crc kubenswrapper[4997]: I1205 09:15:03.479240 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-config-volume" (OuterVolumeSpecName: "config-volume") pod "f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1" (UID: "f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:15:03 crc kubenswrapper[4997]: I1205 09:15:03.479379 4997 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:03 crc kubenswrapper[4997]: I1205 09:15:03.484966 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-kube-api-access-6zqhz" (OuterVolumeSpecName: "kube-api-access-6zqhz") pod "f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1" (UID: "f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1"). InnerVolumeSpecName "kube-api-access-6zqhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:15:03 crc kubenswrapper[4997]: I1205 09:15:03.485026 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1" (UID: "f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:15:03 crc kubenswrapper[4997]: I1205 09:15:03.580973 4997 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:03 crc kubenswrapper[4997]: I1205 09:15:03.581018 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zqhz\" (UniqueName: \"kubernetes.io/projected/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1-kube-api-access-6zqhz\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:03 crc kubenswrapper[4997]: I1205 09:15:03.940596 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr" event={"ID":"f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1","Type":"ContainerDied","Data":"02e0d9a547c716734ea39713b94bf2f3984f749271dcbc86c2991fc366081610"} Dec 05 09:15:03 crc kubenswrapper[4997]: I1205 09:15:03.940878 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02e0d9a547c716734ea39713b94bf2f3984f749271dcbc86c2991fc366081610" Dec 05 09:15:03 crc kubenswrapper[4997]: I1205 09:15:03.940651 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr" Dec 05 09:15:04 crc kubenswrapper[4997]: I1205 09:15:04.409026 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g"] Dec 05 09:15:04 crc kubenswrapper[4997]: I1205 09:15:04.417654 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415390-2kb7g"] Dec 05 09:15:04 crc kubenswrapper[4997]: I1205 09:15:04.750000 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:15:04 crc kubenswrapper[4997]: E1205 09:15:04.750370 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:15:05 crc kubenswrapper[4997]: I1205 09:15:05.765029 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="708bfbf7-22c5-4cbf-9dca-c7220fdd1615" path="/var/lib/kubelet/pods/708bfbf7-22c5-4cbf-9dca-c7220fdd1615/volumes" Dec 05 09:15:07 crc kubenswrapper[4997]: I1205 09:15:07.911428 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kbttp" Dec 05 09:15:07 crc kubenswrapper[4997]: I1205 09:15:07.976423 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kbttp" Dec 05 09:15:08 crc kubenswrapper[4997]: I1205 09:15:08.198860 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kbttp"] Dec 05 09:15:08 crc kubenswrapper[4997]: I1205 09:15:08.990106 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kbttp" podUID="87ad049a-9d68-4136-a547-469dd10d10d2" containerName="registry-server" containerID="cri-o://9f4feb62e7359eeb86ff6a37d3c5d5715b80247d5326de18b6b62192f65255be" gracePeriod=2 Dec 05 09:15:09 crc kubenswrapper[4997]: I1205 09:15:09.478390 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kbttp" Dec 05 09:15:09 crc kubenswrapper[4997]: I1205 09:15:09.533463 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xddv\" (UniqueName: \"kubernetes.io/projected/87ad049a-9d68-4136-a547-469dd10d10d2-kube-api-access-6xddv\") pod \"87ad049a-9d68-4136-a547-469dd10d10d2\" (UID: \"87ad049a-9d68-4136-a547-469dd10d10d2\") " Dec 05 09:15:09 crc kubenswrapper[4997]: I1205 09:15:09.533629 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87ad049a-9d68-4136-a547-469dd10d10d2-utilities\") pod \"87ad049a-9d68-4136-a547-469dd10d10d2\" (UID: \"87ad049a-9d68-4136-a547-469dd10d10d2\") " Dec 05 09:15:09 crc kubenswrapper[4997]: I1205 09:15:09.533687 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87ad049a-9d68-4136-a547-469dd10d10d2-catalog-content\") pod \"87ad049a-9d68-4136-a547-469dd10d10d2\" (UID: \"87ad049a-9d68-4136-a547-469dd10d10d2\") " Dec 05 09:15:09 crc kubenswrapper[4997]: I1205 09:15:09.534865 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87ad049a-9d68-4136-a547-469dd10d10d2-utilities" (OuterVolumeSpecName: "utilities") pod "87ad049a-9d68-4136-a547-469dd10d10d2" (UID: "87ad049a-9d68-4136-a547-469dd10d10d2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:15:09 crc kubenswrapper[4997]: I1205 09:15:09.539398 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87ad049a-9d68-4136-a547-469dd10d10d2-kube-api-access-6xddv" (OuterVolumeSpecName: "kube-api-access-6xddv") pod "87ad049a-9d68-4136-a547-469dd10d10d2" (UID: "87ad049a-9d68-4136-a547-469dd10d10d2"). InnerVolumeSpecName "kube-api-access-6xddv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:15:09 crc kubenswrapper[4997]: I1205 09:15:09.638304 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87ad049a-9d68-4136-a547-469dd10d10d2-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:09 crc kubenswrapper[4997]: I1205 09:15:09.638338 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xddv\" (UniqueName: \"kubernetes.io/projected/87ad049a-9d68-4136-a547-469dd10d10d2-kube-api-access-6xddv\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:09 crc kubenswrapper[4997]: I1205 09:15:09.642165 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87ad049a-9d68-4136-a547-469dd10d10d2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87ad049a-9d68-4136-a547-469dd10d10d2" (UID: "87ad049a-9d68-4136-a547-469dd10d10d2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:15:09 crc kubenswrapper[4997]: I1205 09:15:09.739768 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87ad049a-9d68-4136-a547-469dd10d10d2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:10 crc kubenswrapper[4997]: I1205 09:15:10.004224 4997 generic.go:334] "Generic (PLEG): container finished" podID="87ad049a-9d68-4136-a547-469dd10d10d2" containerID="9f4feb62e7359eeb86ff6a37d3c5d5715b80247d5326de18b6b62192f65255be" exitCode=0 Dec 05 09:15:10 crc kubenswrapper[4997]: I1205 09:15:10.004281 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kbttp" event={"ID":"87ad049a-9d68-4136-a547-469dd10d10d2","Type":"ContainerDied","Data":"9f4feb62e7359eeb86ff6a37d3c5d5715b80247d5326de18b6b62192f65255be"} Dec 05 09:15:10 crc kubenswrapper[4997]: I1205 09:15:10.004314 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kbttp" event={"ID":"87ad049a-9d68-4136-a547-469dd10d10d2","Type":"ContainerDied","Data":"c1a23be416db7eead414cb6cd4e2f1efed89644a451b21c71e25e75b9f739fd7"} Dec 05 09:15:10 crc kubenswrapper[4997]: I1205 09:15:10.004336 4997 scope.go:117] "RemoveContainer" containerID="9f4feb62e7359eeb86ff6a37d3c5d5715b80247d5326de18b6b62192f65255be" Dec 05 09:15:10 crc kubenswrapper[4997]: I1205 09:15:10.004511 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kbttp" Dec 05 09:15:10 crc kubenswrapper[4997]: I1205 09:15:10.034923 4997 scope.go:117] "RemoveContainer" containerID="8de6a82cb0321e633871d7123584e2912d774046af731e3e2706f7d702aa733d" Dec 05 09:15:10 crc kubenswrapper[4997]: I1205 09:15:10.036706 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kbttp"] Dec 05 09:15:10 crc kubenswrapper[4997]: I1205 09:15:10.047383 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kbttp"] Dec 05 09:15:10 crc kubenswrapper[4997]: I1205 09:15:10.063666 4997 scope.go:117] "RemoveContainer" containerID="2a81acdb1361211f1cbf17e92f2f0d733a0a94dd722c2310b1cd522f983dcc35" Dec 05 09:15:10 crc kubenswrapper[4997]: I1205 09:15:10.105139 4997 scope.go:117] "RemoveContainer" containerID="9f4feb62e7359eeb86ff6a37d3c5d5715b80247d5326de18b6b62192f65255be" Dec 05 09:15:10 crc kubenswrapper[4997]: E1205 09:15:10.109781 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f4feb62e7359eeb86ff6a37d3c5d5715b80247d5326de18b6b62192f65255be\": container with ID starting with 9f4feb62e7359eeb86ff6a37d3c5d5715b80247d5326de18b6b62192f65255be not found: ID does not exist" containerID="9f4feb62e7359eeb86ff6a37d3c5d5715b80247d5326de18b6b62192f65255be" Dec 05 09:15:10 crc kubenswrapper[4997]: I1205 09:15:10.110025 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f4feb62e7359eeb86ff6a37d3c5d5715b80247d5326de18b6b62192f65255be"} err="failed to get container status \"9f4feb62e7359eeb86ff6a37d3c5d5715b80247d5326de18b6b62192f65255be\": rpc error: code = NotFound desc = could not find container \"9f4feb62e7359eeb86ff6a37d3c5d5715b80247d5326de18b6b62192f65255be\": container with ID starting with 9f4feb62e7359eeb86ff6a37d3c5d5715b80247d5326de18b6b62192f65255be not found: ID does not exist" Dec 05 09:15:10 crc kubenswrapper[4997]: I1205 09:15:10.110061 4997 scope.go:117] "RemoveContainer" containerID="8de6a82cb0321e633871d7123584e2912d774046af731e3e2706f7d702aa733d" Dec 05 09:15:10 crc kubenswrapper[4997]: E1205 09:15:10.110501 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8de6a82cb0321e633871d7123584e2912d774046af731e3e2706f7d702aa733d\": container with ID starting with 8de6a82cb0321e633871d7123584e2912d774046af731e3e2706f7d702aa733d not found: ID does not exist" containerID="8de6a82cb0321e633871d7123584e2912d774046af731e3e2706f7d702aa733d" Dec 05 09:15:10 crc kubenswrapper[4997]: I1205 09:15:10.110594 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8de6a82cb0321e633871d7123584e2912d774046af731e3e2706f7d702aa733d"} err="failed to get container status \"8de6a82cb0321e633871d7123584e2912d774046af731e3e2706f7d702aa733d\": rpc error: code = NotFound desc = could not find container \"8de6a82cb0321e633871d7123584e2912d774046af731e3e2706f7d702aa733d\": container with ID starting with 8de6a82cb0321e633871d7123584e2912d774046af731e3e2706f7d702aa733d not found: ID does not exist" Dec 05 09:15:10 crc kubenswrapper[4997]: I1205 09:15:10.110677 4997 scope.go:117] "RemoveContainer" containerID="2a81acdb1361211f1cbf17e92f2f0d733a0a94dd722c2310b1cd522f983dcc35" Dec 05 09:15:10 crc kubenswrapper[4997]: E1205 09:15:10.111103 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a81acdb1361211f1cbf17e92f2f0d733a0a94dd722c2310b1cd522f983dcc35\": container with ID starting with 2a81acdb1361211f1cbf17e92f2f0d733a0a94dd722c2310b1cd522f983dcc35 not found: ID does not exist" containerID="2a81acdb1361211f1cbf17e92f2f0d733a0a94dd722c2310b1cd522f983dcc35" Dec 05 09:15:10 crc kubenswrapper[4997]: I1205 09:15:10.111144 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a81acdb1361211f1cbf17e92f2f0d733a0a94dd722c2310b1cd522f983dcc35"} err="failed to get container status \"2a81acdb1361211f1cbf17e92f2f0d733a0a94dd722c2310b1cd522f983dcc35\": rpc error: code = NotFound desc = could not find container \"2a81acdb1361211f1cbf17e92f2f0d733a0a94dd722c2310b1cd522f983dcc35\": container with ID starting with 2a81acdb1361211f1cbf17e92f2f0d733a0a94dd722c2310b1cd522f983dcc35 not found: ID does not exist" Dec 05 09:15:11 crc kubenswrapper[4997]: I1205 09:15:11.762304 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87ad049a-9d68-4136-a547-469dd10d10d2" path="/var/lib/kubelet/pods/87ad049a-9d68-4136-a547-469dd10d10d2/volumes" Dec 05 09:15:19 crc kubenswrapper[4997]: I1205 09:15:19.265109 4997 scope.go:117] "RemoveContainer" containerID="100d4817715ec67fcf1a312c1b88613e5b24a5ce57475e06b138946098010fc0" Dec 05 09:15:19 crc kubenswrapper[4997]: I1205 09:15:19.756488 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:15:19 crc kubenswrapper[4997]: E1205 09:15:19.757072 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:15:32 crc kubenswrapper[4997]: I1205 09:15:32.750271 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:15:32 crc kubenswrapper[4997]: E1205 09:15:32.751490 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:15:42 crc kubenswrapper[4997]: I1205 09:15:42.043210 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-jrxqn"] Dec 05 09:15:42 crc kubenswrapper[4997]: I1205 09:15:42.053567 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-jrxqn"] Dec 05 09:15:43 crc kubenswrapper[4997]: I1205 09:15:43.031475 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-bcea-account-create-update-8vzh5"] Dec 05 09:15:43 crc kubenswrapper[4997]: I1205 09:15:43.042802 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-bcea-account-create-update-8vzh5"] Dec 05 09:15:43 crc kubenswrapper[4997]: I1205 09:15:43.760799 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0036e14d-14f9-4631-903c-8682e6babada" path="/var/lib/kubelet/pods/0036e14d-14f9-4631-903c-8682e6babada/volumes" Dec 05 09:15:43 crc kubenswrapper[4997]: I1205 09:15:43.761363 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d315ff5-173e-4469-856c-74ece8f15669" path="/var/lib/kubelet/pods/5d315ff5-173e-4469-856c-74ece8f15669/volumes" Dec 05 09:15:47 crc kubenswrapper[4997]: I1205 09:15:47.751823 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:15:47 crc kubenswrapper[4997]: E1205 09:15:47.752504 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:15:55 crc kubenswrapper[4997]: I1205 09:15:55.036390 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-5cpz9"] Dec 05 09:15:55 crc kubenswrapper[4997]: I1205 09:15:55.052939 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-5cpz9"] Dec 05 09:15:55 crc kubenswrapper[4997]: I1205 09:15:55.763563 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2519e54-7c30-40f0-a1eb-f576ae2f6fa9" path="/var/lib/kubelet/pods/b2519e54-7c30-40f0-a1eb-f576ae2f6fa9/volumes" Dec 05 09:15:58 crc kubenswrapper[4997]: I1205 09:15:58.749337 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:15:59 crc kubenswrapper[4997]: I1205 09:15:59.494801 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"f8205236afd16f8a0c6d433a5ab02465034afc3f440c01d487df254a7a73dfb2"} Dec 05 09:16:10 crc kubenswrapper[4997]: I1205 09:16:10.600986 4997 generic.go:334] "Generic (PLEG): container finished" podID="82e1e5f9-2d07-4e81-a32a-a35535dd9d2d" containerID="0b6c53e049496e78722cd4d19b8036b4bd04265cb5d20c6bf52f43652381f5fa" exitCode=0 Dec 05 09:16:10 crc kubenswrapper[4997]: I1205 09:16:10.601130 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" event={"ID":"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d","Type":"ContainerDied","Data":"0b6c53e049496e78722cd4d19b8036b4bd04265cb5d20c6bf52f43652381f5fa"} Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.090955 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.135311 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-inventory\") pod \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.135452 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-ssh-key\") pod \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.135487 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmfsw\" (UniqueName: \"kubernetes.io/projected/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-kube-api-access-tmfsw\") pod \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.135514 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-ceph\") pod \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.135534 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-tripleo-cleanup-combined-ca-bundle\") pod \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\" (UID: \"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d\") " Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.142140 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-ceph" (OuterVolumeSpecName: "ceph") pod "82e1e5f9-2d07-4e81-a32a-a35535dd9d2d" (UID: "82e1e5f9-2d07-4e81-a32a-a35535dd9d2d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.142790 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "82e1e5f9-2d07-4e81-a32a-a35535dd9d2d" (UID: "82e1e5f9-2d07-4e81-a32a-a35535dd9d2d"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.147019 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-kube-api-access-tmfsw" (OuterVolumeSpecName: "kube-api-access-tmfsw") pod "82e1e5f9-2d07-4e81-a32a-a35535dd9d2d" (UID: "82e1e5f9-2d07-4e81-a32a-a35535dd9d2d"). InnerVolumeSpecName "kube-api-access-tmfsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.175504 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-inventory" (OuterVolumeSpecName: "inventory") pod "82e1e5f9-2d07-4e81-a32a-a35535dd9d2d" (UID: "82e1e5f9-2d07-4e81-a32a-a35535dd9d2d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.181661 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "82e1e5f9-2d07-4e81-a32a-a35535dd9d2d" (UID: "82e1e5f9-2d07-4e81-a32a-a35535dd9d2d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.237880 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.237915 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.237925 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmfsw\" (UniqueName: \"kubernetes.io/projected/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-kube-api-access-tmfsw\") on node \"crc\" DevicePath \"\"" Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.237935 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.237944 4997 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82e1e5f9-2d07-4e81-a32a-a35535dd9d2d-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.621313 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" event={"ID":"82e1e5f9-2d07-4e81-a32a-a35535dd9d2d","Type":"ContainerDied","Data":"34eb374ce6f81e95f6fa93ebfed5616d5420e98c8a1f2c3c8e8027af1a6958d3"} Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.621667 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34eb374ce6f81e95f6fa93ebfed5616d5420e98c8a1f2c3c8e8027af1a6958d3" Dec 05 09:16:12 crc kubenswrapper[4997]: I1205 09:16:12.621737 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx" Dec 05 09:16:14 crc kubenswrapper[4997]: I1205 09:16:14.063261 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-lsxlp"] Dec 05 09:16:14 crc kubenswrapper[4997]: I1205 09:16:14.073572 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-9580-account-create-update-6cdhz"] Dec 05 09:16:14 crc kubenswrapper[4997]: I1205 09:16:14.082355 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-9580-account-create-update-6cdhz"] Dec 05 09:16:14 crc kubenswrapper[4997]: I1205 09:16:14.090030 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-lsxlp"] Dec 05 09:16:15 crc kubenswrapper[4997]: I1205 09:16:15.766887 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="535f8966-92df-4881-9771-c5dfc02bf193" path="/var/lib/kubelet/pods/535f8966-92df-4881-9771-c5dfc02bf193/volumes" Dec 05 09:16:15 crc kubenswrapper[4997]: I1205 09:16:15.769101 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae61f102-3277-41ea-8c08-086bd0225419" path="/var/lib/kubelet/pods/ae61f102-3277-41ea-8c08-086bd0225419/volumes" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.659024 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q756d"] Dec 05 09:16:18 crc kubenswrapper[4997]: E1205 09:16:18.660166 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87ad049a-9d68-4136-a547-469dd10d10d2" containerName="extract-content" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.660189 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="87ad049a-9d68-4136-a547-469dd10d10d2" containerName="extract-content" Dec 05 09:16:18 crc kubenswrapper[4997]: E1205 09:16:18.660217 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1" containerName="collect-profiles" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.660226 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1" containerName="collect-profiles" Dec 05 09:16:18 crc kubenswrapper[4997]: E1205 09:16:18.660264 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87ad049a-9d68-4136-a547-469dd10d10d2" containerName="extract-utilities" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.660272 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="87ad049a-9d68-4136-a547-469dd10d10d2" containerName="extract-utilities" Dec 05 09:16:18 crc kubenswrapper[4997]: E1205 09:16:18.660299 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87ad049a-9d68-4136-a547-469dd10d10d2" containerName="registry-server" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.660307 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="87ad049a-9d68-4136-a547-469dd10d10d2" containerName="registry-server" Dec 05 09:16:18 crc kubenswrapper[4997]: E1205 09:16:18.660324 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82e1e5f9-2d07-4e81-a32a-a35535dd9d2d" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.660333 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="82e1e5f9-2d07-4e81-a32a-a35535dd9d2d" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.660598 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="87ad049a-9d68-4136-a547-469dd10d10d2" containerName="registry-server" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.660634 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="82e1e5f9-2d07-4e81-a32a-a35535dd9d2d" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.660646 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1" containerName="collect-profiles" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.662531 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q756d" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.680935 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q756d"] Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.762934 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/798d0a0f-e302-48af-8ab9-1640b68805b0-catalog-content\") pod \"certified-operators-q756d\" (UID: \"798d0a0f-e302-48af-8ab9-1640b68805b0\") " pod="openshift-marketplace/certified-operators-q756d" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.763298 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/798d0a0f-e302-48af-8ab9-1640b68805b0-utilities\") pod \"certified-operators-q756d\" (UID: \"798d0a0f-e302-48af-8ab9-1640b68805b0\") " pod="openshift-marketplace/certified-operators-q756d" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.763551 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf8f8\" (UniqueName: \"kubernetes.io/projected/798d0a0f-e302-48af-8ab9-1640b68805b0-kube-api-access-bf8f8\") pod \"certified-operators-q756d\" (UID: \"798d0a0f-e302-48af-8ab9-1640b68805b0\") " pod="openshift-marketplace/certified-operators-q756d" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.865700 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf8f8\" (UniqueName: \"kubernetes.io/projected/798d0a0f-e302-48af-8ab9-1640b68805b0-kube-api-access-bf8f8\") pod \"certified-operators-q756d\" (UID: \"798d0a0f-e302-48af-8ab9-1640b68805b0\") " pod="openshift-marketplace/certified-operators-q756d" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.865782 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/798d0a0f-e302-48af-8ab9-1640b68805b0-catalog-content\") pod \"certified-operators-q756d\" (UID: \"798d0a0f-e302-48af-8ab9-1640b68805b0\") " pod="openshift-marketplace/certified-operators-q756d" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.865813 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/798d0a0f-e302-48af-8ab9-1640b68805b0-utilities\") pod \"certified-operators-q756d\" (UID: \"798d0a0f-e302-48af-8ab9-1640b68805b0\") " pod="openshift-marketplace/certified-operators-q756d" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.866245 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/798d0a0f-e302-48af-8ab9-1640b68805b0-utilities\") pod \"certified-operators-q756d\" (UID: \"798d0a0f-e302-48af-8ab9-1640b68805b0\") " pod="openshift-marketplace/certified-operators-q756d" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.866309 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/798d0a0f-e302-48af-8ab9-1640b68805b0-catalog-content\") pod \"certified-operators-q756d\" (UID: \"798d0a0f-e302-48af-8ab9-1640b68805b0\") " pod="openshift-marketplace/certified-operators-q756d" Dec 05 09:16:18 crc kubenswrapper[4997]: I1205 09:16:18.884592 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf8f8\" (UniqueName: \"kubernetes.io/projected/798d0a0f-e302-48af-8ab9-1640b68805b0-kube-api-access-bf8f8\") pod \"certified-operators-q756d\" (UID: \"798d0a0f-e302-48af-8ab9-1640b68805b0\") " pod="openshift-marketplace/certified-operators-q756d" Dec 05 09:16:19 crc kubenswrapper[4997]: I1205 09:16:19.000141 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q756d" Dec 05 09:16:19 crc kubenswrapper[4997]: I1205 09:16:19.350373 4997 scope.go:117] "RemoveContainer" containerID="6877c786e03b895e087df0c7a627e2791e8fe2ee547c29e3c80826e467e7b60f" Dec 05 09:16:19 crc kubenswrapper[4997]: I1205 09:16:19.353311 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q756d"] Dec 05 09:16:19 crc kubenswrapper[4997]: I1205 09:16:19.420120 4997 scope.go:117] "RemoveContainer" containerID="e6e03db2d9a8ebc1ddf973faa2a742ae658c2c5b445c85b578b67999c85bce05" Dec 05 09:16:19 crc kubenswrapper[4997]: I1205 09:16:19.470690 4997 scope.go:117] "RemoveContainer" containerID="6bfc5691dd87a09f74025bf1c7061f4c87d8119250649b4f0313022fbf2d1cca" Dec 05 09:16:19 crc kubenswrapper[4997]: I1205 09:16:19.511470 4997 scope.go:117] "RemoveContainer" containerID="3597ec59688cc35de0fa3dfb30c5c65b2b793c4c827be3d2df03aa8f54dc9500" Dec 05 09:16:19 crc kubenswrapper[4997]: I1205 09:16:19.530038 4997 scope.go:117] "RemoveContainer" containerID="644de32d915ea6906569388efab86698cefe1488837295c1ca10f7dedb7a5c18" Dec 05 09:16:19 crc kubenswrapper[4997]: I1205 09:16:19.703215 4997 generic.go:334] "Generic (PLEG): container finished" podID="798d0a0f-e302-48af-8ab9-1640b68805b0" containerID="16262c2f0cc56dbf5eea4ac27e9a9c81239ab120c496190ed413ea172a10539b" exitCode=0 Dec 05 09:16:19 crc kubenswrapper[4997]: I1205 09:16:19.703285 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q756d" event={"ID":"798d0a0f-e302-48af-8ab9-1640b68805b0","Type":"ContainerDied","Data":"16262c2f0cc56dbf5eea4ac27e9a9c81239ab120c496190ed413ea172a10539b"} Dec 05 09:16:19 crc kubenswrapper[4997]: I1205 09:16:19.705527 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q756d" event={"ID":"798d0a0f-e302-48af-8ab9-1640b68805b0","Type":"ContainerStarted","Data":"3ac47901ca3a1b4b16d1c7dbc4373cc54a7f95394be592044959e96e8b467b9d"} Dec 05 09:16:19 crc kubenswrapper[4997]: I1205 09:16:19.705757 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:16:20 crc kubenswrapper[4997]: I1205 09:16:20.719523 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q756d" event={"ID":"798d0a0f-e302-48af-8ab9-1640b68805b0","Type":"ContainerStarted","Data":"3e2caee9732b99536841fabdf6e4588ff399bb348b162fc7dc9e3caecda09d27"} Dec 05 09:16:21 crc kubenswrapper[4997]: I1205 09:16:21.732088 4997 generic.go:334] "Generic (PLEG): container finished" podID="798d0a0f-e302-48af-8ab9-1640b68805b0" containerID="3e2caee9732b99536841fabdf6e4588ff399bb348b162fc7dc9e3caecda09d27" exitCode=0 Dec 05 09:16:21 crc kubenswrapper[4997]: I1205 09:16:21.732140 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q756d" event={"ID":"798d0a0f-e302-48af-8ab9-1640b68805b0","Type":"ContainerDied","Data":"3e2caee9732b99536841fabdf6e4588ff399bb348b162fc7dc9e3caecda09d27"} Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.746204 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-97rb6"] Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.748452 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q756d" event={"ID":"798d0a0f-e302-48af-8ab9-1640b68805b0","Type":"ContainerStarted","Data":"a8118e992d3435b5c33f2a900c3191c6a582053d6ee2590fd33e3ad516d53974"} Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.748577 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.752951 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.753163 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.753325 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.755880 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.761600 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-networker-r2z6x"] Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.763374 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.767552 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.771315 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-4542r" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.780813 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-97rb6"] Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.794345 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-networker-r2z6x"] Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.798205 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q756d" podStartSLOduration=2.378239731 podStartE2EDuration="4.798191722s" podCreationTimestamp="2025-12-05 09:16:18 +0000 UTC" firstStartedPulling="2025-12-05 09:16:19.705370283 +0000 UTC m=+8480.234277554" lastFinishedPulling="2025-12-05 09:16:22.125322284 +0000 UTC m=+8482.654229545" observedRunningTime="2025-12-05 09:16:22.785363106 +0000 UTC m=+8483.314270377" watchObservedRunningTime="2025-12-05 09:16:22.798191722 +0000 UTC m=+8483.327098983" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.847755 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8ctt\" (UniqueName: \"kubernetes.io/projected/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-kube-api-access-r8ctt\") pod \"bootstrap-openstack-openstack-cell1-97rb6\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.847886 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-inventory\") pod \"bootstrap-openstack-openstack-networker-r2z6x\" (UID: \"95a0bda3-99d3-46df-8ca6-fd85db893d66\") " pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.847926 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xpnl\" (UniqueName: \"kubernetes.io/projected/95a0bda3-99d3-46df-8ca6-fd85db893d66-kube-api-access-7xpnl\") pod \"bootstrap-openstack-openstack-networker-r2z6x\" (UID: \"95a0bda3-99d3-46df-8ca6-fd85db893d66\") " pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.847954 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-networker-r2z6x\" (UID: \"95a0bda3-99d3-46df-8ca6-fd85db893d66\") " pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.847980 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-ceph\") pod \"bootstrap-openstack-openstack-cell1-97rb6\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.847997 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-97rb6\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.848044 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-ssh-key\") pod \"bootstrap-openstack-openstack-networker-r2z6x\" (UID: \"95a0bda3-99d3-46df-8ca6-fd85db893d66\") " pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.848083 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-inventory\") pod \"bootstrap-openstack-openstack-cell1-97rb6\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.848100 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-97rb6\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.949882 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xpnl\" (UniqueName: \"kubernetes.io/projected/95a0bda3-99d3-46df-8ca6-fd85db893d66-kube-api-access-7xpnl\") pod \"bootstrap-openstack-openstack-networker-r2z6x\" (UID: \"95a0bda3-99d3-46df-8ca6-fd85db893d66\") " pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.950270 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-networker-r2z6x\" (UID: \"95a0bda3-99d3-46df-8ca6-fd85db893d66\") " pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.950409 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-97rb6\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.950534 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-ceph\") pod \"bootstrap-openstack-openstack-cell1-97rb6\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.950725 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-ssh-key\") pod \"bootstrap-openstack-openstack-networker-r2z6x\" (UID: \"95a0bda3-99d3-46df-8ca6-fd85db893d66\") " pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.950881 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-inventory\") pod \"bootstrap-openstack-openstack-cell1-97rb6\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.951025 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-97rb6\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.951673 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8ctt\" (UniqueName: \"kubernetes.io/projected/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-kube-api-access-r8ctt\") pod \"bootstrap-openstack-openstack-cell1-97rb6\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.952068 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-inventory\") pod \"bootstrap-openstack-openstack-networker-r2z6x\" (UID: \"95a0bda3-99d3-46df-8ca6-fd85db893d66\") " pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.956659 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-ssh-key\") pod \"bootstrap-openstack-openstack-networker-r2z6x\" (UID: \"95a0bda3-99d3-46df-8ca6-fd85db893d66\") " pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.958278 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-networker-r2z6x\" (UID: \"95a0bda3-99d3-46df-8ca6-fd85db893d66\") " pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.958486 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-ceph\") pod \"bootstrap-openstack-openstack-cell1-97rb6\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.960084 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-97rb6\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.960405 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-inventory\") pod \"bootstrap-openstack-openstack-cell1-97rb6\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.962931 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-97rb6\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.967248 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-inventory\") pod \"bootstrap-openstack-openstack-networker-r2z6x\" (UID: \"95a0bda3-99d3-46df-8ca6-fd85db893d66\") " pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.978150 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8ctt\" (UniqueName: \"kubernetes.io/projected/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-kube-api-access-r8ctt\") pod \"bootstrap-openstack-openstack-cell1-97rb6\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:16:22 crc kubenswrapper[4997]: I1205 09:16:22.978717 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xpnl\" (UniqueName: \"kubernetes.io/projected/95a0bda3-99d3-46df-8ca6-fd85db893d66-kube-api-access-7xpnl\") pod \"bootstrap-openstack-openstack-networker-r2z6x\" (UID: \"95a0bda3-99d3-46df-8ca6-fd85db893d66\") " pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" Dec 05 09:16:23 crc kubenswrapper[4997]: I1205 09:16:23.086917 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:16:23 crc kubenswrapper[4997]: I1205 09:16:23.094301 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" Dec 05 09:16:23 crc kubenswrapper[4997]: I1205 09:16:23.660769 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-97rb6"] Dec 05 09:16:23 crc kubenswrapper[4997]: W1205 09:16:23.751325 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95a0bda3_99d3_46df_8ca6_fd85db893d66.slice/crio-0ea90a87b307638b9403f77080e8b3a38575c848ea4277e0219bb79f5eb880a2 WatchSource:0}: Error finding container 0ea90a87b307638b9403f77080e8b3a38575c848ea4277e0219bb79f5eb880a2: Status 404 returned error can't find the container with id 0ea90a87b307638b9403f77080e8b3a38575c848ea4277e0219bb79f5eb880a2 Dec 05 09:16:23 crc kubenswrapper[4997]: I1205 09:16:23.763335 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-networker-r2z6x"] Dec 05 09:16:23 crc kubenswrapper[4997]: I1205 09:16:23.763382 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" event={"ID":"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00","Type":"ContainerStarted","Data":"3829e6ea368ecada53c103b3a1e975c2bf0e14699b01559530b3d523dc716a68"} Dec 05 09:16:24 crc kubenswrapper[4997]: I1205 09:16:24.766965 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" event={"ID":"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00","Type":"ContainerStarted","Data":"876d6cc467489ff1e2b366d251f171f44ef132fcdb5a004aa4a2f3406b067df3"} Dec 05 09:16:24 crc kubenswrapper[4997]: I1205 09:16:24.781568 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" event={"ID":"95a0bda3-99d3-46df-8ca6-fd85db893d66","Type":"ContainerStarted","Data":"deecce5663936bdaea0b480b991af1b92b6484e851695fa0e9fd58be159f34f5"} Dec 05 09:16:24 crc kubenswrapper[4997]: I1205 09:16:24.781630 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" event={"ID":"95a0bda3-99d3-46df-8ca6-fd85db893d66","Type":"ContainerStarted","Data":"0ea90a87b307638b9403f77080e8b3a38575c848ea4277e0219bb79f5eb880a2"} Dec 05 09:16:24 crc kubenswrapper[4997]: I1205 09:16:24.795120 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" podStartSLOduration=2.376620541 podStartE2EDuration="2.795096021s" podCreationTimestamp="2025-12-05 09:16:22 +0000 UTC" firstStartedPulling="2025-12-05 09:16:23.665390247 +0000 UTC m=+8484.194297508" lastFinishedPulling="2025-12-05 09:16:24.083865727 +0000 UTC m=+8484.612772988" observedRunningTime="2025-12-05 09:16:24.792699537 +0000 UTC m=+8485.321606808" watchObservedRunningTime="2025-12-05 09:16:24.795096021 +0000 UTC m=+8485.324003292" Dec 05 09:16:24 crc kubenswrapper[4997]: I1205 09:16:24.837306 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" podStartSLOduration=2.332797108 podStartE2EDuration="2.83728245s" podCreationTimestamp="2025-12-05 09:16:22 +0000 UTC" firstStartedPulling="2025-12-05 09:16:23.755439739 +0000 UTC m=+8484.284347000" lastFinishedPulling="2025-12-05 09:16:24.259925081 +0000 UTC m=+8484.788832342" observedRunningTime="2025-12-05 09:16:24.824601667 +0000 UTC m=+8485.353508938" watchObservedRunningTime="2025-12-05 09:16:24.83728245 +0000 UTC m=+8485.366189711" Dec 05 09:16:26 crc kubenswrapper[4997]: I1205 09:16:26.031120 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-nlscc"] Dec 05 09:16:26 crc kubenswrapper[4997]: I1205 09:16:26.048270 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-nlscc"] Dec 05 09:16:27 crc kubenswrapper[4997]: I1205 09:16:27.759939 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c864cae-1e34-46f0-8f85-c83c8963a6bc" path="/var/lib/kubelet/pods/6c864cae-1e34-46f0-8f85-c83c8963a6bc/volumes" Dec 05 09:16:29 crc kubenswrapper[4997]: I1205 09:16:29.001743 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q756d" Dec 05 09:16:29 crc kubenswrapper[4997]: I1205 09:16:29.002135 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q756d" Dec 05 09:16:29 crc kubenswrapper[4997]: I1205 09:16:29.144987 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q756d" Dec 05 09:16:29 crc kubenswrapper[4997]: I1205 09:16:29.872185 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q756d" Dec 05 09:16:29 crc kubenswrapper[4997]: I1205 09:16:29.923587 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q756d"] Dec 05 09:16:31 crc kubenswrapper[4997]: I1205 09:16:31.839967 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q756d" podUID="798d0a0f-e302-48af-8ab9-1640b68805b0" containerName="registry-server" containerID="cri-o://a8118e992d3435b5c33f2a900c3191c6a582053d6ee2590fd33e3ad516d53974" gracePeriod=2 Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.285890 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q756d" Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.449607 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/798d0a0f-e302-48af-8ab9-1640b68805b0-utilities\") pod \"798d0a0f-e302-48af-8ab9-1640b68805b0\" (UID: \"798d0a0f-e302-48af-8ab9-1640b68805b0\") " Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.449791 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/798d0a0f-e302-48af-8ab9-1640b68805b0-catalog-content\") pod \"798d0a0f-e302-48af-8ab9-1640b68805b0\" (UID: \"798d0a0f-e302-48af-8ab9-1640b68805b0\") " Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.450025 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf8f8\" (UniqueName: \"kubernetes.io/projected/798d0a0f-e302-48af-8ab9-1640b68805b0-kube-api-access-bf8f8\") pod \"798d0a0f-e302-48af-8ab9-1640b68805b0\" (UID: \"798d0a0f-e302-48af-8ab9-1640b68805b0\") " Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.451650 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/798d0a0f-e302-48af-8ab9-1640b68805b0-utilities" (OuterVolumeSpecName: "utilities") pod "798d0a0f-e302-48af-8ab9-1640b68805b0" (UID: "798d0a0f-e302-48af-8ab9-1640b68805b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.457828 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/798d0a0f-e302-48af-8ab9-1640b68805b0-kube-api-access-bf8f8" (OuterVolumeSpecName: "kube-api-access-bf8f8") pod "798d0a0f-e302-48af-8ab9-1640b68805b0" (UID: "798d0a0f-e302-48af-8ab9-1640b68805b0"). InnerVolumeSpecName "kube-api-access-bf8f8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.498346 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/798d0a0f-e302-48af-8ab9-1640b68805b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "798d0a0f-e302-48af-8ab9-1640b68805b0" (UID: "798d0a0f-e302-48af-8ab9-1640b68805b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.553389 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf8f8\" (UniqueName: \"kubernetes.io/projected/798d0a0f-e302-48af-8ab9-1640b68805b0-kube-api-access-bf8f8\") on node \"crc\" DevicePath \"\"" Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.553446 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/798d0a0f-e302-48af-8ab9-1640b68805b0-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.553464 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/798d0a0f-e302-48af-8ab9-1640b68805b0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.852748 4997 generic.go:334] "Generic (PLEG): container finished" podID="798d0a0f-e302-48af-8ab9-1640b68805b0" containerID="a8118e992d3435b5c33f2a900c3191c6a582053d6ee2590fd33e3ad516d53974" exitCode=0 Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.852842 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q756d" event={"ID":"798d0a0f-e302-48af-8ab9-1640b68805b0","Type":"ContainerDied","Data":"a8118e992d3435b5c33f2a900c3191c6a582053d6ee2590fd33e3ad516d53974"} Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.853157 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q756d" event={"ID":"798d0a0f-e302-48af-8ab9-1640b68805b0","Type":"ContainerDied","Data":"3ac47901ca3a1b4b16d1c7dbc4373cc54a7f95394be592044959e96e8b467b9d"} Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.853182 4997 scope.go:117] "RemoveContainer" containerID="a8118e992d3435b5c33f2a900c3191c6a582053d6ee2590fd33e3ad516d53974" Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.852883 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q756d" Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.877416 4997 scope.go:117] "RemoveContainer" containerID="3e2caee9732b99536841fabdf6e4588ff399bb348b162fc7dc9e3caecda09d27" Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.891225 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q756d"] Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.901586 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-q756d"] Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.911417 4997 scope.go:117] "RemoveContainer" containerID="16262c2f0cc56dbf5eea4ac27e9a9c81239ab120c496190ed413ea172a10539b" Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.952408 4997 scope.go:117] "RemoveContainer" containerID="a8118e992d3435b5c33f2a900c3191c6a582053d6ee2590fd33e3ad516d53974" Dec 05 09:16:32 crc kubenswrapper[4997]: E1205 09:16:32.952937 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8118e992d3435b5c33f2a900c3191c6a582053d6ee2590fd33e3ad516d53974\": container with ID starting with a8118e992d3435b5c33f2a900c3191c6a582053d6ee2590fd33e3ad516d53974 not found: ID does not exist" containerID="a8118e992d3435b5c33f2a900c3191c6a582053d6ee2590fd33e3ad516d53974" Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.952977 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8118e992d3435b5c33f2a900c3191c6a582053d6ee2590fd33e3ad516d53974"} err="failed to get container status \"a8118e992d3435b5c33f2a900c3191c6a582053d6ee2590fd33e3ad516d53974\": rpc error: code = NotFound desc = could not find container \"a8118e992d3435b5c33f2a900c3191c6a582053d6ee2590fd33e3ad516d53974\": container with ID starting with a8118e992d3435b5c33f2a900c3191c6a582053d6ee2590fd33e3ad516d53974 not found: ID does not exist" Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.953002 4997 scope.go:117] "RemoveContainer" containerID="3e2caee9732b99536841fabdf6e4588ff399bb348b162fc7dc9e3caecda09d27" Dec 05 09:16:32 crc kubenswrapper[4997]: E1205 09:16:32.953240 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e2caee9732b99536841fabdf6e4588ff399bb348b162fc7dc9e3caecda09d27\": container with ID starting with 3e2caee9732b99536841fabdf6e4588ff399bb348b162fc7dc9e3caecda09d27 not found: ID does not exist" containerID="3e2caee9732b99536841fabdf6e4588ff399bb348b162fc7dc9e3caecda09d27" Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.953327 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e2caee9732b99536841fabdf6e4588ff399bb348b162fc7dc9e3caecda09d27"} err="failed to get container status \"3e2caee9732b99536841fabdf6e4588ff399bb348b162fc7dc9e3caecda09d27\": rpc error: code = NotFound desc = could not find container \"3e2caee9732b99536841fabdf6e4588ff399bb348b162fc7dc9e3caecda09d27\": container with ID starting with 3e2caee9732b99536841fabdf6e4588ff399bb348b162fc7dc9e3caecda09d27 not found: ID does not exist" Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.953433 4997 scope.go:117] "RemoveContainer" containerID="16262c2f0cc56dbf5eea4ac27e9a9c81239ab120c496190ed413ea172a10539b" Dec 05 09:16:32 crc kubenswrapper[4997]: E1205 09:16:32.953777 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16262c2f0cc56dbf5eea4ac27e9a9c81239ab120c496190ed413ea172a10539b\": container with ID starting with 16262c2f0cc56dbf5eea4ac27e9a9c81239ab120c496190ed413ea172a10539b not found: ID does not exist" containerID="16262c2f0cc56dbf5eea4ac27e9a9c81239ab120c496190ed413ea172a10539b" Dec 05 09:16:32 crc kubenswrapper[4997]: I1205 09:16:32.953834 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16262c2f0cc56dbf5eea4ac27e9a9c81239ab120c496190ed413ea172a10539b"} err="failed to get container status \"16262c2f0cc56dbf5eea4ac27e9a9c81239ab120c496190ed413ea172a10539b\": rpc error: code = NotFound desc = could not find container \"16262c2f0cc56dbf5eea4ac27e9a9c81239ab120c496190ed413ea172a10539b\": container with ID starting with 16262c2f0cc56dbf5eea4ac27e9a9c81239ab120c496190ed413ea172a10539b not found: ID does not exist" Dec 05 09:16:33 crc kubenswrapper[4997]: I1205 09:16:33.761151 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="798d0a0f-e302-48af-8ab9-1640b68805b0" path="/var/lib/kubelet/pods/798d0a0f-e302-48af-8ab9-1640b68805b0/volumes" Dec 05 09:17:19 crc kubenswrapper[4997]: I1205 09:17:19.647094 4997 scope.go:117] "RemoveContainer" containerID="4e03ef7591ecc13bc654b364b37b1e8b4da87249ab0b97600dc99b42895c3b90" Dec 05 09:18:19 crc kubenswrapper[4997]: I1205 09:18:19.770091 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:18:19 crc kubenswrapper[4997]: I1205 09:18:19.770682 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:18:49 crc kubenswrapper[4997]: I1205 09:18:49.769745 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:18:49 crc kubenswrapper[4997]: I1205 09:18:49.770508 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:19:19 crc kubenswrapper[4997]: I1205 09:19:19.771326 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:19:19 crc kubenswrapper[4997]: I1205 09:19:19.772310 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:19:19 crc kubenswrapper[4997]: I1205 09:19:19.772382 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 09:19:19 crc kubenswrapper[4997]: I1205 09:19:19.774083 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f8205236afd16f8a0c6d433a5ab02465034afc3f440c01d487df254a7a73dfb2"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:19:19 crc kubenswrapper[4997]: I1205 09:19:19.774271 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://f8205236afd16f8a0c6d433a5ab02465034afc3f440c01d487df254a7a73dfb2" gracePeriod=600 Dec 05 09:19:20 crc kubenswrapper[4997]: I1205 09:19:20.479256 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="f8205236afd16f8a0c6d433a5ab02465034afc3f440c01d487df254a7a73dfb2" exitCode=0 Dec 05 09:19:20 crc kubenswrapper[4997]: I1205 09:19:20.479370 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"f8205236afd16f8a0c6d433a5ab02465034afc3f440c01d487df254a7a73dfb2"} Dec 05 09:19:20 crc kubenswrapper[4997]: I1205 09:19:20.480258 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a"} Dec 05 09:19:20 crc kubenswrapper[4997]: I1205 09:19:20.480374 4997 scope.go:117] "RemoveContainer" containerID="e6d271b834369b6363e20100fcc389ad14b07ed8a1f6fcabcd9aa8d1a60962f9" Dec 05 09:19:23 crc kubenswrapper[4997]: I1205 09:19:23.516993 4997 generic.go:334] "Generic (PLEG): container finished" podID="e45c1dd1-eab5-4852-bb9d-ecee8ab68e00" containerID="876d6cc467489ff1e2b366d251f171f44ef132fcdb5a004aa4a2f3406b067df3" exitCode=0 Dec 05 09:19:23 crc kubenswrapper[4997]: I1205 09:19:23.517065 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" event={"ID":"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00","Type":"ContainerDied","Data":"876d6cc467489ff1e2b366d251f171f44ef132fcdb5a004aa4a2f3406b067df3"} Dec 05 09:19:24 crc kubenswrapper[4997]: I1205 09:19:24.992433 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.043119 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8ctt\" (UniqueName: \"kubernetes.io/projected/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-kube-api-access-r8ctt\") pod \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.043198 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-inventory\") pod \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.043431 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-ssh-key\") pod \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.043586 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-ceph\") pod \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.043672 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-bootstrap-combined-ca-bundle\") pod \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\" (UID: \"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00\") " Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.053655 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-kube-api-access-r8ctt" (OuterVolumeSpecName: "kube-api-access-r8ctt") pod "e45c1dd1-eab5-4852-bb9d-ecee8ab68e00" (UID: "e45c1dd1-eab5-4852-bb9d-ecee8ab68e00"). InnerVolumeSpecName "kube-api-access-r8ctt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.053511 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-ceph" (OuterVolumeSpecName: "ceph") pod "e45c1dd1-eab5-4852-bb9d-ecee8ab68e00" (UID: "e45c1dd1-eab5-4852-bb9d-ecee8ab68e00"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.060545 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "e45c1dd1-eab5-4852-bb9d-ecee8ab68e00" (UID: "e45c1dd1-eab5-4852-bb9d-ecee8ab68e00"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.080750 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-inventory" (OuterVolumeSpecName: "inventory") pod "e45c1dd1-eab5-4852-bb9d-ecee8ab68e00" (UID: "e45c1dd1-eab5-4852-bb9d-ecee8ab68e00"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.082345 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e45c1dd1-eab5-4852-bb9d-ecee8ab68e00" (UID: "e45c1dd1-eab5-4852-bb9d-ecee8ab68e00"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.146058 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.146325 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.146500 4997 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.146656 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8ctt\" (UniqueName: \"kubernetes.io/projected/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-kube-api-access-r8ctt\") on node \"crc\" DevicePath \"\"" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.146791 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e45c1dd1-eab5-4852-bb9d-ecee8ab68e00-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.538441 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" event={"ID":"e45c1dd1-eab5-4852-bb9d-ecee8ab68e00","Type":"ContainerDied","Data":"3829e6ea368ecada53c103b3a1e975c2bf0e14699b01559530b3d523dc716a68"} Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.538530 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3829e6ea368ecada53c103b3a1e975c2bf0e14699b01559530b3d523dc716a68" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.538471 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-97rb6" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.659359 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-b22xv"] Dec 05 09:19:25 crc kubenswrapper[4997]: E1205 09:19:25.659815 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="798d0a0f-e302-48af-8ab9-1640b68805b0" containerName="extract-utilities" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.659838 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="798d0a0f-e302-48af-8ab9-1640b68805b0" containerName="extract-utilities" Dec 05 09:19:25 crc kubenswrapper[4997]: E1205 09:19:25.659858 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="798d0a0f-e302-48af-8ab9-1640b68805b0" containerName="extract-content" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.659865 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="798d0a0f-e302-48af-8ab9-1640b68805b0" containerName="extract-content" Dec 05 09:19:25 crc kubenswrapper[4997]: E1205 09:19:25.659880 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="798d0a0f-e302-48af-8ab9-1640b68805b0" containerName="registry-server" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.659886 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="798d0a0f-e302-48af-8ab9-1640b68805b0" containerName="registry-server" Dec 05 09:19:25 crc kubenswrapper[4997]: E1205 09:19:25.659909 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e45c1dd1-eab5-4852-bb9d-ecee8ab68e00" containerName="bootstrap-openstack-openstack-cell1" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.659917 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e45c1dd1-eab5-4852-bb9d-ecee8ab68e00" containerName="bootstrap-openstack-openstack-cell1" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.660348 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e45c1dd1-eab5-4852-bb9d-ecee8ab68e00" containerName="bootstrap-openstack-openstack-cell1" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.660374 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="798d0a0f-e302-48af-8ab9-1640b68805b0" containerName="registry-server" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.661147 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-b22xv" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.664230 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.664430 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.670006 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-b22xv"] Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.761193 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-ssh-key\") pod \"download-cache-openstack-openstack-cell1-b22xv\" (UID: \"999c70de-3370-4ce0-947a-b585dcd879d7\") " pod="openstack/download-cache-openstack-openstack-cell1-b22xv" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.761251 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndhrr\" (UniqueName: \"kubernetes.io/projected/999c70de-3370-4ce0-947a-b585dcd879d7-kube-api-access-ndhrr\") pod \"download-cache-openstack-openstack-cell1-b22xv\" (UID: \"999c70de-3370-4ce0-947a-b585dcd879d7\") " pod="openstack/download-cache-openstack-openstack-cell1-b22xv" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.761545 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-ceph\") pod \"download-cache-openstack-openstack-cell1-b22xv\" (UID: \"999c70de-3370-4ce0-947a-b585dcd879d7\") " pod="openstack/download-cache-openstack-openstack-cell1-b22xv" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.761606 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-inventory\") pod \"download-cache-openstack-openstack-cell1-b22xv\" (UID: \"999c70de-3370-4ce0-947a-b585dcd879d7\") " pod="openstack/download-cache-openstack-openstack-cell1-b22xv" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.863336 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-ceph\") pod \"download-cache-openstack-openstack-cell1-b22xv\" (UID: \"999c70de-3370-4ce0-947a-b585dcd879d7\") " pod="openstack/download-cache-openstack-openstack-cell1-b22xv" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.863384 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-inventory\") pod \"download-cache-openstack-openstack-cell1-b22xv\" (UID: \"999c70de-3370-4ce0-947a-b585dcd879d7\") " pod="openstack/download-cache-openstack-openstack-cell1-b22xv" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.863479 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-ssh-key\") pod \"download-cache-openstack-openstack-cell1-b22xv\" (UID: \"999c70de-3370-4ce0-947a-b585dcd879d7\") " pod="openstack/download-cache-openstack-openstack-cell1-b22xv" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.863505 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndhrr\" (UniqueName: \"kubernetes.io/projected/999c70de-3370-4ce0-947a-b585dcd879d7-kube-api-access-ndhrr\") pod \"download-cache-openstack-openstack-cell1-b22xv\" (UID: \"999c70de-3370-4ce0-947a-b585dcd879d7\") " pod="openstack/download-cache-openstack-openstack-cell1-b22xv" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.873402 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-ssh-key\") pod \"download-cache-openstack-openstack-cell1-b22xv\" (UID: \"999c70de-3370-4ce0-947a-b585dcd879d7\") " pod="openstack/download-cache-openstack-openstack-cell1-b22xv" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.873433 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-inventory\") pod \"download-cache-openstack-openstack-cell1-b22xv\" (UID: \"999c70de-3370-4ce0-947a-b585dcd879d7\") " pod="openstack/download-cache-openstack-openstack-cell1-b22xv" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.873433 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-ceph\") pod \"download-cache-openstack-openstack-cell1-b22xv\" (UID: \"999c70de-3370-4ce0-947a-b585dcd879d7\") " pod="openstack/download-cache-openstack-openstack-cell1-b22xv" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.879395 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndhrr\" (UniqueName: \"kubernetes.io/projected/999c70de-3370-4ce0-947a-b585dcd879d7-kube-api-access-ndhrr\") pod \"download-cache-openstack-openstack-cell1-b22xv\" (UID: \"999c70de-3370-4ce0-947a-b585dcd879d7\") " pod="openstack/download-cache-openstack-openstack-cell1-b22xv" Dec 05 09:19:25 crc kubenswrapper[4997]: I1205 09:19:25.983308 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-b22xv" Dec 05 09:19:26 crc kubenswrapper[4997]: I1205 09:19:26.490205 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-b22xv"] Dec 05 09:19:26 crc kubenswrapper[4997]: I1205 09:19:26.550179 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-b22xv" event={"ID":"999c70de-3370-4ce0-947a-b585dcd879d7","Type":"ContainerStarted","Data":"9f19eae73e9d40ad5d766f6c0d36224400cf2b9900c5caca5cb442fb3937bb40"} Dec 05 09:19:27 crc kubenswrapper[4997]: I1205 09:19:27.559823 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-b22xv" event={"ID":"999c70de-3370-4ce0-947a-b585dcd879d7","Type":"ContainerStarted","Data":"d5f519bfe710282a2d845da1bee779fa7aba49a5c0a165e9d8dafcf447703c4b"} Dec 05 09:19:27 crc kubenswrapper[4997]: I1205 09:19:27.563832 4997 generic.go:334] "Generic (PLEG): container finished" podID="95a0bda3-99d3-46df-8ca6-fd85db893d66" containerID="deecce5663936bdaea0b480b991af1b92b6484e851695fa0e9fd58be159f34f5" exitCode=0 Dec 05 09:19:27 crc kubenswrapper[4997]: I1205 09:19:27.563866 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" event={"ID":"95a0bda3-99d3-46df-8ca6-fd85db893d66","Type":"ContainerDied","Data":"deecce5663936bdaea0b480b991af1b92b6484e851695fa0e9fd58be159f34f5"} Dec 05 09:19:27 crc kubenswrapper[4997]: I1205 09:19:27.583926 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-b22xv" podStartSLOduration=2.175611804 podStartE2EDuration="2.583903918s" podCreationTimestamp="2025-12-05 09:19:25 +0000 UTC" firstStartedPulling="2025-12-05 09:19:26.497865604 +0000 UTC m=+8667.026772865" lastFinishedPulling="2025-12-05 09:19:26.906157728 +0000 UTC m=+8667.435064979" observedRunningTime="2025-12-05 09:19:27.579295714 +0000 UTC m=+8668.108202995" watchObservedRunningTime="2025-12-05 09:19:27.583903918 +0000 UTC m=+8668.112811179" Dec 05 09:19:28 crc kubenswrapper[4997]: I1205 09:19:28.986355 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.041095 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-bootstrap-combined-ca-bundle\") pod \"95a0bda3-99d3-46df-8ca6-fd85db893d66\" (UID: \"95a0bda3-99d3-46df-8ca6-fd85db893d66\") " Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.041178 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-inventory\") pod \"95a0bda3-99d3-46df-8ca6-fd85db893d66\" (UID: \"95a0bda3-99d3-46df-8ca6-fd85db893d66\") " Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.041268 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-ssh-key\") pod \"95a0bda3-99d3-46df-8ca6-fd85db893d66\" (UID: \"95a0bda3-99d3-46df-8ca6-fd85db893d66\") " Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.041354 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xpnl\" (UniqueName: \"kubernetes.io/projected/95a0bda3-99d3-46df-8ca6-fd85db893d66-kube-api-access-7xpnl\") pod \"95a0bda3-99d3-46df-8ca6-fd85db893d66\" (UID: \"95a0bda3-99d3-46df-8ca6-fd85db893d66\") " Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.046908 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "95a0bda3-99d3-46df-8ca6-fd85db893d66" (UID: "95a0bda3-99d3-46df-8ca6-fd85db893d66"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.047759 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95a0bda3-99d3-46df-8ca6-fd85db893d66-kube-api-access-7xpnl" (OuterVolumeSpecName: "kube-api-access-7xpnl") pod "95a0bda3-99d3-46df-8ca6-fd85db893d66" (UID: "95a0bda3-99d3-46df-8ca6-fd85db893d66"). InnerVolumeSpecName "kube-api-access-7xpnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.073024 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-inventory" (OuterVolumeSpecName: "inventory") pod "95a0bda3-99d3-46df-8ca6-fd85db893d66" (UID: "95a0bda3-99d3-46df-8ca6-fd85db893d66"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.085596 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "95a0bda3-99d3-46df-8ca6-fd85db893d66" (UID: "95a0bda3-99d3-46df-8ca6-fd85db893d66"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.144587 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.144649 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xpnl\" (UniqueName: \"kubernetes.io/projected/95a0bda3-99d3-46df-8ca6-fd85db893d66-kube-api-access-7xpnl\") on node \"crc\" DevicePath \"\"" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.144659 4997 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.144669 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95a0bda3-99d3-46df-8ca6-fd85db893d66-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.585359 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" event={"ID":"95a0bda3-99d3-46df-8ca6-fd85db893d66","Type":"ContainerDied","Data":"0ea90a87b307638b9403f77080e8b3a38575c848ea4277e0219bb79f5eb880a2"} Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.585706 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ea90a87b307638b9403f77080e8b3a38575c848ea4277e0219bb79f5eb880a2" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.585448 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-r2z6x" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.679700 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-networker-nsk64"] Dec 05 09:19:29 crc kubenswrapper[4997]: E1205 09:19:29.680214 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a0bda3-99d3-46df-8ca6-fd85db893d66" containerName="bootstrap-openstack-openstack-networker" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.680237 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a0bda3-99d3-46df-8ca6-fd85db893d66" containerName="bootstrap-openstack-openstack-networker" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.680452 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="95a0bda3-99d3-46df-8ca6-fd85db893d66" containerName="bootstrap-openstack-openstack-networker" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.681340 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-nsk64" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.684724 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-4542r" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.689823 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.692499 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-networker-nsk64"] Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.761796 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c2dr\" (UniqueName: \"kubernetes.io/projected/73bb544f-2408-4f58-94eb-7b5116d41848-kube-api-access-6c2dr\") pod \"download-cache-openstack-openstack-networker-nsk64\" (UID: \"73bb544f-2408-4f58-94eb-7b5116d41848\") " pod="openstack/download-cache-openstack-openstack-networker-nsk64" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.761944 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73bb544f-2408-4f58-94eb-7b5116d41848-inventory\") pod \"download-cache-openstack-openstack-networker-nsk64\" (UID: \"73bb544f-2408-4f58-94eb-7b5116d41848\") " pod="openstack/download-cache-openstack-openstack-networker-nsk64" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.762042 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73bb544f-2408-4f58-94eb-7b5116d41848-ssh-key\") pod \"download-cache-openstack-openstack-networker-nsk64\" (UID: \"73bb544f-2408-4f58-94eb-7b5116d41848\") " pod="openstack/download-cache-openstack-openstack-networker-nsk64" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.864008 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c2dr\" (UniqueName: \"kubernetes.io/projected/73bb544f-2408-4f58-94eb-7b5116d41848-kube-api-access-6c2dr\") pod \"download-cache-openstack-openstack-networker-nsk64\" (UID: \"73bb544f-2408-4f58-94eb-7b5116d41848\") " pod="openstack/download-cache-openstack-openstack-networker-nsk64" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.864166 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73bb544f-2408-4f58-94eb-7b5116d41848-inventory\") pod \"download-cache-openstack-openstack-networker-nsk64\" (UID: \"73bb544f-2408-4f58-94eb-7b5116d41848\") " pod="openstack/download-cache-openstack-openstack-networker-nsk64" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.864737 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73bb544f-2408-4f58-94eb-7b5116d41848-ssh-key\") pod \"download-cache-openstack-openstack-networker-nsk64\" (UID: \"73bb544f-2408-4f58-94eb-7b5116d41848\") " pod="openstack/download-cache-openstack-openstack-networker-nsk64" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.871240 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73bb544f-2408-4f58-94eb-7b5116d41848-inventory\") pod \"download-cache-openstack-openstack-networker-nsk64\" (UID: \"73bb544f-2408-4f58-94eb-7b5116d41848\") " pod="openstack/download-cache-openstack-openstack-networker-nsk64" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.874222 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73bb544f-2408-4f58-94eb-7b5116d41848-ssh-key\") pod \"download-cache-openstack-openstack-networker-nsk64\" (UID: \"73bb544f-2408-4f58-94eb-7b5116d41848\") " pod="openstack/download-cache-openstack-openstack-networker-nsk64" Dec 05 09:19:29 crc kubenswrapper[4997]: I1205 09:19:29.881120 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c2dr\" (UniqueName: \"kubernetes.io/projected/73bb544f-2408-4f58-94eb-7b5116d41848-kube-api-access-6c2dr\") pod \"download-cache-openstack-openstack-networker-nsk64\" (UID: \"73bb544f-2408-4f58-94eb-7b5116d41848\") " pod="openstack/download-cache-openstack-openstack-networker-nsk64" Dec 05 09:19:30 crc kubenswrapper[4997]: I1205 09:19:30.000979 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-nsk64" Dec 05 09:19:30 crc kubenswrapper[4997]: I1205 09:19:30.529497 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-networker-nsk64"] Dec 05 09:19:30 crc kubenswrapper[4997]: W1205 09:19:30.533857 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73bb544f_2408_4f58_94eb_7b5116d41848.slice/crio-18313ce21b4471e17df7d488989583df33658d479a5ae4d7468a7894273c3a3f WatchSource:0}: Error finding container 18313ce21b4471e17df7d488989583df33658d479a5ae4d7468a7894273c3a3f: Status 404 returned error can't find the container with id 18313ce21b4471e17df7d488989583df33658d479a5ae4d7468a7894273c3a3f Dec 05 09:19:30 crc kubenswrapper[4997]: I1205 09:19:30.597913 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-nsk64" event={"ID":"73bb544f-2408-4f58-94eb-7b5116d41848","Type":"ContainerStarted","Data":"18313ce21b4471e17df7d488989583df33658d479a5ae4d7468a7894273c3a3f"} Dec 05 09:19:31 crc kubenswrapper[4997]: I1205 09:19:31.610289 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-nsk64" event={"ID":"73bb544f-2408-4f58-94eb-7b5116d41848","Type":"ContainerStarted","Data":"d55a70cf786a00d5ea6d8f0bb1ff258a4096bd5a49c1f5d7120cf05512763ef8"} Dec 05 09:19:31 crc kubenswrapper[4997]: I1205 09:19:31.636788 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-networker-nsk64" podStartSLOduration=2.236171074 podStartE2EDuration="2.63676586s" podCreationTimestamp="2025-12-05 09:19:29 +0000 UTC" firstStartedPulling="2025-12-05 09:19:30.536192164 +0000 UTC m=+8671.065099425" lastFinishedPulling="2025-12-05 09:19:30.93678694 +0000 UTC m=+8671.465694211" observedRunningTime="2025-12-05 09:19:31.62712918 +0000 UTC m=+8672.156036451" watchObservedRunningTime="2025-12-05 09:19:31.63676586 +0000 UTC m=+8672.165673121" Dec 05 09:20:19 crc kubenswrapper[4997]: I1205 09:20:19.548597 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wmpnm"] Dec 05 09:20:19 crc kubenswrapper[4997]: I1205 09:20:19.551541 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wmpnm" Dec 05 09:20:19 crc kubenswrapper[4997]: I1205 09:20:19.562257 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wmpnm"] Dec 05 09:20:19 crc kubenswrapper[4997]: I1205 09:20:19.696927 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-utilities\") pod \"redhat-marketplace-wmpnm\" (UID: \"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5\") " pod="openshift-marketplace/redhat-marketplace-wmpnm" Dec 05 09:20:19 crc kubenswrapper[4997]: I1205 09:20:19.697098 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-catalog-content\") pod \"redhat-marketplace-wmpnm\" (UID: \"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5\") " pod="openshift-marketplace/redhat-marketplace-wmpnm" Dec 05 09:20:19 crc kubenswrapper[4997]: I1205 09:20:19.697136 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf7tl\" (UniqueName: \"kubernetes.io/projected/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-kube-api-access-jf7tl\") pod \"redhat-marketplace-wmpnm\" (UID: \"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5\") " pod="openshift-marketplace/redhat-marketplace-wmpnm" Dec 05 09:20:19 crc kubenswrapper[4997]: I1205 09:20:19.798983 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-utilities\") pod \"redhat-marketplace-wmpnm\" (UID: \"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5\") " pod="openshift-marketplace/redhat-marketplace-wmpnm" Dec 05 09:20:19 crc kubenswrapper[4997]: I1205 09:20:19.799221 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-catalog-content\") pod \"redhat-marketplace-wmpnm\" (UID: \"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5\") " pod="openshift-marketplace/redhat-marketplace-wmpnm" Dec 05 09:20:19 crc kubenswrapper[4997]: I1205 09:20:19.799277 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf7tl\" (UniqueName: \"kubernetes.io/projected/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-kube-api-access-jf7tl\") pod \"redhat-marketplace-wmpnm\" (UID: \"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5\") " pod="openshift-marketplace/redhat-marketplace-wmpnm" Dec 05 09:20:19 crc kubenswrapper[4997]: I1205 09:20:19.799643 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-utilities\") pod \"redhat-marketplace-wmpnm\" (UID: \"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5\") " pod="openshift-marketplace/redhat-marketplace-wmpnm" Dec 05 09:20:19 crc kubenswrapper[4997]: I1205 09:20:19.799805 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-catalog-content\") pod \"redhat-marketplace-wmpnm\" (UID: \"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5\") " pod="openshift-marketplace/redhat-marketplace-wmpnm" Dec 05 09:20:19 crc kubenswrapper[4997]: I1205 09:20:19.904756 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf7tl\" (UniqueName: \"kubernetes.io/projected/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-kube-api-access-jf7tl\") pod \"redhat-marketplace-wmpnm\" (UID: \"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5\") " pod="openshift-marketplace/redhat-marketplace-wmpnm" Dec 05 09:20:20 crc kubenswrapper[4997]: I1205 09:20:20.180765 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wmpnm" Dec 05 09:20:20 crc kubenswrapper[4997]: I1205 09:20:20.646575 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wmpnm"] Dec 05 09:20:21 crc kubenswrapper[4997]: I1205 09:20:21.087027 4997 generic.go:334] "Generic (PLEG): container finished" podID="ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5" containerID="41e0f431aac72e9b3fe844ce70880758482ef345de5fa95c3c69dbc93722539c" exitCode=0 Dec 05 09:20:21 crc kubenswrapper[4997]: I1205 09:20:21.087128 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmpnm" event={"ID":"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5","Type":"ContainerDied","Data":"41e0f431aac72e9b3fe844ce70880758482ef345de5fa95c3c69dbc93722539c"} Dec 05 09:20:21 crc kubenswrapper[4997]: I1205 09:20:21.087309 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmpnm" event={"ID":"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5","Type":"ContainerStarted","Data":"c7e45457ed35925e83b8d359d8b7ac24ed31365afd994697b92a55ac2ff3b8dd"} Dec 05 09:20:23 crc kubenswrapper[4997]: I1205 09:20:23.110989 4997 generic.go:334] "Generic (PLEG): container finished" podID="ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5" containerID="6b9d674a5c14301883628e64374e1c25ff0f334f1765b5e99cc67ee0aef97174" exitCode=0 Dec 05 09:20:23 crc kubenswrapper[4997]: I1205 09:20:23.111080 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmpnm" event={"ID":"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5","Type":"ContainerDied","Data":"6b9d674a5c14301883628e64374e1c25ff0f334f1765b5e99cc67ee0aef97174"} Dec 05 09:20:24 crc kubenswrapper[4997]: I1205 09:20:24.121914 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmpnm" event={"ID":"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5","Type":"ContainerStarted","Data":"a58b0decd30df22b3c3f0d908f5ad211578a2f7f0a23e5b92b58efe81515ca2a"} Dec 05 09:20:24 crc kubenswrapper[4997]: I1205 09:20:24.145117 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wmpnm" podStartSLOduration=2.743252542 podStartE2EDuration="5.145094665s" podCreationTimestamp="2025-12-05 09:20:19 +0000 UTC" firstStartedPulling="2025-12-05 09:20:21.088733729 +0000 UTC m=+8721.617640990" lastFinishedPulling="2025-12-05 09:20:23.490575862 +0000 UTC m=+8724.019483113" observedRunningTime="2025-12-05 09:20:24.141005784 +0000 UTC m=+8724.669913065" watchObservedRunningTime="2025-12-05 09:20:24.145094665 +0000 UTC m=+8724.674001926" Dec 05 09:20:30 crc kubenswrapper[4997]: I1205 09:20:30.181781 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wmpnm" Dec 05 09:20:30 crc kubenswrapper[4997]: I1205 09:20:30.182381 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wmpnm" Dec 05 09:20:30 crc kubenswrapper[4997]: I1205 09:20:30.253903 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wmpnm" Dec 05 09:20:30 crc kubenswrapper[4997]: I1205 09:20:30.952882 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mrgx4"] Dec 05 09:20:30 crc kubenswrapper[4997]: I1205 09:20:30.955785 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mrgx4" Dec 05 09:20:30 crc kubenswrapper[4997]: I1205 09:20:30.964752 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mrgx4"] Dec 05 09:20:31 crc kubenswrapper[4997]: I1205 09:20:31.135461 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ds8m\" (UniqueName: \"kubernetes.io/projected/6b101195-57bb-44f8-a8a5-90d32686c822-kube-api-access-2ds8m\") pod \"community-operators-mrgx4\" (UID: \"6b101195-57bb-44f8-a8a5-90d32686c822\") " pod="openshift-marketplace/community-operators-mrgx4" Dec 05 09:20:31 crc kubenswrapper[4997]: I1205 09:20:31.135674 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b101195-57bb-44f8-a8a5-90d32686c822-utilities\") pod \"community-operators-mrgx4\" (UID: \"6b101195-57bb-44f8-a8a5-90d32686c822\") " pod="openshift-marketplace/community-operators-mrgx4" Dec 05 09:20:31 crc kubenswrapper[4997]: I1205 09:20:31.135855 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b101195-57bb-44f8-a8a5-90d32686c822-catalog-content\") pod \"community-operators-mrgx4\" (UID: \"6b101195-57bb-44f8-a8a5-90d32686c822\") " pod="openshift-marketplace/community-operators-mrgx4" Dec 05 09:20:31 crc kubenswrapper[4997]: I1205 09:20:31.236789 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wmpnm" Dec 05 09:20:31 crc kubenswrapper[4997]: I1205 09:20:31.237888 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ds8m\" (UniqueName: \"kubernetes.io/projected/6b101195-57bb-44f8-a8a5-90d32686c822-kube-api-access-2ds8m\") pod \"community-operators-mrgx4\" (UID: \"6b101195-57bb-44f8-a8a5-90d32686c822\") " pod="openshift-marketplace/community-operators-mrgx4" Dec 05 09:20:31 crc kubenswrapper[4997]: I1205 09:20:31.237936 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b101195-57bb-44f8-a8a5-90d32686c822-utilities\") pod \"community-operators-mrgx4\" (UID: \"6b101195-57bb-44f8-a8a5-90d32686c822\") " pod="openshift-marketplace/community-operators-mrgx4" Dec 05 09:20:31 crc kubenswrapper[4997]: I1205 09:20:31.237985 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b101195-57bb-44f8-a8a5-90d32686c822-catalog-content\") pod \"community-operators-mrgx4\" (UID: \"6b101195-57bb-44f8-a8a5-90d32686c822\") " pod="openshift-marketplace/community-operators-mrgx4" Dec 05 09:20:31 crc kubenswrapper[4997]: I1205 09:20:31.238447 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b101195-57bb-44f8-a8a5-90d32686c822-catalog-content\") pod \"community-operators-mrgx4\" (UID: \"6b101195-57bb-44f8-a8a5-90d32686c822\") " pod="openshift-marketplace/community-operators-mrgx4" Dec 05 09:20:31 crc kubenswrapper[4997]: I1205 09:20:31.238566 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b101195-57bb-44f8-a8a5-90d32686c822-utilities\") pod \"community-operators-mrgx4\" (UID: \"6b101195-57bb-44f8-a8a5-90d32686c822\") " pod="openshift-marketplace/community-operators-mrgx4" Dec 05 09:20:31 crc kubenswrapper[4997]: I1205 09:20:31.271002 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ds8m\" (UniqueName: \"kubernetes.io/projected/6b101195-57bb-44f8-a8a5-90d32686c822-kube-api-access-2ds8m\") pod \"community-operators-mrgx4\" (UID: \"6b101195-57bb-44f8-a8a5-90d32686c822\") " pod="openshift-marketplace/community-operators-mrgx4" Dec 05 09:20:31 crc kubenswrapper[4997]: I1205 09:20:31.277715 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mrgx4" Dec 05 09:20:31 crc kubenswrapper[4997]: I1205 09:20:31.823878 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mrgx4"] Dec 05 09:20:32 crc kubenswrapper[4997]: I1205 09:20:32.190888 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mrgx4" event={"ID":"6b101195-57bb-44f8-a8a5-90d32686c822","Type":"ContainerStarted","Data":"3437628c65a953eed7bf5b568c21aa98eabdbc6f9379d165b77b30951c846932"} Dec 05 09:20:33 crc kubenswrapper[4997]: I1205 09:20:33.202905 4997 generic.go:334] "Generic (PLEG): container finished" podID="6b101195-57bb-44f8-a8a5-90d32686c822" containerID="929d28a13721caf563dd38f4238ba4b89eec20d221cd34edd457e718374f1bc9" exitCode=0 Dec 05 09:20:33 crc kubenswrapper[4997]: I1205 09:20:33.203007 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mrgx4" event={"ID":"6b101195-57bb-44f8-a8a5-90d32686c822","Type":"ContainerDied","Data":"929d28a13721caf563dd38f4238ba4b89eec20d221cd34edd457e718374f1bc9"} Dec 05 09:20:33 crc kubenswrapper[4997]: I1205 09:20:33.493992 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wmpnm"] Dec 05 09:20:33 crc kubenswrapper[4997]: I1205 09:20:33.494253 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wmpnm" podUID="ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5" containerName="registry-server" containerID="cri-o://a58b0decd30df22b3c3f0d908f5ad211578a2f7f0a23e5b92b58efe81515ca2a" gracePeriod=2 Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.010068 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wmpnm" Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.108746 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jf7tl\" (UniqueName: \"kubernetes.io/projected/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-kube-api-access-jf7tl\") pod \"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5\" (UID: \"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5\") " Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.108965 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-utilities\") pod \"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5\" (UID: \"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5\") " Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.109020 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-catalog-content\") pod \"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5\" (UID: \"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5\") " Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.109760 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-utilities" (OuterVolumeSpecName: "utilities") pod "ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5" (UID: "ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.119787 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-kube-api-access-jf7tl" (OuterVolumeSpecName: "kube-api-access-jf7tl") pod "ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5" (UID: "ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5"). InnerVolumeSpecName "kube-api-access-jf7tl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.129533 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5" (UID: "ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.211499 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.211539 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.211552 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jf7tl\" (UniqueName: \"kubernetes.io/projected/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5-kube-api-access-jf7tl\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.225165 4997 generic.go:334] "Generic (PLEG): container finished" podID="ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5" containerID="a58b0decd30df22b3c3f0d908f5ad211578a2f7f0a23e5b92b58efe81515ca2a" exitCode=0 Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.225245 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmpnm" event={"ID":"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5","Type":"ContainerDied","Data":"a58b0decd30df22b3c3f0d908f5ad211578a2f7f0a23e5b92b58efe81515ca2a"} Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.225250 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wmpnm" Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.225276 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmpnm" event={"ID":"ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5","Type":"ContainerDied","Data":"c7e45457ed35925e83b8d359d8b7ac24ed31365afd994697b92a55ac2ff3b8dd"} Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.225296 4997 scope.go:117] "RemoveContainer" containerID="a58b0decd30df22b3c3f0d908f5ad211578a2f7f0a23e5b92b58efe81515ca2a" Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.236789 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mrgx4" event={"ID":"6b101195-57bb-44f8-a8a5-90d32686c822","Type":"ContainerStarted","Data":"6a2aa418966990ca43f2e43a23de12e87410867fee9c4b26df50e10648652606"} Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.261971 4997 scope.go:117] "RemoveContainer" containerID="6b9d674a5c14301883628e64374e1c25ff0f334f1765b5e99cc67ee0aef97174" Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.283606 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wmpnm"] Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.287769 4997 scope.go:117] "RemoveContainer" containerID="41e0f431aac72e9b3fe844ce70880758482ef345de5fa95c3c69dbc93722539c" Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.292884 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wmpnm"] Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.307229 4997 scope.go:117] "RemoveContainer" containerID="a58b0decd30df22b3c3f0d908f5ad211578a2f7f0a23e5b92b58efe81515ca2a" Dec 05 09:20:34 crc kubenswrapper[4997]: E1205 09:20:34.308039 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a58b0decd30df22b3c3f0d908f5ad211578a2f7f0a23e5b92b58efe81515ca2a\": container with ID starting with a58b0decd30df22b3c3f0d908f5ad211578a2f7f0a23e5b92b58efe81515ca2a not found: ID does not exist" containerID="a58b0decd30df22b3c3f0d908f5ad211578a2f7f0a23e5b92b58efe81515ca2a" Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.308149 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a58b0decd30df22b3c3f0d908f5ad211578a2f7f0a23e5b92b58efe81515ca2a"} err="failed to get container status \"a58b0decd30df22b3c3f0d908f5ad211578a2f7f0a23e5b92b58efe81515ca2a\": rpc error: code = NotFound desc = could not find container \"a58b0decd30df22b3c3f0d908f5ad211578a2f7f0a23e5b92b58efe81515ca2a\": container with ID starting with a58b0decd30df22b3c3f0d908f5ad211578a2f7f0a23e5b92b58efe81515ca2a not found: ID does not exist" Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.308249 4997 scope.go:117] "RemoveContainer" containerID="6b9d674a5c14301883628e64374e1c25ff0f334f1765b5e99cc67ee0aef97174" Dec 05 09:20:34 crc kubenswrapper[4997]: E1205 09:20:34.308641 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b9d674a5c14301883628e64374e1c25ff0f334f1765b5e99cc67ee0aef97174\": container with ID starting with 6b9d674a5c14301883628e64374e1c25ff0f334f1765b5e99cc67ee0aef97174 not found: ID does not exist" containerID="6b9d674a5c14301883628e64374e1c25ff0f334f1765b5e99cc67ee0aef97174" Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.308735 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b9d674a5c14301883628e64374e1c25ff0f334f1765b5e99cc67ee0aef97174"} err="failed to get container status \"6b9d674a5c14301883628e64374e1c25ff0f334f1765b5e99cc67ee0aef97174\": rpc error: code = NotFound desc = could not find container \"6b9d674a5c14301883628e64374e1c25ff0f334f1765b5e99cc67ee0aef97174\": container with ID starting with 6b9d674a5c14301883628e64374e1c25ff0f334f1765b5e99cc67ee0aef97174 not found: ID does not exist" Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.308825 4997 scope.go:117] "RemoveContainer" containerID="41e0f431aac72e9b3fe844ce70880758482ef345de5fa95c3c69dbc93722539c" Dec 05 09:20:34 crc kubenswrapper[4997]: E1205 09:20:34.309224 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41e0f431aac72e9b3fe844ce70880758482ef345de5fa95c3c69dbc93722539c\": container with ID starting with 41e0f431aac72e9b3fe844ce70880758482ef345de5fa95c3c69dbc93722539c not found: ID does not exist" containerID="41e0f431aac72e9b3fe844ce70880758482ef345de5fa95c3c69dbc93722539c" Dec 05 09:20:34 crc kubenswrapper[4997]: I1205 09:20:34.309313 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41e0f431aac72e9b3fe844ce70880758482ef345de5fa95c3c69dbc93722539c"} err="failed to get container status \"41e0f431aac72e9b3fe844ce70880758482ef345de5fa95c3c69dbc93722539c\": rpc error: code = NotFound desc = could not find container \"41e0f431aac72e9b3fe844ce70880758482ef345de5fa95c3c69dbc93722539c\": container with ID starting with 41e0f431aac72e9b3fe844ce70880758482ef345de5fa95c3c69dbc93722539c not found: ID does not exist" Dec 05 09:20:35 crc kubenswrapper[4997]: I1205 09:20:35.250167 4997 generic.go:334] "Generic (PLEG): container finished" podID="6b101195-57bb-44f8-a8a5-90d32686c822" containerID="6a2aa418966990ca43f2e43a23de12e87410867fee9c4b26df50e10648652606" exitCode=0 Dec 05 09:20:35 crc kubenswrapper[4997]: I1205 09:20:35.250281 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mrgx4" event={"ID":"6b101195-57bb-44f8-a8a5-90d32686c822","Type":"ContainerDied","Data":"6a2aa418966990ca43f2e43a23de12e87410867fee9c4b26df50e10648652606"} Dec 05 09:20:35 crc kubenswrapper[4997]: I1205 09:20:35.762888 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5" path="/var/lib/kubelet/pods/ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5/volumes" Dec 05 09:20:36 crc kubenswrapper[4997]: I1205 09:20:36.260970 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mrgx4" event={"ID":"6b101195-57bb-44f8-a8a5-90d32686c822","Type":"ContainerStarted","Data":"f008b3dcd5f7be586f6cd0560c15dc16e53d579a6e5da2ad0b02d5809863465f"} Dec 05 09:20:36 crc kubenswrapper[4997]: I1205 09:20:36.281211 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mrgx4" podStartSLOduration=3.8392182 podStartE2EDuration="6.281189126s" podCreationTimestamp="2025-12-05 09:20:30 +0000 UTC" firstStartedPulling="2025-12-05 09:20:33.206828354 +0000 UTC m=+8733.735735625" lastFinishedPulling="2025-12-05 09:20:35.64879927 +0000 UTC m=+8736.177706551" observedRunningTime="2025-12-05 09:20:36.277377812 +0000 UTC m=+8736.806285093" watchObservedRunningTime="2025-12-05 09:20:36.281189126 +0000 UTC m=+8736.810096387" Dec 05 09:20:41 crc kubenswrapper[4997]: I1205 09:20:41.279122 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mrgx4" Dec 05 09:20:41 crc kubenswrapper[4997]: I1205 09:20:41.279723 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mrgx4" Dec 05 09:20:41 crc kubenswrapper[4997]: I1205 09:20:41.327635 4997 generic.go:334] "Generic (PLEG): container finished" podID="73bb544f-2408-4f58-94eb-7b5116d41848" containerID="d55a70cf786a00d5ea6d8f0bb1ff258a4096bd5a49c1f5d7120cf05512763ef8" exitCode=0 Dec 05 09:20:41 crc kubenswrapper[4997]: I1205 09:20:41.327953 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-nsk64" event={"ID":"73bb544f-2408-4f58-94eb-7b5116d41848","Type":"ContainerDied","Data":"d55a70cf786a00d5ea6d8f0bb1ff258a4096bd5a49c1f5d7120cf05512763ef8"} Dec 05 09:20:41 crc kubenswrapper[4997]: I1205 09:20:41.335194 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mrgx4" Dec 05 09:20:41 crc kubenswrapper[4997]: I1205 09:20:41.381314 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mrgx4" Dec 05 09:20:41 crc kubenswrapper[4997]: I1205 09:20:41.577228 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mrgx4"] Dec 05 09:20:42 crc kubenswrapper[4997]: I1205 09:20:42.897464 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-nsk64" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.022601 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73bb544f-2408-4f58-94eb-7b5116d41848-inventory\") pod \"73bb544f-2408-4f58-94eb-7b5116d41848\" (UID: \"73bb544f-2408-4f58-94eb-7b5116d41848\") " Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.022879 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6c2dr\" (UniqueName: \"kubernetes.io/projected/73bb544f-2408-4f58-94eb-7b5116d41848-kube-api-access-6c2dr\") pod \"73bb544f-2408-4f58-94eb-7b5116d41848\" (UID: \"73bb544f-2408-4f58-94eb-7b5116d41848\") " Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.022933 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73bb544f-2408-4f58-94eb-7b5116d41848-ssh-key\") pod \"73bb544f-2408-4f58-94eb-7b5116d41848\" (UID: \"73bb544f-2408-4f58-94eb-7b5116d41848\") " Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.034791 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73bb544f-2408-4f58-94eb-7b5116d41848-kube-api-access-6c2dr" (OuterVolumeSpecName: "kube-api-access-6c2dr") pod "73bb544f-2408-4f58-94eb-7b5116d41848" (UID: "73bb544f-2408-4f58-94eb-7b5116d41848"). InnerVolumeSpecName "kube-api-access-6c2dr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.050730 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73bb544f-2408-4f58-94eb-7b5116d41848-inventory" (OuterVolumeSpecName: "inventory") pod "73bb544f-2408-4f58-94eb-7b5116d41848" (UID: "73bb544f-2408-4f58-94eb-7b5116d41848"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.074033 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73bb544f-2408-4f58-94eb-7b5116d41848-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "73bb544f-2408-4f58-94eb-7b5116d41848" (UID: "73bb544f-2408-4f58-94eb-7b5116d41848"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.125913 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73bb544f-2408-4f58-94eb-7b5116d41848-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.126050 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73bb544f-2408-4f58-94eb-7b5116d41848-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.126070 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6c2dr\" (UniqueName: \"kubernetes.io/projected/73bb544f-2408-4f58-94eb-7b5116d41848-kube-api-access-6c2dr\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.353392 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-nsk64" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.353391 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-nsk64" event={"ID":"73bb544f-2408-4f58-94eb-7b5116d41848","Type":"ContainerDied","Data":"18313ce21b4471e17df7d488989583df33658d479a5ae4d7468a7894273c3a3f"} Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.353445 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18313ce21b4471e17df7d488989583df33658d479a5ae4d7468a7894273c3a3f" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.353557 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mrgx4" podUID="6b101195-57bb-44f8-a8a5-90d32686c822" containerName="registry-server" containerID="cri-o://f008b3dcd5f7be586f6cd0560c15dc16e53d579a6e5da2ad0b02d5809863465f" gracePeriod=2 Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.456130 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-networker-6pc2p"] Dec 05 09:20:43 crc kubenswrapper[4997]: E1205 09:20:43.456920 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73bb544f-2408-4f58-94eb-7b5116d41848" containerName="download-cache-openstack-openstack-networker" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.456942 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="73bb544f-2408-4f58-94eb-7b5116d41848" containerName="download-cache-openstack-openstack-networker" Dec 05 09:20:43 crc kubenswrapper[4997]: E1205 09:20:43.456966 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5" containerName="registry-server" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.456975 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5" containerName="registry-server" Dec 05 09:20:43 crc kubenswrapper[4997]: E1205 09:20:43.456990 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5" containerName="extract-utilities" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.456998 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5" containerName="extract-utilities" Dec 05 09:20:43 crc kubenswrapper[4997]: E1205 09:20:43.457014 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5" containerName="extract-content" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.457022 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5" containerName="extract-content" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.457262 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac1b8572-81c0-4747-9d7d-0a1ffdc0aff5" containerName="registry-server" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.457280 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="73bb544f-2408-4f58-94eb-7b5116d41848" containerName="download-cache-openstack-openstack-networker" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.458190 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-6pc2p" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.461456 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-4542r" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.462599 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.465902 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-networker-6pc2p"] Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.636586 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zgwq\" (UniqueName: \"kubernetes.io/projected/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-kube-api-access-5zgwq\") pod \"configure-network-openstack-openstack-networker-6pc2p\" (UID: \"10cd3056-9662-42dd-b5c4-9b4aeed3da3a\") " pod="openstack/configure-network-openstack-openstack-networker-6pc2p" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.637362 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-inventory\") pod \"configure-network-openstack-openstack-networker-6pc2p\" (UID: \"10cd3056-9662-42dd-b5c4-9b4aeed3da3a\") " pod="openstack/configure-network-openstack-openstack-networker-6pc2p" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.637973 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-ssh-key\") pod \"configure-network-openstack-openstack-networker-6pc2p\" (UID: \"10cd3056-9662-42dd-b5c4-9b4aeed3da3a\") " pod="openstack/configure-network-openstack-openstack-networker-6pc2p" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.740146 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-ssh-key\") pod \"configure-network-openstack-openstack-networker-6pc2p\" (UID: \"10cd3056-9662-42dd-b5c4-9b4aeed3da3a\") " pod="openstack/configure-network-openstack-openstack-networker-6pc2p" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.740240 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zgwq\" (UniqueName: \"kubernetes.io/projected/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-kube-api-access-5zgwq\") pod \"configure-network-openstack-openstack-networker-6pc2p\" (UID: \"10cd3056-9662-42dd-b5c4-9b4aeed3da3a\") " pod="openstack/configure-network-openstack-openstack-networker-6pc2p" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.740367 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-inventory\") pod \"configure-network-openstack-openstack-networker-6pc2p\" (UID: \"10cd3056-9662-42dd-b5c4-9b4aeed3da3a\") " pod="openstack/configure-network-openstack-openstack-networker-6pc2p" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.745823 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-inventory\") pod \"configure-network-openstack-openstack-networker-6pc2p\" (UID: \"10cd3056-9662-42dd-b5c4-9b4aeed3da3a\") " pod="openstack/configure-network-openstack-openstack-networker-6pc2p" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.747670 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-ssh-key\") pod \"configure-network-openstack-openstack-networker-6pc2p\" (UID: \"10cd3056-9662-42dd-b5c4-9b4aeed3da3a\") " pod="openstack/configure-network-openstack-openstack-networker-6pc2p" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.766396 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zgwq\" (UniqueName: \"kubernetes.io/projected/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-kube-api-access-5zgwq\") pod \"configure-network-openstack-openstack-networker-6pc2p\" (UID: \"10cd3056-9662-42dd-b5c4-9b4aeed3da3a\") " pod="openstack/configure-network-openstack-openstack-networker-6pc2p" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.819629 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-6pc2p" Dec 05 09:20:43 crc kubenswrapper[4997]: I1205 09:20:43.917377 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mrgx4" Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.046293 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ds8m\" (UniqueName: \"kubernetes.io/projected/6b101195-57bb-44f8-a8a5-90d32686c822-kube-api-access-2ds8m\") pod \"6b101195-57bb-44f8-a8a5-90d32686c822\" (UID: \"6b101195-57bb-44f8-a8a5-90d32686c822\") " Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.046453 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b101195-57bb-44f8-a8a5-90d32686c822-utilities\") pod \"6b101195-57bb-44f8-a8a5-90d32686c822\" (UID: \"6b101195-57bb-44f8-a8a5-90d32686c822\") " Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.046690 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b101195-57bb-44f8-a8a5-90d32686c822-catalog-content\") pod \"6b101195-57bb-44f8-a8a5-90d32686c822\" (UID: \"6b101195-57bb-44f8-a8a5-90d32686c822\") " Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.048707 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b101195-57bb-44f8-a8a5-90d32686c822-utilities" (OuterVolumeSpecName: "utilities") pod "6b101195-57bb-44f8-a8a5-90d32686c822" (UID: "6b101195-57bb-44f8-a8a5-90d32686c822"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.060133 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b101195-57bb-44f8-a8a5-90d32686c822-kube-api-access-2ds8m" (OuterVolumeSpecName: "kube-api-access-2ds8m") pod "6b101195-57bb-44f8-a8a5-90d32686c822" (UID: "6b101195-57bb-44f8-a8a5-90d32686c822"). InnerVolumeSpecName "kube-api-access-2ds8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.115629 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b101195-57bb-44f8-a8a5-90d32686c822-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b101195-57bb-44f8-a8a5-90d32686c822" (UID: "6b101195-57bb-44f8-a8a5-90d32686c822"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.149150 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b101195-57bb-44f8-a8a5-90d32686c822-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.149197 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ds8m\" (UniqueName: \"kubernetes.io/projected/6b101195-57bb-44f8-a8a5-90d32686c822-kube-api-access-2ds8m\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.149210 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b101195-57bb-44f8-a8a5-90d32686c822-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.363586 4997 generic.go:334] "Generic (PLEG): container finished" podID="6b101195-57bb-44f8-a8a5-90d32686c822" containerID="f008b3dcd5f7be586f6cd0560c15dc16e53d579a6e5da2ad0b02d5809863465f" exitCode=0 Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.363653 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mrgx4" event={"ID":"6b101195-57bb-44f8-a8a5-90d32686c822","Type":"ContainerDied","Data":"f008b3dcd5f7be586f6cd0560c15dc16e53d579a6e5da2ad0b02d5809863465f"} Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.363718 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mrgx4" event={"ID":"6b101195-57bb-44f8-a8a5-90d32686c822","Type":"ContainerDied","Data":"3437628c65a953eed7bf5b568c21aa98eabdbc6f9379d165b77b30951c846932"} Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.363739 4997 scope.go:117] "RemoveContainer" containerID="f008b3dcd5f7be586f6cd0560c15dc16e53d579a6e5da2ad0b02d5809863465f" Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.363675 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mrgx4" Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.396887 4997 scope.go:117] "RemoveContainer" containerID="6a2aa418966990ca43f2e43a23de12e87410867fee9c4b26df50e10648652606" Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.406057 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-networker-6pc2p"] Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.425376 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mrgx4"] Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.435831 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mrgx4"] Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.460599 4997 scope.go:117] "RemoveContainer" containerID="929d28a13721caf563dd38f4238ba4b89eec20d221cd34edd457e718374f1bc9" Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.480474 4997 scope.go:117] "RemoveContainer" containerID="f008b3dcd5f7be586f6cd0560c15dc16e53d579a6e5da2ad0b02d5809863465f" Dec 05 09:20:44 crc kubenswrapper[4997]: E1205 09:20:44.480988 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f008b3dcd5f7be586f6cd0560c15dc16e53d579a6e5da2ad0b02d5809863465f\": container with ID starting with f008b3dcd5f7be586f6cd0560c15dc16e53d579a6e5da2ad0b02d5809863465f not found: ID does not exist" containerID="f008b3dcd5f7be586f6cd0560c15dc16e53d579a6e5da2ad0b02d5809863465f" Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.481038 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f008b3dcd5f7be586f6cd0560c15dc16e53d579a6e5da2ad0b02d5809863465f"} err="failed to get container status \"f008b3dcd5f7be586f6cd0560c15dc16e53d579a6e5da2ad0b02d5809863465f\": rpc error: code = NotFound desc = could not find container \"f008b3dcd5f7be586f6cd0560c15dc16e53d579a6e5da2ad0b02d5809863465f\": container with ID starting with f008b3dcd5f7be586f6cd0560c15dc16e53d579a6e5da2ad0b02d5809863465f not found: ID does not exist" Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.481062 4997 scope.go:117] "RemoveContainer" containerID="6a2aa418966990ca43f2e43a23de12e87410867fee9c4b26df50e10648652606" Dec 05 09:20:44 crc kubenswrapper[4997]: E1205 09:20:44.481419 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a2aa418966990ca43f2e43a23de12e87410867fee9c4b26df50e10648652606\": container with ID starting with 6a2aa418966990ca43f2e43a23de12e87410867fee9c4b26df50e10648652606 not found: ID does not exist" containerID="6a2aa418966990ca43f2e43a23de12e87410867fee9c4b26df50e10648652606" Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.481467 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a2aa418966990ca43f2e43a23de12e87410867fee9c4b26df50e10648652606"} err="failed to get container status \"6a2aa418966990ca43f2e43a23de12e87410867fee9c4b26df50e10648652606\": rpc error: code = NotFound desc = could not find container \"6a2aa418966990ca43f2e43a23de12e87410867fee9c4b26df50e10648652606\": container with ID starting with 6a2aa418966990ca43f2e43a23de12e87410867fee9c4b26df50e10648652606 not found: ID does not exist" Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.481498 4997 scope.go:117] "RemoveContainer" containerID="929d28a13721caf563dd38f4238ba4b89eec20d221cd34edd457e718374f1bc9" Dec 05 09:20:44 crc kubenswrapper[4997]: E1205 09:20:44.481918 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"929d28a13721caf563dd38f4238ba4b89eec20d221cd34edd457e718374f1bc9\": container with ID starting with 929d28a13721caf563dd38f4238ba4b89eec20d221cd34edd457e718374f1bc9 not found: ID does not exist" containerID="929d28a13721caf563dd38f4238ba4b89eec20d221cd34edd457e718374f1bc9" Dec 05 09:20:44 crc kubenswrapper[4997]: I1205 09:20:44.482125 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"929d28a13721caf563dd38f4238ba4b89eec20d221cd34edd457e718374f1bc9"} err="failed to get container status \"929d28a13721caf563dd38f4238ba4b89eec20d221cd34edd457e718374f1bc9\": rpc error: code = NotFound desc = could not find container \"929d28a13721caf563dd38f4238ba4b89eec20d221cd34edd457e718374f1bc9\": container with ID starting with 929d28a13721caf563dd38f4238ba4b89eec20d221cd34edd457e718374f1bc9 not found: ID does not exist" Dec 05 09:20:45 crc kubenswrapper[4997]: I1205 09:20:45.375894 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-6pc2p" event={"ID":"10cd3056-9662-42dd-b5c4-9b4aeed3da3a","Type":"ContainerStarted","Data":"cfb60e6a31492b07b136e1721634918e1f62cf0a3f89d527ed0355d51d1bf226"} Dec 05 09:20:45 crc kubenswrapper[4997]: I1205 09:20:45.375935 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-6pc2p" event={"ID":"10cd3056-9662-42dd-b5c4-9b4aeed3da3a","Type":"ContainerStarted","Data":"1de64ed1c675cf3d1d050e7eda4426e7210e5eee9e6f220e074578c19dc4efaa"} Dec 05 09:20:45 crc kubenswrapper[4997]: I1205 09:20:45.395339 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-networker-6pc2p" podStartSLOduration=2.009666277 podStartE2EDuration="2.395319549s" podCreationTimestamp="2025-12-05 09:20:43 +0000 UTC" firstStartedPulling="2025-12-05 09:20:44.40907905 +0000 UTC m=+8744.937986311" lastFinishedPulling="2025-12-05 09:20:44.794732322 +0000 UTC m=+8745.323639583" observedRunningTime="2025-12-05 09:20:45.392253187 +0000 UTC m=+8745.921160448" watchObservedRunningTime="2025-12-05 09:20:45.395319549 +0000 UTC m=+8745.924226810" Dec 05 09:20:45 crc kubenswrapper[4997]: I1205 09:20:45.760513 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b101195-57bb-44f8-a8a5-90d32686c822" path="/var/lib/kubelet/pods/6b101195-57bb-44f8-a8a5-90d32686c822/volumes" Dec 05 09:20:57 crc kubenswrapper[4997]: I1205 09:20:57.483943 4997 generic.go:334] "Generic (PLEG): container finished" podID="999c70de-3370-4ce0-947a-b585dcd879d7" containerID="d5f519bfe710282a2d845da1bee779fa7aba49a5c0a165e9d8dafcf447703c4b" exitCode=0 Dec 05 09:20:57 crc kubenswrapper[4997]: I1205 09:20:57.484028 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-b22xv" event={"ID":"999c70de-3370-4ce0-947a-b585dcd879d7","Type":"ContainerDied","Data":"d5f519bfe710282a2d845da1bee779fa7aba49a5c0a165e9d8dafcf447703c4b"} Dec 05 09:20:58 crc kubenswrapper[4997]: I1205 09:20:58.986961 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-b22xv" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.063790 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndhrr\" (UniqueName: \"kubernetes.io/projected/999c70de-3370-4ce0-947a-b585dcd879d7-kube-api-access-ndhrr\") pod \"999c70de-3370-4ce0-947a-b585dcd879d7\" (UID: \"999c70de-3370-4ce0-947a-b585dcd879d7\") " Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.063831 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-inventory\") pod \"999c70de-3370-4ce0-947a-b585dcd879d7\" (UID: \"999c70de-3370-4ce0-947a-b585dcd879d7\") " Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.063908 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-ceph\") pod \"999c70de-3370-4ce0-947a-b585dcd879d7\" (UID: \"999c70de-3370-4ce0-947a-b585dcd879d7\") " Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.063984 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-ssh-key\") pod \"999c70de-3370-4ce0-947a-b585dcd879d7\" (UID: \"999c70de-3370-4ce0-947a-b585dcd879d7\") " Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.076043 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-ceph" (OuterVolumeSpecName: "ceph") pod "999c70de-3370-4ce0-947a-b585dcd879d7" (UID: "999c70de-3370-4ce0-947a-b585dcd879d7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.076195 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/999c70de-3370-4ce0-947a-b585dcd879d7-kube-api-access-ndhrr" (OuterVolumeSpecName: "kube-api-access-ndhrr") pod "999c70de-3370-4ce0-947a-b585dcd879d7" (UID: "999c70de-3370-4ce0-947a-b585dcd879d7"). InnerVolumeSpecName "kube-api-access-ndhrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.092174 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-inventory" (OuterVolumeSpecName: "inventory") pod "999c70de-3370-4ce0-947a-b585dcd879d7" (UID: "999c70de-3370-4ce0-947a-b585dcd879d7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.096446 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "999c70de-3370-4ce0-947a-b585dcd879d7" (UID: "999c70de-3370-4ce0-947a-b585dcd879d7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.166864 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.167056 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndhrr\" (UniqueName: \"kubernetes.io/projected/999c70de-3370-4ce0-947a-b585dcd879d7-kube-api-access-ndhrr\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.167123 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.167213 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/999c70de-3370-4ce0-947a-b585dcd879d7-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.501743 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-b22xv" event={"ID":"999c70de-3370-4ce0-947a-b585dcd879d7","Type":"ContainerDied","Data":"9f19eae73e9d40ad5d766f6c0d36224400cf2b9900c5caca5cb442fb3937bb40"} Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.502016 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f19eae73e9d40ad5d766f6c0d36224400cf2b9900c5caca5cb442fb3937bb40" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.501811 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-b22xv" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.660401 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-rlp6r"] Dec 05 09:20:59 crc kubenswrapper[4997]: E1205 09:20:59.661085 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b101195-57bb-44f8-a8a5-90d32686c822" containerName="extract-content" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.661109 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b101195-57bb-44f8-a8a5-90d32686c822" containerName="extract-content" Dec 05 09:20:59 crc kubenswrapper[4997]: E1205 09:20:59.661126 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b101195-57bb-44f8-a8a5-90d32686c822" containerName="extract-utilities" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.661135 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b101195-57bb-44f8-a8a5-90d32686c822" containerName="extract-utilities" Dec 05 09:20:59 crc kubenswrapper[4997]: E1205 09:20:59.661156 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="999c70de-3370-4ce0-947a-b585dcd879d7" containerName="download-cache-openstack-openstack-cell1" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.661165 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="999c70de-3370-4ce0-947a-b585dcd879d7" containerName="download-cache-openstack-openstack-cell1" Dec 05 09:20:59 crc kubenswrapper[4997]: E1205 09:20:59.661210 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b101195-57bb-44f8-a8a5-90d32686c822" containerName="registry-server" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.661220 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b101195-57bb-44f8-a8a5-90d32686c822" containerName="registry-server" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.661493 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="999c70de-3370-4ce0-947a-b585dcd879d7" containerName="download-cache-openstack-openstack-cell1" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.661523 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b101195-57bb-44f8-a8a5-90d32686c822" containerName="registry-server" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.662417 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.665844 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.666107 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.678404 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-rlp6r"] Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.780443 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-inventory\") pod \"configure-network-openstack-openstack-cell1-rlp6r\" (UID: \"efdefcdd-f76f-4767-af25-b730587e394b\") " pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.780724 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bd29\" (UniqueName: \"kubernetes.io/projected/efdefcdd-f76f-4767-af25-b730587e394b-kube-api-access-8bd29\") pod \"configure-network-openstack-openstack-cell1-rlp6r\" (UID: \"efdefcdd-f76f-4767-af25-b730587e394b\") " pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.780848 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-ceph\") pod \"configure-network-openstack-openstack-cell1-rlp6r\" (UID: \"efdefcdd-f76f-4767-af25-b730587e394b\") " pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.780915 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-ssh-key\") pod \"configure-network-openstack-openstack-cell1-rlp6r\" (UID: \"efdefcdd-f76f-4767-af25-b730587e394b\") " pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.882708 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-inventory\") pod \"configure-network-openstack-openstack-cell1-rlp6r\" (UID: \"efdefcdd-f76f-4767-af25-b730587e394b\") " pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.882827 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bd29\" (UniqueName: \"kubernetes.io/projected/efdefcdd-f76f-4767-af25-b730587e394b-kube-api-access-8bd29\") pod \"configure-network-openstack-openstack-cell1-rlp6r\" (UID: \"efdefcdd-f76f-4767-af25-b730587e394b\") " pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.882854 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-ceph\") pod \"configure-network-openstack-openstack-cell1-rlp6r\" (UID: \"efdefcdd-f76f-4767-af25-b730587e394b\") " pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.882894 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-ssh-key\") pod \"configure-network-openstack-openstack-cell1-rlp6r\" (UID: \"efdefcdd-f76f-4767-af25-b730587e394b\") " pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.886969 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.889626 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-ceph\") pod \"configure-network-openstack-openstack-cell1-rlp6r\" (UID: \"efdefcdd-f76f-4767-af25-b730587e394b\") " pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.891109 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-ssh-key\") pod \"configure-network-openstack-openstack-cell1-rlp6r\" (UID: \"efdefcdd-f76f-4767-af25-b730587e394b\") " pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.899255 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-inventory\") pod \"configure-network-openstack-openstack-cell1-rlp6r\" (UID: \"efdefcdd-f76f-4767-af25-b730587e394b\") " pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.901699 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bd29\" (UniqueName: \"kubernetes.io/projected/efdefcdd-f76f-4767-af25-b730587e394b-kube-api-access-8bd29\") pod \"configure-network-openstack-openstack-cell1-rlp6r\" (UID: \"efdefcdd-f76f-4767-af25-b730587e394b\") " pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.987068 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:20:59 crc kubenswrapper[4997]: I1205 09:20:59.995227 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" Dec 05 09:21:01 crc kubenswrapper[4997]: W1205 09:21:01.184802 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefdefcdd_f76f_4767_af25_b730587e394b.slice/crio-2d2566c63faa3a915fc206fe28fc2ddebb2a3ef5e0857383f139e34c5d4d1172 WatchSource:0}: Error finding container 2d2566c63faa3a915fc206fe28fc2ddebb2a3ef5e0857383f139e34c5d4d1172: Status 404 returned error can't find the container with id 2d2566c63faa3a915fc206fe28fc2ddebb2a3ef5e0857383f139e34c5d4d1172 Dec 05 09:21:01 crc kubenswrapper[4997]: I1205 09:21:01.190077 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-rlp6r"] Dec 05 09:21:01 crc kubenswrapper[4997]: I1205 09:21:01.520867 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" event={"ID":"efdefcdd-f76f-4767-af25-b730587e394b","Type":"ContainerStarted","Data":"2d2566c63faa3a915fc206fe28fc2ddebb2a3ef5e0857383f139e34c5d4d1172"} Dec 05 09:21:02 crc kubenswrapper[4997]: I1205 09:21:02.533774 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" event={"ID":"efdefcdd-f76f-4767-af25-b730587e394b","Type":"ContainerStarted","Data":"f401824e7c4dd78993db6aaad587a8c387a21cefe692dba0f5c47b5750b32090"} Dec 05 09:21:02 crc kubenswrapper[4997]: I1205 09:21:02.551795 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" podStartSLOduration=3.109197946 podStartE2EDuration="3.551771876s" podCreationTimestamp="2025-12-05 09:20:59 +0000 UTC" firstStartedPulling="2025-12-05 09:21:01.187727525 +0000 UTC m=+8761.716634786" lastFinishedPulling="2025-12-05 09:21:01.630301415 +0000 UTC m=+8762.159208716" observedRunningTime="2025-12-05 09:21:02.547311265 +0000 UTC m=+8763.076218556" watchObservedRunningTime="2025-12-05 09:21:02.551771876 +0000 UTC m=+8763.080679147" Dec 05 09:21:49 crc kubenswrapper[4997]: I1205 09:21:49.769951 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:21:49 crc kubenswrapper[4997]: I1205 09:21:49.770525 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:21:50 crc kubenswrapper[4997]: I1205 09:21:50.043850 4997 generic.go:334] "Generic (PLEG): container finished" podID="10cd3056-9662-42dd-b5c4-9b4aeed3da3a" containerID="cfb60e6a31492b07b136e1721634918e1f62cf0a3f89d527ed0355d51d1bf226" exitCode=0 Dec 05 09:21:50 crc kubenswrapper[4997]: I1205 09:21:50.043917 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-6pc2p" event={"ID":"10cd3056-9662-42dd-b5c4-9b4aeed3da3a","Type":"ContainerDied","Data":"cfb60e6a31492b07b136e1721634918e1f62cf0a3f89d527ed0355d51d1bf226"} Dec 05 09:21:51 crc kubenswrapper[4997]: I1205 09:21:51.517418 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-6pc2p" Dec 05 09:21:51 crc kubenswrapper[4997]: I1205 09:21:51.637530 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-ssh-key\") pod \"10cd3056-9662-42dd-b5c4-9b4aeed3da3a\" (UID: \"10cd3056-9662-42dd-b5c4-9b4aeed3da3a\") " Dec 05 09:21:51 crc kubenswrapper[4997]: I1205 09:21:51.637739 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-inventory\") pod \"10cd3056-9662-42dd-b5c4-9b4aeed3da3a\" (UID: \"10cd3056-9662-42dd-b5c4-9b4aeed3da3a\") " Dec 05 09:21:51 crc kubenswrapper[4997]: I1205 09:21:51.637905 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zgwq\" (UniqueName: \"kubernetes.io/projected/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-kube-api-access-5zgwq\") pod \"10cd3056-9662-42dd-b5c4-9b4aeed3da3a\" (UID: \"10cd3056-9662-42dd-b5c4-9b4aeed3da3a\") " Dec 05 09:21:51 crc kubenswrapper[4997]: I1205 09:21:51.644128 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-kube-api-access-5zgwq" (OuterVolumeSpecName: "kube-api-access-5zgwq") pod "10cd3056-9662-42dd-b5c4-9b4aeed3da3a" (UID: "10cd3056-9662-42dd-b5c4-9b4aeed3da3a"). InnerVolumeSpecName "kube-api-access-5zgwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:21:51 crc kubenswrapper[4997]: I1205 09:21:51.665020 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "10cd3056-9662-42dd-b5c4-9b4aeed3da3a" (UID: "10cd3056-9662-42dd-b5c4-9b4aeed3da3a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:21:51 crc kubenswrapper[4997]: I1205 09:21:51.670158 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-inventory" (OuterVolumeSpecName: "inventory") pod "10cd3056-9662-42dd-b5c4-9b4aeed3da3a" (UID: "10cd3056-9662-42dd-b5c4-9b4aeed3da3a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:21:51 crc kubenswrapper[4997]: I1205 09:21:51.740983 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zgwq\" (UniqueName: \"kubernetes.io/projected/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-kube-api-access-5zgwq\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:51 crc kubenswrapper[4997]: I1205 09:21:51.741021 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:51 crc kubenswrapper[4997]: I1205 09:21:51.741032 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10cd3056-9662-42dd-b5c4-9b4aeed3da3a-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.066064 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-6pc2p" event={"ID":"10cd3056-9662-42dd-b5c4-9b4aeed3da3a","Type":"ContainerDied","Data":"1de64ed1c675cf3d1d050e7eda4426e7210e5eee9e6f220e074578c19dc4efaa"} Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.066114 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1de64ed1c675cf3d1d050e7eda4426e7210e5eee9e6f220e074578c19dc4efaa" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.066115 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-6pc2p" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.162215 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-networker-s9llr"] Dec 05 09:21:52 crc kubenswrapper[4997]: E1205 09:21:52.162762 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10cd3056-9662-42dd-b5c4-9b4aeed3da3a" containerName="configure-network-openstack-openstack-networker" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.162787 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="10cd3056-9662-42dd-b5c4-9b4aeed3da3a" containerName="configure-network-openstack-openstack-networker" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.163095 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="10cd3056-9662-42dd-b5c4-9b4aeed3da3a" containerName="configure-network-openstack-openstack-networker" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.164222 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-s9llr" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.167235 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-4542r" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.167677 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.187556 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-networker-s9llr"] Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.250950 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2qtf\" (UniqueName: \"kubernetes.io/projected/10ee156b-3801-425b-b9cc-33c89f2912cd-kube-api-access-q2qtf\") pod \"validate-network-openstack-openstack-networker-s9llr\" (UID: \"10ee156b-3801-425b-b9cc-33c89f2912cd\") " pod="openstack/validate-network-openstack-openstack-networker-s9llr" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.251176 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10ee156b-3801-425b-b9cc-33c89f2912cd-ssh-key\") pod \"validate-network-openstack-openstack-networker-s9llr\" (UID: \"10ee156b-3801-425b-b9cc-33c89f2912cd\") " pod="openstack/validate-network-openstack-openstack-networker-s9llr" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.251230 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10ee156b-3801-425b-b9cc-33c89f2912cd-inventory\") pod \"validate-network-openstack-openstack-networker-s9llr\" (UID: \"10ee156b-3801-425b-b9cc-33c89f2912cd\") " pod="openstack/validate-network-openstack-openstack-networker-s9llr" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.353050 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10ee156b-3801-425b-b9cc-33c89f2912cd-ssh-key\") pod \"validate-network-openstack-openstack-networker-s9llr\" (UID: \"10ee156b-3801-425b-b9cc-33c89f2912cd\") " pod="openstack/validate-network-openstack-openstack-networker-s9llr" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.353112 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10ee156b-3801-425b-b9cc-33c89f2912cd-inventory\") pod \"validate-network-openstack-openstack-networker-s9llr\" (UID: \"10ee156b-3801-425b-b9cc-33c89f2912cd\") " pod="openstack/validate-network-openstack-openstack-networker-s9llr" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.353224 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2qtf\" (UniqueName: \"kubernetes.io/projected/10ee156b-3801-425b-b9cc-33c89f2912cd-kube-api-access-q2qtf\") pod \"validate-network-openstack-openstack-networker-s9llr\" (UID: \"10ee156b-3801-425b-b9cc-33c89f2912cd\") " pod="openstack/validate-network-openstack-openstack-networker-s9llr" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.357705 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10ee156b-3801-425b-b9cc-33c89f2912cd-ssh-key\") pod \"validate-network-openstack-openstack-networker-s9llr\" (UID: \"10ee156b-3801-425b-b9cc-33c89f2912cd\") " pod="openstack/validate-network-openstack-openstack-networker-s9llr" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.368260 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10ee156b-3801-425b-b9cc-33c89f2912cd-inventory\") pod \"validate-network-openstack-openstack-networker-s9llr\" (UID: \"10ee156b-3801-425b-b9cc-33c89f2912cd\") " pod="openstack/validate-network-openstack-openstack-networker-s9llr" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.372485 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2qtf\" (UniqueName: \"kubernetes.io/projected/10ee156b-3801-425b-b9cc-33c89f2912cd-kube-api-access-q2qtf\") pod \"validate-network-openstack-openstack-networker-s9llr\" (UID: \"10ee156b-3801-425b-b9cc-33c89f2912cd\") " pod="openstack/validate-network-openstack-openstack-networker-s9llr" Dec 05 09:21:52 crc kubenswrapper[4997]: I1205 09:21:52.481733 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-s9llr" Dec 05 09:21:53 crc kubenswrapper[4997]: I1205 09:21:53.024847 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-networker-s9llr"] Dec 05 09:21:53 crc kubenswrapper[4997]: I1205 09:21:53.040042 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:21:53 crc kubenswrapper[4997]: I1205 09:21:53.076689 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-s9llr" event={"ID":"10ee156b-3801-425b-b9cc-33c89f2912cd","Type":"ContainerStarted","Data":"6dc83c35e31852964096c199bd5e9f9be3d624ac446fc3f0e70cb267c656061e"} Dec 05 09:21:54 crc kubenswrapper[4997]: I1205 09:21:54.086342 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-s9llr" event={"ID":"10ee156b-3801-425b-b9cc-33c89f2912cd","Type":"ContainerStarted","Data":"79657fb5cd778342dfb2b6e93cc7b462eea51e9293d526fc2f1d670a81711f99"} Dec 05 09:21:54 crc kubenswrapper[4997]: I1205 09:21:54.100628 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-networker-s9llr" podStartSLOduration=1.470005438 podStartE2EDuration="2.100588714s" podCreationTimestamp="2025-12-05 09:21:52 +0000 UTC" firstStartedPulling="2025-12-05 09:21:53.039775071 +0000 UTC m=+8813.568682332" lastFinishedPulling="2025-12-05 09:21:53.670358347 +0000 UTC m=+8814.199265608" observedRunningTime="2025-12-05 09:21:54.098799315 +0000 UTC m=+8814.627706586" watchObservedRunningTime="2025-12-05 09:21:54.100588714 +0000 UTC m=+8814.629495975" Dec 05 09:22:00 crc kubenswrapper[4997]: I1205 09:22:00.149059 4997 generic.go:334] "Generic (PLEG): container finished" podID="10ee156b-3801-425b-b9cc-33c89f2912cd" containerID="79657fb5cd778342dfb2b6e93cc7b462eea51e9293d526fc2f1d670a81711f99" exitCode=0 Dec 05 09:22:00 crc kubenswrapper[4997]: I1205 09:22:00.149147 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-s9llr" event={"ID":"10ee156b-3801-425b-b9cc-33c89f2912cd","Type":"ContainerDied","Data":"79657fb5cd778342dfb2b6e93cc7b462eea51e9293d526fc2f1d670a81711f99"} Dec 05 09:22:01 crc kubenswrapper[4997]: I1205 09:22:01.722266 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-s9llr" Dec 05 09:22:01 crc kubenswrapper[4997]: I1205 09:22:01.862753 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10ee156b-3801-425b-b9cc-33c89f2912cd-ssh-key\") pod \"10ee156b-3801-425b-b9cc-33c89f2912cd\" (UID: \"10ee156b-3801-425b-b9cc-33c89f2912cd\") " Dec 05 09:22:01 crc kubenswrapper[4997]: I1205 09:22:01.862844 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2qtf\" (UniqueName: \"kubernetes.io/projected/10ee156b-3801-425b-b9cc-33c89f2912cd-kube-api-access-q2qtf\") pod \"10ee156b-3801-425b-b9cc-33c89f2912cd\" (UID: \"10ee156b-3801-425b-b9cc-33c89f2912cd\") " Dec 05 09:22:01 crc kubenswrapper[4997]: I1205 09:22:01.862885 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10ee156b-3801-425b-b9cc-33c89f2912cd-inventory\") pod \"10ee156b-3801-425b-b9cc-33c89f2912cd\" (UID: \"10ee156b-3801-425b-b9cc-33c89f2912cd\") " Dec 05 09:22:01 crc kubenswrapper[4997]: I1205 09:22:01.868228 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10ee156b-3801-425b-b9cc-33c89f2912cd-kube-api-access-q2qtf" (OuterVolumeSpecName: "kube-api-access-q2qtf") pod "10ee156b-3801-425b-b9cc-33c89f2912cd" (UID: "10ee156b-3801-425b-b9cc-33c89f2912cd"). InnerVolumeSpecName "kube-api-access-q2qtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:22:01 crc kubenswrapper[4997]: I1205 09:22:01.900841 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10ee156b-3801-425b-b9cc-33c89f2912cd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "10ee156b-3801-425b-b9cc-33c89f2912cd" (UID: "10ee156b-3801-425b-b9cc-33c89f2912cd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:22:01 crc kubenswrapper[4997]: I1205 09:22:01.900871 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10ee156b-3801-425b-b9cc-33c89f2912cd-inventory" (OuterVolumeSpecName: "inventory") pod "10ee156b-3801-425b-b9cc-33c89f2912cd" (UID: "10ee156b-3801-425b-b9cc-33c89f2912cd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:22:01 crc kubenswrapper[4997]: I1205 09:22:01.965699 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10ee156b-3801-425b-b9cc-33c89f2912cd-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:01 crc kubenswrapper[4997]: I1205 09:22:01.965768 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2qtf\" (UniqueName: \"kubernetes.io/projected/10ee156b-3801-425b-b9cc-33c89f2912cd-kube-api-access-q2qtf\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:01 crc kubenswrapper[4997]: I1205 09:22:01.965788 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10ee156b-3801-425b-b9cc-33c89f2912cd-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.180887 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-s9llr" event={"ID":"10ee156b-3801-425b-b9cc-33c89f2912cd","Type":"ContainerDied","Data":"6dc83c35e31852964096c199bd5e9f9be3d624ac446fc3f0e70cb267c656061e"} Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.181226 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6dc83c35e31852964096c199bd5e9f9be3d624ac446fc3f0e70cb267c656061e" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.181132 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-s9llr" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.182203 4997 generic.go:334] "Generic (PLEG): container finished" podID="efdefcdd-f76f-4767-af25-b730587e394b" containerID="f401824e7c4dd78993db6aaad587a8c387a21cefe692dba0f5c47b5750b32090" exitCode=0 Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.182228 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" event={"ID":"efdefcdd-f76f-4767-af25-b730587e394b","Type":"ContainerDied","Data":"f401824e7c4dd78993db6aaad587a8c387a21cefe692dba0f5c47b5750b32090"} Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.243691 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-networker-xmfpv"] Dec 05 09:22:02 crc kubenswrapper[4997]: E1205 09:22:02.244591 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10ee156b-3801-425b-b9cc-33c89f2912cd" containerName="validate-network-openstack-openstack-networker" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.244645 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="10ee156b-3801-425b-b9cc-33c89f2912cd" containerName="validate-network-openstack-openstack-networker" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.245109 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="10ee156b-3801-425b-b9cc-33c89f2912cd" containerName="validate-network-openstack-openstack-networker" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.246229 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-xmfpv" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.252071 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-4542r" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.252648 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.269697 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-networker-xmfpv"] Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.375300 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d81a1b59-323c-4254-8e95-b50acb2eb309-ssh-key\") pod \"install-os-openstack-openstack-networker-xmfpv\" (UID: \"d81a1b59-323c-4254-8e95-b50acb2eb309\") " pod="openstack/install-os-openstack-openstack-networker-xmfpv" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.375429 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj7t7\" (UniqueName: \"kubernetes.io/projected/d81a1b59-323c-4254-8e95-b50acb2eb309-kube-api-access-lj7t7\") pod \"install-os-openstack-openstack-networker-xmfpv\" (UID: \"d81a1b59-323c-4254-8e95-b50acb2eb309\") " pod="openstack/install-os-openstack-openstack-networker-xmfpv" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.375554 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d81a1b59-323c-4254-8e95-b50acb2eb309-inventory\") pod \"install-os-openstack-openstack-networker-xmfpv\" (UID: \"d81a1b59-323c-4254-8e95-b50acb2eb309\") " pod="openstack/install-os-openstack-openstack-networker-xmfpv" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.477089 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d81a1b59-323c-4254-8e95-b50acb2eb309-ssh-key\") pod \"install-os-openstack-openstack-networker-xmfpv\" (UID: \"d81a1b59-323c-4254-8e95-b50acb2eb309\") " pod="openstack/install-os-openstack-openstack-networker-xmfpv" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.477204 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj7t7\" (UniqueName: \"kubernetes.io/projected/d81a1b59-323c-4254-8e95-b50acb2eb309-kube-api-access-lj7t7\") pod \"install-os-openstack-openstack-networker-xmfpv\" (UID: \"d81a1b59-323c-4254-8e95-b50acb2eb309\") " pod="openstack/install-os-openstack-openstack-networker-xmfpv" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.477291 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d81a1b59-323c-4254-8e95-b50acb2eb309-inventory\") pod \"install-os-openstack-openstack-networker-xmfpv\" (UID: \"d81a1b59-323c-4254-8e95-b50acb2eb309\") " pod="openstack/install-os-openstack-openstack-networker-xmfpv" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.482224 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d81a1b59-323c-4254-8e95-b50acb2eb309-ssh-key\") pod \"install-os-openstack-openstack-networker-xmfpv\" (UID: \"d81a1b59-323c-4254-8e95-b50acb2eb309\") " pod="openstack/install-os-openstack-openstack-networker-xmfpv" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.483567 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d81a1b59-323c-4254-8e95-b50acb2eb309-inventory\") pod \"install-os-openstack-openstack-networker-xmfpv\" (UID: \"d81a1b59-323c-4254-8e95-b50acb2eb309\") " pod="openstack/install-os-openstack-openstack-networker-xmfpv" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.495768 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj7t7\" (UniqueName: \"kubernetes.io/projected/d81a1b59-323c-4254-8e95-b50acb2eb309-kube-api-access-lj7t7\") pod \"install-os-openstack-openstack-networker-xmfpv\" (UID: \"d81a1b59-323c-4254-8e95-b50acb2eb309\") " pod="openstack/install-os-openstack-openstack-networker-xmfpv" Dec 05 09:22:02 crc kubenswrapper[4997]: I1205 09:22:02.566104 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-xmfpv" Dec 05 09:22:03 crc kubenswrapper[4997]: I1205 09:22:03.074969 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-networker-xmfpv"] Dec 05 09:22:03 crc kubenswrapper[4997]: W1205 09:22:03.077812 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd81a1b59_323c_4254_8e95_b50acb2eb309.slice/crio-7a5d52a5421654e431d87dd404bf8d20d83ca2523aeee14fb487e282ecd40903 WatchSource:0}: Error finding container 7a5d52a5421654e431d87dd404bf8d20d83ca2523aeee14fb487e282ecd40903: Status 404 returned error can't find the container with id 7a5d52a5421654e431d87dd404bf8d20d83ca2523aeee14fb487e282ecd40903 Dec 05 09:22:03 crc kubenswrapper[4997]: I1205 09:22:03.195256 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-xmfpv" event={"ID":"d81a1b59-323c-4254-8e95-b50acb2eb309","Type":"ContainerStarted","Data":"7a5d52a5421654e431d87dd404bf8d20d83ca2523aeee14fb487e282ecd40903"} Dec 05 09:22:03 crc kubenswrapper[4997]: I1205 09:22:03.781778 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" Dec 05 09:22:03 crc kubenswrapper[4997]: I1205 09:22:03.908972 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bd29\" (UniqueName: \"kubernetes.io/projected/efdefcdd-f76f-4767-af25-b730587e394b-kube-api-access-8bd29\") pod \"efdefcdd-f76f-4767-af25-b730587e394b\" (UID: \"efdefcdd-f76f-4767-af25-b730587e394b\") " Dec 05 09:22:03 crc kubenswrapper[4997]: I1205 09:22:03.909318 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-inventory\") pod \"efdefcdd-f76f-4767-af25-b730587e394b\" (UID: \"efdefcdd-f76f-4767-af25-b730587e394b\") " Dec 05 09:22:03 crc kubenswrapper[4997]: I1205 09:22:03.909420 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-ceph\") pod \"efdefcdd-f76f-4767-af25-b730587e394b\" (UID: \"efdefcdd-f76f-4767-af25-b730587e394b\") " Dec 05 09:22:03 crc kubenswrapper[4997]: I1205 09:22:03.909508 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-ssh-key\") pod \"efdefcdd-f76f-4767-af25-b730587e394b\" (UID: \"efdefcdd-f76f-4767-af25-b730587e394b\") " Dec 05 09:22:03 crc kubenswrapper[4997]: I1205 09:22:03.915280 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-ceph" (OuterVolumeSpecName: "ceph") pod "efdefcdd-f76f-4767-af25-b730587e394b" (UID: "efdefcdd-f76f-4767-af25-b730587e394b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:22:03 crc kubenswrapper[4997]: I1205 09:22:03.915316 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdefcdd-f76f-4767-af25-b730587e394b-kube-api-access-8bd29" (OuterVolumeSpecName: "kube-api-access-8bd29") pod "efdefcdd-f76f-4767-af25-b730587e394b" (UID: "efdefcdd-f76f-4767-af25-b730587e394b"). InnerVolumeSpecName "kube-api-access-8bd29". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:22:03 crc kubenswrapper[4997]: I1205 09:22:03.938610 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-inventory" (OuterVolumeSpecName: "inventory") pod "efdefcdd-f76f-4767-af25-b730587e394b" (UID: "efdefcdd-f76f-4767-af25-b730587e394b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:22:03 crc kubenswrapper[4997]: I1205 09:22:03.941780 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "efdefcdd-f76f-4767-af25-b730587e394b" (UID: "efdefcdd-f76f-4767-af25-b730587e394b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.011560 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.011597 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.011611 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bd29\" (UniqueName: \"kubernetes.io/projected/efdefcdd-f76f-4767-af25-b730587e394b-kube-api-access-8bd29\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.011636 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/efdefcdd-f76f-4767-af25-b730587e394b-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.209475 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-xmfpv" event={"ID":"d81a1b59-323c-4254-8e95-b50acb2eb309","Type":"ContainerStarted","Data":"033bf04f8e06fbe0d050eb2c65892d5fd31e04fd277d4239932da1040447b5c6"} Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.217493 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" event={"ID":"efdefcdd-f76f-4767-af25-b730587e394b","Type":"ContainerDied","Data":"2d2566c63faa3a915fc206fe28fc2ddebb2a3ef5e0857383f139e34c5d4d1172"} Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.217584 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d2566c63faa3a915fc206fe28fc2ddebb2a3ef5e0857383f139e34c5d4d1172" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.217656 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-rlp6r" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.236844 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-networker-xmfpv" podStartSLOduration=1.696672231 podStartE2EDuration="2.236827796s" podCreationTimestamp="2025-12-05 09:22:02 +0000 UTC" firstStartedPulling="2025-12-05 09:22:03.079694032 +0000 UTC m=+8823.608601293" lastFinishedPulling="2025-12-05 09:22:03.619849597 +0000 UTC m=+8824.148756858" observedRunningTime="2025-12-05 09:22:04.235673465 +0000 UTC m=+8824.764580726" watchObservedRunningTime="2025-12-05 09:22:04.236827796 +0000 UTC m=+8824.765735057" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.288257 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-4gmlh"] Dec 05 09:22:04 crc kubenswrapper[4997]: E1205 09:22:04.289044 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efdefcdd-f76f-4767-af25-b730587e394b" containerName="configure-network-openstack-openstack-cell1" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.289067 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="efdefcdd-f76f-4767-af25-b730587e394b" containerName="configure-network-openstack-openstack-cell1" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.289374 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="efdefcdd-f76f-4767-af25-b730587e394b" containerName="configure-network-openstack-openstack-cell1" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.290497 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.292825 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.293003 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.314508 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-4gmlh"] Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.325267 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j7b5\" (UniqueName: \"kubernetes.io/projected/8c00bc35-9325-4d86-b9d5-0871b4842954-kube-api-access-4j7b5\") pod \"validate-network-openstack-openstack-cell1-4gmlh\" (UID: \"8c00bc35-9325-4d86-b9d5-0871b4842954\") " pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.325455 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-ceph\") pod \"validate-network-openstack-openstack-cell1-4gmlh\" (UID: \"8c00bc35-9325-4d86-b9d5-0871b4842954\") " pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.325511 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-inventory\") pod \"validate-network-openstack-openstack-cell1-4gmlh\" (UID: \"8c00bc35-9325-4d86-b9d5-0871b4842954\") " pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.325545 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-ssh-key\") pod \"validate-network-openstack-openstack-cell1-4gmlh\" (UID: \"8c00bc35-9325-4d86-b9d5-0871b4842954\") " pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.427244 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-ceph\") pod \"validate-network-openstack-openstack-cell1-4gmlh\" (UID: \"8c00bc35-9325-4d86-b9d5-0871b4842954\") " pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.427334 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-inventory\") pod \"validate-network-openstack-openstack-cell1-4gmlh\" (UID: \"8c00bc35-9325-4d86-b9d5-0871b4842954\") " pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.427361 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-ssh-key\") pod \"validate-network-openstack-openstack-cell1-4gmlh\" (UID: \"8c00bc35-9325-4d86-b9d5-0871b4842954\") " pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.427451 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j7b5\" (UniqueName: \"kubernetes.io/projected/8c00bc35-9325-4d86-b9d5-0871b4842954-kube-api-access-4j7b5\") pod \"validate-network-openstack-openstack-cell1-4gmlh\" (UID: \"8c00bc35-9325-4d86-b9d5-0871b4842954\") " pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.433042 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-inventory\") pod \"validate-network-openstack-openstack-cell1-4gmlh\" (UID: \"8c00bc35-9325-4d86-b9d5-0871b4842954\") " pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.433175 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-ceph\") pod \"validate-network-openstack-openstack-cell1-4gmlh\" (UID: \"8c00bc35-9325-4d86-b9d5-0871b4842954\") " pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.434203 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-ssh-key\") pod \"validate-network-openstack-openstack-cell1-4gmlh\" (UID: \"8c00bc35-9325-4d86-b9d5-0871b4842954\") " pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.446277 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j7b5\" (UniqueName: \"kubernetes.io/projected/8c00bc35-9325-4d86-b9d5-0871b4842954-kube-api-access-4j7b5\") pod \"validate-network-openstack-openstack-cell1-4gmlh\" (UID: \"8c00bc35-9325-4d86-b9d5-0871b4842954\") " pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" Dec 05 09:22:04 crc kubenswrapper[4997]: I1205 09:22:04.608757 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" Dec 05 09:22:05 crc kubenswrapper[4997]: I1205 09:22:05.643804 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-4gmlh"] Dec 05 09:22:05 crc kubenswrapper[4997]: W1205 09:22:05.687108 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c00bc35_9325_4d86_b9d5_0871b4842954.slice/crio-3d747674390da941a22a043125122643ab3cfcefeb23272198b35c54383a8724 WatchSource:0}: Error finding container 3d747674390da941a22a043125122643ab3cfcefeb23272198b35c54383a8724: Status 404 returned error can't find the container with id 3d747674390da941a22a043125122643ab3cfcefeb23272198b35c54383a8724 Dec 05 09:22:06 crc kubenswrapper[4997]: I1205 09:22:06.269134 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" event={"ID":"8c00bc35-9325-4d86-b9d5-0871b4842954","Type":"ContainerStarted","Data":"3d747674390da941a22a043125122643ab3cfcefeb23272198b35c54383a8724"} Dec 05 09:22:07 crc kubenswrapper[4997]: I1205 09:22:07.280883 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" event={"ID":"8c00bc35-9325-4d86-b9d5-0871b4842954","Type":"ContainerStarted","Data":"f7a32368801c99404e7b2ab750973318be16e0ae4c16ff0c0d01548ca09bf46f"} Dec 05 09:22:07 crc kubenswrapper[4997]: I1205 09:22:07.302713 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" podStartSLOduration=2.839753149 podStartE2EDuration="3.302690888s" podCreationTimestamp="2025-12-05 09:22:04 +0000 UTC" firstStartedPulling="2025-12-05 09:22:05.697222239 +0000 UTC m=+8826.226129500" lastFinishedPulling="2025-12-05 09:22:06.160159968 +0000 UTC m=+8826.689067239" observedRunningTime="2025-12-05 09:22:07.296268825 +0000 UTC m=+8827.825176106" watchObservedRunningTime="2025-12-05 09:22:07.302690888 +0000 UTC m=+8827.831598149" Dec 05 09:22:12 crc kubenswrapper[4997]: I1205 09:22:12.330288 4997 generic.go:334] "Generic (PLEG): container finished" podID="8c00bc35-9325-4d86-b9d5-0871b4842954" containerID="f7a32368801c99404e7b2ab750973318be16e0ae4c16ff0c0d01548ca09bf46f" exitCode=0 Dec 05 09:22:12 crc kubenswrapper[4997]: I1205 09:22:12.330361 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" event={"ID":"8c00bc35-9325-4d86-b9d5-0871b4842954","Type":"ContainerDied","Data":"f7a32368801c99404e7b2ab750973318be16e0ae4c16ff0c0d01548ca09bf46f"} Dec 05 09:22:13 crc kubenswrapper[4997]: I1205 09:22:13.780982 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" Dec 05 09:22:13 crc kubenswrapper[4997]: I1205 09:22:13.931428 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4j7b5\" (UniqueName: \"kubernetes.io/projected/8c00bc35-9325-4d86-b9d5-0871b4842954-kube-api-access-4j7b5\") pod \"8c00bc35-9325-4d86-b9d5-0871b4842954\" (UID: \"8c00bc35-9325-4d86-b9d5-0871b4842954\") " Dec 05 09:22:13 crc kubenswrapper[4997]: I1205 09:22:13.931529 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-ceph\") pod \"8c00bc35-9325-4d86-b9d5-0871b4842954\" (UID: \"8c00bc35-9325-4d86-b9d5-0871b4842954\") " Dec 05 09:22:13 crc kubenswrapper[4997]: I1205 09:22:13.931566 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-inventory\") pod \"8c00bc35-9325-4d86-b9d5-0871b4842954\" (UID: \"8c00bc35-9325-4d86-b9d5-0871b4842954\") " Dec 05 09:22:13 crc kubenswrapper[4997]: I1205 09:22:13.931665 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-ssh-key\") pod \"8c00bc35-9325-4d86-b9d5-0871b4842954\" (UID: \"8c00bc35-9325-4d86-b9d5-0871b4842954\") " Dec 05 09:22:13 crc kubenswrapper[4997]: I1205 09:22:13.938209 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-ceph" (OuterVolumeSpecName: "ceph") pod "8c00bc35-9325-4d86-b9d5-0871b4842954" (UID: "8c00bc35-9325-4d86-b9d5-0871b4842954"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:22:13 crc kubenswrapper[4997]: I1205 09:22:13.938365 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c00bc35-9325-4d86-b9d5-0871b4842954-kube-api-access-4j7b5" (OuterVolumeSpecName: "kube-api-access-4j7b5") pod "8c00bc35-9325-4d86-b9d5-0871b4842954" (UID: "8c00bc35-9325-4d86-b9d5-0871b4842954"). InnerVolumeSpecName "kube-api-access-4j7b5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:22:13 crc kubenswrapper[4997]: I1205 09:22:13.961875 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-inventory" (OuterVolumeSpecName: "inventory") pod "8c00bc35-9325-4d86-b9d5-0871b4842954" (UID: "8c00bc35-9325-4d86-b9d5-0871b4842954"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:22:13 crc kubenswrapper[4997]: I1205 09:22:13.962294 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8c00bc35-9325-4d86-b9d5-0871b4842954" (UID: "8c00bc35-9325-4d86-b9d5-0871b4842954"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.034116 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.034404 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4j7b5\" (UniqueName: \"kubernetes.io/projected/8c00bc35-9325-4d86-b9d5-0871b4842954-kube-api-access-4j7b5\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.034417 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.034426 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c00bc35-9325-4d86-b9d5-0871b4842954-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.350023 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" event={"ID":"8c00bc35-9325-4d86-b9d5-0871b4842954","Type":"ContainerDied","Data":"3d747674390da941a22a043125122643ab3cfcefeb23272198b35c54383a8724"} Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.350061 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-4gmlh" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.350069 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d747674390da941a22a043125122643ab3cfcefeb23272198b35c54383a8724" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.420380 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-bxss2"] Dec 05 09:22:14 crc kubenswrapper[4997]: E1205 09:22:14.420835 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c00bc35-9325-4d86-b9d5-0871b4842954" containerName="validate-network-openstack-openstack-cell1" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.420856 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c00bc35-9325-4d86-b9d5-0871b4842954" containerName="validate-network-openstack-openstack-cell1" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.421052 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c00bc35-9325-4d86-b9d5-0871b4842954" containerName="validate-network-openstack-openstack-cell1" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.421883 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-bxss2" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.423575 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.424058 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.444088 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-bxss2"] Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.544279 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-inventory\") pod \"install-os-openstack-openstack-cell1-bxss2\" (UID: \"a9704a5b-2b0a-4786-b151-29cddb30973f\") " pod="openstack/install-os-openstack-openstack-cell1-bxss2" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.544332 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-ssh-key\") pod \"install-os-openstack-openstack-cell1-bxss2\" (UID: \"a9704a5b-2b0a-4786-b151-29cddb30973f\") " pod="openstack/install-os-openstack-openstack-cell1-bxss2" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.544374 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-ceph\") pod \"install-os-openstack-openstack-cell1-bxss2\" (UID: \"a9704a5b-2b0a-4786-b151-29cddb30973f\") " pod="openstack/install-os-openstack-openstack-cell1-bxss2" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.544582 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkvbk\" (UniqueName: \"kubernetes.io/projected/a9704a5b-2b0a-4786-b151-29cddb30973f-kube-api-access-wkvbk\") pod \"install-os-openstack-openstack-cell1-bxss2\" (UID: \"a9704a5b-2b0a-4786-b151-29cddb30973f\") " pod="openstack/install-os-openstack-openstack-cell1-bxss2" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.647142 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-inventory\") pod \"install-os-openstack-openstack-cell1-bxss2\" (UID: \"a9704a5b-2b0a-4786-b151-29cddb30973f\") " pod="openstack/install-os-openstack-openstack-cell1-bxss2" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.647202 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-ssh-key\") pod \"install-os-openstack-openstack-cell1-bxss2\" (UID: \"a9704a5b-2b0a-4786-b151-29cddb30973f\") " pod="openstack/install-os-openstack-openstack-cell1-bxss2" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.647257 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-ceph\") pod \"install-os-openstack-openstack-cell1-bxss2\" (UID: \"a9704a5b-2b0a-4786-b151-29cddb30973f\") " pod="openstack/install-os-openstack-openstack-cell1-bxss2" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.647344 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkvbk\" (UniqueName: \"kubernetes.io/projected/a9704a5b-2b0a-4786-b151-29cddb30973f-kube-api-access-wkvbk\") pod \"install-os-openstack-openstack-cell1-bxss2\" (UID: \"a9704a5b-2b0a-4786-b151-29cddb30973f\") " pod="openstack/install-os-openstack-openstack-cell1-bxss2" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.652363 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-inventory\") pod \"install-os-openstack-openstack-cell1-bxss2\" (UID: \"a9704a5b-2b0a-4786-b151-29cddb30973f\") " pod="openstack/install-os-openstack-openstack-cell1-bxss2" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.652581 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-ssh-key\") pod \"install-os-openstack-openstack-cell1-bxss2\" (UID: \"a9704a5b-2b0a-4786-b151-29cddb30973f\") " pod="openstack/install-os-openstack-openstack-cell1-bxss2" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.656237 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-ceph\") pod \"install-os-openstack-openstack-cell1-bxss2\" (UID: \"a9704a5b-2b0a-4786-b151-29cddb30973f\") " pod="openstack/install-os-openstack-openstack-cell1-bxss2" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.666635 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkvbk\" (UniqueName: \"kubernetes.io/projected/a9704a5b-2b0a-4786-b151-29cddb30973f-kube-api-access-wkvbk\") pod \"install-os-openstack-openstack-cell1-bxss2\" (UID: \"a9704a5b-2b0a-4786-b151-29cddb30973f\") " pod="openstack/install-os-openstack-openstack-cell1-bxss2" Dec 05 09:22:14 crc kubenswrapper[4997]: I1205 09:22:14.739497 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-bxss2" Dec 05 09:22:15 crc kubenswrapper[4997]: I1205 09:22:15.315700 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-bxss2"] Dec 05 09:22:15 crc kubenswrapper[4997]: I1205 09:22:15.367566 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-bxss2" event={"ID":"a9704a5b-2b0a-4786-b151-29cddb30973f","Type":"ContainerStarted","Data":"2c28e00ecf5086f0abbd421b6973da11c7076b9638c75d8f732a893afa048f0e"} Dec 05 09:22:16 crc kubenswrapper[4997]: I1205 09:22:16.378303 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-bxss2" event={"ID":"a9704a5b-2b0a-4786-b151-29cddb30973f","Type":"ContainerStarted","Data":"61b8ad8c78a38e14b868c885f35ab661104d46cb9da8284a9066c8c3b7f4a8f7"} Dec 05 09:22:16 crc kubenswrapper[4997]: I1205 09:22:16.395669 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-bxss2" podStartSLOduration=1.9600074090000001 podStartE2EDuration="2.395649801s" podCreationTimestamp="2025-12-05 09:22:14 +0000 UTC" firstStartedPulling="2025-12-05 09:22:15.32455762 +0000 UTC m=+8835.853464881" lastFinishedPulling="2025-12-05 09:22:15.760200012 +0000 UTC m=+8836.289107273" observedRunningTime="2025-12-05 09:22:16.394550531 +0000 UTC m=+8836.923457812" watchObservedRunningTime="2025-12-05 09:22:16.395649801 +0000 UTC m=+8836.924557082" Dec 05 09:22:19 crc kubenswrapper[4997]: I1205 09:22:19.770429 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:22:19 crc kubenswrapper[4997]: I1205 09:22:19.770999 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:22:49 crc kubenswrapper[4997]: I1205 09:22:49.770813 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:22:49 crc kubenswrapper[4997]: I1205 09:22:49.771373 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:22:49 crc kubenswrapper[4997]: I1205 09:22:49.771421 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 09:22:49 crc kubenswrapper[4997]: I1205 09:22:49.772250 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:22:49 crc kubenswrapper[4997]: I1205 09:22:49.772309 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" gracePeriod=600 Dec 05 09:22:49 crc kubenswrapper[4997]: E1205 09:22:49.898242 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:22:50 crc kubenswrapper[4997]: I1205 09:22:50.727916 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" exitCode=0 Dec 05 09:22:50 crc kubenswrapper[4997]: I1205 09:22:50.727969 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a"} Dec 05 09:22:50 crc kubenswrapper[4997]: I1205 09:22:50.728311 4997 scope.go:117] "RemoveContainer" containerID="f8205236afd16f8a0c6d433a5ab02465034afc3f440c01d487df254a7a73dfb2" Dec 05 09:22:50 crc kubenswrapper[4997]: I1205 09:22:50.729756 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:22:50 crc kubenswrapper[4997]: E1205 09:22:50.730455 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:22:54 crc kubenswrapper[4997]: I1205 09:22:54.773948 4997 generic.go:334] "Generic (PLEG): container finished" podID="d81a1b59-323c-4254-8e95-b50acb2eb309" containerID="033bf04f8e06fbe0d050eb2c65892d5fd31e04fd277d4239932da1040447b5c6" exitCode=0 Dec 05 09:22:54 crc kubenswrapper[4997]: I1205 09:22:54.774210 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-xmfpv" event={"ID":"d81a1b59-323c-4254-8e95-b50acb2eb309","Type":"ContainerDied","Data":"033bf04f8e06fbe0d050eb2c65892d5fd31e04fd277d4239932da1040447b5c6"} Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.262132 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-xmfpv" Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.343974 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lj7t7\" (UniqueName: \"kubernetes.io/projected/d81a1b59-323c-4254-8e95-b50acb2eb309-kube-api-access-lj7t7\") pod \"d81a1b59-323c-4254-8e95-b50acb2eb309\" (UID: \"d81a1b59-323c-4254-8e95-b50acb2eb309\") " Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.344209 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d81a1b59-323c-4254-8e95-b50acb2eb309-inventory\") pod \"d81a1b59-323c-4254-8e95-b50acb2eb309\" (UID: \"d81a1b59-323c-4254-8e95-b50acb2eb309\") " Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.344392 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d81a1b59-323c-4254-8e95-b50acb2eb309-ssh-key\") pod \"d81a1b59-323c-4254-8e95-b50acb2eb309\" (UID: \"d81a1b59-323c-4254-8e95-b50acb2eb309\") " Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.356035 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d81a1b59-323c-4254-8e95-b50acb2eb309-kube-api-access-lj7t7" (OuterVolumeSpecName: "kube-api-access-lj7t7") pod "d81a1b59-323c-4254-8e95-b50acb2eb309" (UID: "d81a1b59-323c-4254-8e95-b50acb2eb309"). InnerVolumeSpecName "kube-api-access-lj7t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.372382 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d81a1b59-323c-4254-8e95-b50acb2eb309-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d81a1b59-323c-4254-8e95-b50acb2eb309" (UID: "d81a1b59-323c-4254-8e95-b50acb2eb309"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.385816 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d81a1b59-323c-4254-8e95-b50acb2eb309-inventory" (OuterVolumeSpecName: "inventory") pod "d81a1b59-323c-4254-8e95-b50acb2eb309" (UID: "d81a1b59-323c-4254-8e95-b50acb2eb309"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.447494 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d81a1b59-323c-4254-8e95-b50acb2eb309-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.447536 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lj7t7\" (UniqueName: \"kubernetes.io/projected/d81a1b59-323c-4254-8e95-b50acb2eb309-kube-api-access-lj7t7\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.447549 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d81a1b59-323c-4254-8e95-b50acb2eb309-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.806194 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-xmfpv" event={"ID":"d81a1b59-323c-4254-8e95-b50acb2eb309","Type":"ContainerDied","Data":"7a5d52a5421654e431d87dd404bf8d20d83ca2523aeee14fb487e282ecd40903"} Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.806257 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a5d52a5421654e431d87dd404bf8d20d83ca2523aeee14fb487e282ecd40903" Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.806292 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-xmfpv" Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.958313 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-networker-29djd"] Dec 05 09:22:56 crc kubenswrapper[4997]: E1205 09:22:56.958837 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d81a1b59-323c-4254-8e95-b50acb2eb309" containerName="install-os-openstack-openstack-networker" Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.958860 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d81a1b59-323c-4254-8e95-b50acb2eb309" containerName="install-os-openstack-openstack-networker" Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.959109 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="d81a1b59-323c-4254-8e95-b50acb2eb309" containerName="install-os-openstack-openstack-networker" Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.959843 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-29djd" Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.962057 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-4542r" Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.966702 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Dec 05 09:22:56 crc kubenswrapper[4997]: I1205 09:22:56.978021 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-networker-29djd"] Dec 05 09:22:57 crc kubenswrapper[4997]: I1205 09:22:57.060378 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpks7\" (UniqueName: \"kubernetes.io/projected/888fe326-b46b-45f0-9dfe-6875d4ef6def-kube-api-access-cpks7\") pod \"configure-os-openstack-openstack-networker-29djd\" (UID: \"888fe326-b46b-45f0-9dfe-6875d4ef6def\") " pod="openstack/configure-os-openstack-openstack-networker-29djd" Dec 05 09:22:57 crc kubenswrapper[4997]: I1205 09:22:57.061210 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/888fe326-b46b-45f0-9dfe-6875d4ef6def-inventory\") pod \"configure-os-openstack-openstack-networker-29djd\" (UID: \"888fe326-b46b-45f0-9dfe-6875d4ef6def\") " pod="openstack/configure-os-openstack-openstack-networker-29djd" Dec 05 09:22:57 crc kubenswrapper[4997]: I1205 09:22:57.061563 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/888fe326-b46b-45f0-9dfe-6875d4ef6def-ssh-key\") pod \"configure-os-openstack-openstack-networker-29djd\" (UID: \"888fe326-b46b-45f0-9dfe-6875d4ef6def\") " pod="openstack/configure-os-openstack-openstack-networker-29djd" Dec 05 09:22:57 crc kubenswrapper[4997]: I1205 09:22:57.163349 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/888fe326-b46b-45f0-9dfe-6875d4ef6def-inventory\") pod \"configure-os-openstack-openstack-networker-29djd\" (UID: \"888fe326-b46b-45f0-9dfe-6875d4ef6def\") " pod="openstack/configure-os-openstack-openstack-networker-29djd" Dec 05 09:22:57 crc kubenswrapper[4997]: I1205 09:22:57.163419 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/888fe326-b46b-45f0-9dfe-6875d4ef6def-ssh-key\") pod \"configure-os-openstack-openstack-networker-29djd\" (UID: \"888fe326-b46b-45f0-9dfe-6875d4ef6def\") " pod="openstack/configure-os-openstack-openstack-networker-29djd" Dec 05 09:22:57 crc kubenswrapper[4997]: I1205 09:22:57.163452 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpks7\" (UniqueName: \"kubernetes.io/projected/888fe326-b46b-45f0-9dfe-6875d4ef6def-kube-api-access-cpks7\") pod \"configure-os-openstack-openstack-networker-29djd\" (UID: \"888fe326-b46b-45f0-9dfe-6875d4ef6def\") " pod="openstack/configure-os-openstack-openstack-networker-29djd" Dec 05 09:22:57 crc kubenswrapper[4997]: I1205 09:22:57.167240 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/888fe326-b46b-45f0-9dfe-6875d4ef6def-ssh-key\") pod \"configure-os-openstack-openstack-networker-29djd\" (UID: \"888fe326-b46b-45f0-9dfe-6875d4ef6def\") " pod="openstack/configure-os-openstack-openstack-networker-29djd" Dec 05 09:22:57 crc kubenswrapper[4997]: I1205 09:22:57.167447 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/888fe326-b46b-45f0-9dfe-6875d4ef6def-inventory\") pod \"configure-os-openstack-openstack-networker-29djd\" (UID: \"888fe326-b46b-45f0-9dfe-6875d4ef6def\") " pod="openstack/configure-os-openstack-openstack-networker-29djd" Dec 05 09:22:57 crc kubenswrapper[4997]: I1205 09:22:57.182654 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpks7\" (UniqueName: \"kubernetes.io/projected/888fe326-b46b-45f0-9dfe-6875d4ef6def-kube-api-access-cpks7\") pod \"configure-os-openstack-openstack-networker-29djd\" (UID: \"888fe326-b46b-45f0-9dfe-6875d4ef6def\") " pod="openstack/configure-os-openstack-openstack-networker-29djd" Dec 05 09:22:57 crc kubenswrapper[4997]: I1205 09:22:57.279955 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-29djd" Dec 05 09:22:57 crc kubenswrapper[4997]: I1205 09:22:57.868517 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-networker-29djd"] Dec 05 09:22:58 crc kubenswrapper[4997]: I1205 09:22:58.831011 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-29djd" event={"ID":"888fe326-b46b-45f0-9dfe-6875d4ef6def","Type":"ContainerStarted","Data":"d257960344b22351752e1bbd527be00dd43de6edb194fe30c281a31a1bd63d2d"} Dec 05 09:22:58 crc kubenswrapper[4997]: I1205 09:22:58.832476 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-29djd" event={"ID":"888fe326-b46b-45f0-9dfe-6875d4ef6def","Type":"ContainerStarted","Data":"1966c82d357f927d0529d84058ab0e66eb454eec45f159ce3019c9504f1c6913"} Dec 05 09:22:58 crc kubenswrapper[4997]: I1205 09:22:58.856404 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-networker-29djd" podStartSLOduration=2.480480567 podStartE2EDuration="2.856380458s" podCreationTimestamp="2025-12-05 09:22:56 +0000 UTC" firstStartedPulling="2025-12-05 09:22:57.873307872 +0000 UTC m=+8878.402215133" lastFinishedPulling="2025-12-05 09:22:58.249207763 +0000 UTC m=+8878.778115024" observedRunningTime="2025-12-05 09:22:58.844566118 +0000 UTC m=+8879.373473389" watchObservedRunningTime="2025-12-05 09:22:58.856380458 +0000 UTC m=+8879.385287719" Dec 05 09:23:03 crc kubenswrapper[4997]: I1205 09:23:03.752673 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:23:03 crc kubenswrapper[4997]: E1205 09:23:03.753377 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:23:06 crc kubenswrapper[4997]: I1205 09:23:06.908872 4997 generic.go:334] "Generic (PLEG): container finished" podID="a9704a5b-2b0a-4786-b151-29cddb30973f" containerID="61b8ad8c78a38e14b868c885f35ab661104d46cb9da8284a9066c8c3b7f4a8f7" exitCode=0 Dec 05 09:23:06 crc kubenswrapper[4997]: I1205 09:23:06.908971 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-bxss2" event={"ID":"a9704a5b-2b0a-4786-b151-29cddb30973f","Type":"ContainerDied","Data":"61b8ad8c78a38e14b868c885f35ab661104d46cb9da8284a9066c8c3b7f4a8f7"} Dec 05 09:23:08 crc kubenswrapper[4997]: I1205 09:23:08.434495 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-bxss2" Dec 05 09:23:08 crc kubenswrapper[4997]: I1205 09:23:08.593452 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkvbk\" (UniqueName: \"kubernetes.io/projected/a9704a5b-2b0a-4786-b151-29cddb30973f-kube-api-access-wkvbk\") pod \"a9704a5b-2b0a-4786-b151-29cddb30973f\" (UID: \"a9704a5b-2b0a-4786-b151-29cddb30973f\") " Dec 05 09:23:08 crc kubenswrapper[4997]: I1205 09:23:08.593862 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-ceph\") pod \"a9704a5b-2b0a-4786-b151-29cddb30973f\" (UID: \"a9704a5b-2b0a-4786-b151-29cddb30973f\") " Dec 05 09:23:08 crc kubenswrapper[4997]: I1205 09:23:08.593972 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-inventory\") pod \"a9704a5b-2b0a-4786-b151-29cddb30973f\" (UID: \"a9704a5b-2b0a-4786-b151-29cddb30973f\") " Dec 05 09:23:08 crc kubenswrapper[4997]: I1205 09:23:08.594079 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-ssh-key\") pod \"a9704a5b-2b0a-4786-b151-29cddb30973f\" (UID: \"a9704a5b-2b0a-4786-b151-29cddb30973f\") " Dec 05 09:23:08 crc kubenswrapper[4997]: I1205 09:23:08.598732 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9704a5b-2b0a-4786-b151-29cddb30973f-kube-api-access-wkvbk" (OuterVolumeSpecName: "kube-api-access-wkvbk") pod "a9704a5b-2b0a-4786-b151-29cddb30973f" (UID: "a9704a5b-2b0a-4786-b151-29cddb30973f"). InnerVolumeSpecName "kube-api-access-wkvbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:23:08 crc kubenswrapper[4997]: I1205 09:23:08.600036 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-ceph" (OuterVolumeSpecName: "ceph") pod "a9704a5b-2b0a-4786-b151-29cddb30973f" (UID: "a9704a5b-2b0a-4786-b151-29cddb30973f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:23:08 crc kubenswrapper[4997]: I1205 09:23:08.621312 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a9704a5b-2b0a-4786-b151-29cddb30973f" (UID: "a9704a5b-2b0a-4786-b151-29cddb30973f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:23:08 crc kubenswrapper[4997]: I1205 09:23:08.624086 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-inventory" (OuterVolumeSpecName: "inventory") pod "a9704a5b-2b0a-4786-b151-29cddb30973f" (UID: "a9704a5b-2b0a-4786-b151-29cddb30973f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:23:08 crc kubenswrapper[4997]: I1205 09:23:08.696815 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkvbk\" (UniqueName: \"kubernetes.io/projected/a9704a5b-2b0a-4786-b151-29cddb30973f-kube-api-access-wkvbk\") on node \"crc\" DevicePath \"\"" Dec 05 09:23:08 crc kubenswrapper[4997]: I1205 09:23:08.697062 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:23:08 crc kubenswrapper[4997]: I1205 09:23:08.697183 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:23:08 crc kubenswrapper[4997]: I1205 09:23:08.697261 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9704a5b-2b0a-4786-b151-29cddb30973f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:23:08 crc kubenswrapper[4997]: I1205 09:23:08.956744 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-bxss2" event={"ID":"a9704a5b-2b0a-4786-b151-29cddb30973f","Type":"ContainerDied","Data":"2c28e00ecf5086f0abbd421b6973da11c7076b9638c75d8f732a893afa048f0e"} Dec 05 09:23:08 crc kubenswrapper[4997]: I1205 09:23:08.957057 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c28e00ecf5086f0abbd421b6973da11c7076b9638c75d8f732a893afa048f0e" Dec 05 09:23:08 crc kubenswrapper[4997]: I1205 09:23:08.956789 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-bxss2" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.005986 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-jsbwt"] Dec 05 09:23:09 crc kubenswrapper[4997]: E1205 09:23:09.006460 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9704a5b-2b0a-4786-b151-29cddb30973f" containerName="install-os-openstack-openstack-cell1" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.006484 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9704a5b-2b0a-4786-b151-29cddb30973f" containerName="install-os-openstack-openstack-cell1" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.006777 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9704a5b-2b0a-4786-b151-29cddb30973f" containerName="install-os-openstack-openstack-cell1" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.007588 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.010563 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.019795 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.020812 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-jsbwt"] Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.207064 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-ceph\") pod \"configure-os-openstack-openstack-cell1-jsbwt\" (UID: \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\") " pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.207125 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wk7s\" (UniqueName: \"kubernetes.io/projected/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-kube-api-access-8wk7s\") pod \"configure-os-openstack-openstack-cell1-jsbwt\" (UID: \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\") " pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.207150 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-ssh-key\") pod \"configure-os-openstack-openstack-cell1-jsbwt\" (UID: \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\") " pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.207316 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-inventory\") pod \"configure-os-openstack-openstack-cell1-jsbwt\" (UID: \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\") " pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.308998 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-ceph\") pod \"configure-os-openstack-openstack-cell1-jsbwt\" (UID: \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\") " pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.309058 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wk7s\" (UniqueName: \"kubernetes.io/projected/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-kube-api-access-8wk7s\") pod \"configure-os-openstack-openstack-cell1-jsbwt\" (UID: \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\") " pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.309086 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-ssh-key\") pod \"configure-os-openstack-openstack-cell1-jsbwt\" (UID: \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\") " pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.309191 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-inventory\") pod \"configure-os-openstack-openstack-cell1-jsbwt\" (UID: \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\") " pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.314194 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-ceph\") pod \"configure-os-openstack-openstack-cell1-jsbwt\" (UID: \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\") " pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.314339 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-ssh-key\") pod \"configure-os-openstack-openstack-cell1-jsbwt\" (UID: \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\") " pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.314651 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-inventory\") pod \"configure-os-openstack-openstack-cell1-jsbwt\" (UID: \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\") " pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.325319 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wk7s\" (UniqueName: \"kubernetes.io/projected/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-kube-api-access-8wk7s\") pod \"configure-os-openstack-openstack-cell1-jsbwt\" (UID: \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\") " pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.332362 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.686702 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-jsbwt"] Dec 05 09:23:09 crc kubenswrapper[4997]: I1205 09:23:09.966357 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" event={"ID":"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4","Type":"ContainerStarted","Data":"0d585084a3626ee981933050f86fc8ccae6fb1878959d3922a3418535f63d995"} Dec 05 09:23:10 crc kubenswrapper[4997]: I1205 09:23:10.996105 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" event={"ID":"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4","Type":"ContainerStarted","Data":"e43784f5a0adf9a298358043721f398ccba08541a6fa81b4e4f01451af4e3632"} Dec 05 09:23:11 crc kubenswrapper[4997]: I1205 09:23:11.020915 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" podStartSLOduration=2.6215989840000002 podStartE2EDuration="3.020882485s" podCreationTimestamp="2025-12-05 09:23:08 +0000 UTC" firstStartedPulling="2025-12-05 09:23:09.694802339 +0000 UTC m=+8890.223709600" lastFinishedPulling="2025-12-05 09:23:10.09408584 +0000 UTC m=+8890.622993101" observedRunningTime="2025-12-05 09:23:11.010751821 +0000 UTC m=+8891.539659112" watchObservedRunningTime="2025-12-05 09:23:11.020882485 +0000 UTC m=+8891.549789756" Dec 05 09:23:17 crc kubenswrapper[4997]: I1205 09:23:17.750960 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:23:17 crc kubenswrapper[4997]: E1205 09:23:17.752468 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:23:32 crc kubenswrapper[4997]: I1205 09:23:32.749214 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:23:32 crc kubenswrapper[4997]: E1205 09:23:32.750116 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:23:46 crc kubenswrapper[4997]: I1205 09:23:46.749495 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:23:46 crc kubenswrapper[4997]: E1205 09:23:46.750239 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:23:52 crc kubenswrapper[4997]: I1205 09:23:52.444055 4997 generic.go:334] "Generic (PLEG): container finished" podID="888fe326-b46b-45f0-9dfe-6875d4ef6def" containerID="d257960344b22351752e1bbd527be00dd43de6edb194fe30c281a31a1bd63d2d" exitCode=0 Dec 05 09:23:52 crc kubenswrapper[4997]: I1205 09:23:52.444125 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-29djd" event={"ID":"888fe326-b46b-45f0-9dfe-6875d4ef6def","Type":"ContainerDied","Data":"d257960344b22351752e1bbd527be00dd43de6edb194fe30c281a31a1bd63d2d"} Dec 05 09:23:53 crc kubenswrapper[4997]: I1205 09:23:53.932476 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-29djd" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.040886 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpks7\" (UniqueName: \"kubernetes.io/projected/888fe326-b46b-45f0-9dfe-6875d4ef6def-kube-api-access-cpks7\") pod \"888fe326-b46b-45f0-9dfe-6875d4ef6def\" (UID: \"888fe326-b46b-45f0-9dfe-6875d4ef6def\") " Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.041151 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/888fe326-b46b-45f0-9dfe-6875d4ef6def-ssh-key\") pod \"888fe326-b46b-45f0-9dfe-6875d4ef6def\" (UID: \"888fe326-b46b-45f0-9dfe-6875d4ef6def\") " Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.041236 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/888fe326-b46b-45f0-9dfe-6875d4ef6def-inventory\") pod \"888fe326-b46b-45f0-9dfe-6875d4ef6def\" (UID: \"888fe326-b46b-45f0-9dfe-6875d4ef6def\") " Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.068105 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/888fe326-b46b-45f0-9dfe-6875d4ef6def-kube-api-access-cpks7" (OuterVolumeSpecName: "kube-api-access-cpks7") pod "888fe326-b46b-45f0-9dfe-6875d4ef6def" (UID: "888fe326-b46b-45f0-9dfe-6875d4ef6def"). InnerVolumeSpecName "kube-api-access-cpks7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.075813 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/888fe326-b46b-45f0-9dfe-6875d4ef6def-inventory" (OuterVolumeSpecName: "inventory") pod "888fe326-b46b-45f0-9dfe-6875d4ef6def" (UID: "888fe326-b46b-45f0-9dfe-6875d4ef6def"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.083812 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/888fe326-b46b-45f0-9dfe-6875d4ef6def-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "888fe326-b46b-45f0-9dfe-6875d4ef6def" (UID: "888fe326-b46b-45f0-9dfe-6875d4ef6def"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.144383 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpks7\" (UniqueName: \"kubernetes.io/projected/888fe326-b46b-45f0-9dfe-6875d4ef6def-kube-api-access-cpks7\") on node \"crc\" DevicePath \"\"" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.144462 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/888fe326-b46b-45f0-9dfe-6875d4ef6def-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.144474 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/888fe326-b46b-45f0-9dfe-6875d4ef6def-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.465070 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-29djd" event={"ID":"888fe326-b46b-45f0-9dfe-6875d4ef6def","Type":"ContainerDied","Data":"1966c82d357f927d0529d84058ab0e66eb454eec45f159ce3019c9504f1c6913"} Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.465122 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1966c82d357f927d0529d84058ab0e66eb454eec45f159ce3019c9504f1c6913" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.465188 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-29djd" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.555457 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-networker-spxhg"] Dec 05 09:23:54 crc kubenswrapper[4997]: E1205 09:23:54.556188 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="888fe326-b46b-45f0-9dfe-6875d4ef6def" containerName="configure-os-openstack-openstack-networker" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.556205 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="888fe326-b46b-45f0-9dfe-6875d4ef6def" containerName="configure-os-openstack-openstack-networker" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.556398 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="888fe326-b46b-45f0-9dfe-6875d4ef6def" containerName="configure-os-openstack-openstack-networker" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.557120 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-spxhg" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.560924 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.560937 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-4542r" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.567512 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-networker-spxhg"] Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.653886 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c036120-5481-45e6-8575-5df04a770fd1-ssh-key\") pod \"run-os-openstack-openstack-networker-spxhg\" (UID: \"9c036120-5481-45e6-8575-5df04a770fd1\") " pod="openstack/run-os-openstack-openstack-networker-spxhg" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.653980 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgng8\" (UniqueName: \"kubernetes.io/projected/9c036120-5481-45e6-8575-5df04a770fd1-kube-api-access-sgng8\") pod \"run-os-openstack-openstack-networker-spxhg\" (UID: \"9c036120-5481-45e6-8575-5df04a770fd1\") " pod="openstack/run-os-openstack-openstack-networker-spxhg" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.654078 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c036120-5481-45e6-8575-5df04a770fd1-inventory\") pod \"run-os-openstack-openstack-networker-spxhg\" (UID: \"9c036120-5481-45e6-8575-5df04a770fd1\") " pod="openstack/run-os-openstack-openstack-networker-spxhg" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.756561 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c036120-5481-45e6-8575-5df04a770fd1-inventory\") pod \"run-os-openstack-openstack-networker-spxhg\" (UID: \"9c036120-5481-45e6-8575-5df04a770fd1\") " pod="openstack/run-os-openstack-openstack-networker-spxhg" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.756718 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c036120-5481-45e6-8575-5df04a770fd1-ssh-key\") pod \"run-os-openstack-openstack-networker-spxhg\" (UID: \"9c036120-5481-45e6-8575-5df04a770fd1\") " pod="openstack/run-os-openstack-openstack-networker-spxhg" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.756787 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgng8\" (UniqueName: \"kubernetes.io/projected/9c036120-5481-45e6-8575-5df04a770fd1-kube-api-access-sgng8\") pod \"run-os-openstack-openstack-networker-spxhg\" (UID: \"9c036120-5481-45e6-8575-5df04a770fd1\") " pod="openstack/run-os-openstack-openstack-networker-spxhg" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.761952 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c036120-5481-45e6-8575-5df04a770fd1-ssh-key\") pod \"run-os-openstack-openstack-networker-spxhg\" (UID: \"9c036120-5481-45e6-8575-5df04a770fd1\") " pod="openstack/run-os-openstack-openstack-networker-spxhg" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.763063 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c036120-5481-45e6-8575-5df04a770fd1-inventory\") pod \"run-os-openstack-openstack-networker-spxhg\" (UID: \"9c036120-5481-45e6-8575-5df04a770fd1\") " pod="openstack/run-os-openstack-openstack-networker-spxhg" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.774721 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgng8\" (UniqueName: \"kubernetes.io/projected/9c036120-5481-45e6-8575-5df04a770fd1-kube-api-access-sgng8\") pod \"run-os-openstack-openstack-networker-spxhg\" (UID: \"9c036120-5481-45e6-8575-5df04a770fd1\") " pod="openstack/run-os-openstack-openstack-networker-spxhg" Dec 05 09:23:54 crc kubenswrapper[4997]: I1205 09:23:54.877242 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-spxhg" Dec 05 09:23:55 crc kubenswrapper[4997]: I1205 09:23:55.418913 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-networker-spxhg"] Dec 05 09:23:55 crc kubenswrapper[4997]: I1205 09:23:55.475017 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-spxhg" event={"ID":"9c036120-5481-45e6-8575-5df04a770fd1","Type":"ContainerStarted","Data":"6a3b25be823055e067912b584459fbb354ef346222256a241808e921c348f914"} Dec 05 09:23:56 crc kubenswrapper[4997]: I1205 09:23:56.486866 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-spxhg" event={"ID":"9c036120-5481-45e6-8575-5df04a770fd1","Type":"ContainerStarted","Data":"3db26440c735414fd25fe966a6214eebc202a8e97d9e8e458af2b70e1f392bc9"} Dec 05 09:23:56 crc kubenswrapper[4997]: I1205 09:23:56.509326 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-networker-spxhg" podStartSLOduration=1.9708937450000001 podStartE2EDuration="2.509307153s" podCreationTimestamp="2025-12-05 09:23:54 +0000 UTC" firstStartedPulling="2025-12-05 09:23:55.425508979 +0000 UTC m=+8935.954416240" lastFinishedPulling="2025-12-05 09:23:55.963922367 +0000 UTC m=+8936.492829648" observedRunningTime="2025-12-05 09:23:56.502120739 +0000 UTC m=+8937.031028010" watchObservedRunningTime="2025-12-05 09:23:56.509307153 +0000 UTC m=+8937.038214414" Dec 05 09:24:00 crc kubenswrapper[4997]: I1205 09:24:00.749973 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:24:00 crc kubenswrapper[4997]: E1205 09:24:00.750881 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:24:06 crc kubenswrapper[4997]: I1205 09:24:06.601796 4997 generic.go:334] "Generic (PLEG): container finished" podID="9c036120-5481-45e6-8575-5df04a770fd1" containerID="3db26440c735414fd25fe966a6214eebc202a8e97d9e8e458af2b70e1f392bc9" exitCode=0 Dec 05 09:24:06 crc kubenswrapper[4997]: I1205 09:24:06.601880 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-spxhg" event={"ID":"9c036120-5481-45e6-8575-5df04a770fd1","Type":"ContainerDied","Data":"3db26440c735414fd25fe966a6214eebc202a8e97d9e8e458af2b70e1f392bc9"} Dec 05 09:24:07 crc kubenswrapper[4997]: I1205 09:24:07.628174 4997 generic.go:334] "Generic (PLEG): container finished" podID="cfe38bed-b2b8-4d90-b74b-d82cea4b63a4" containerID="e43784f5a0adf9a298358043721f398ccba08541a6fa81b4e4f01451af4e3632" exitCode=0 Dec 05 09:24:07 crc kubenswrapper[4997]: I1205 09:24:07.628450 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" event={"ID":"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4","Type":"ContainerDied","Data":"e43784f5a0adf9a298358043721f398ccba08541a6fa81b4e4f01451af4e3632"} Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.057909 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-spxhg" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.123296 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c036120-5481-45e6-8575-5df04a770fd1-inventory\") pod \"9c036120-5481-45e6-8575-5df04a770fd1\" (UID: \"9c036120-5481-45e6-8575-5df04a770fd1\") " Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.126595 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgng8\" (UniqueName: \"kubernetes.io/projected/9c036120-5481-45e6-8575-5df04a770fd1-kube-api-access-sgng8\") pod \"9c036120-5481-45e6-8575-5df04a770fd1\" (UID: \"9c036120-5481-45e6-8575-5df04a770fd1\") " Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.126685 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c036120-5481-45e6-8575-5df04a770fd1-ssh-key\") pod \"9c036120-5481-45e6-8575-5df04a770fd1\" (UID: \"9c036120-5481-45e6-8575-5df04a770fd1\") " Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.139412 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c036120-5481-45e6-8575-5df04a770fd1-kube-api-access-sgng8" (OuterVolumeSpecName: "kube-api-access-sgng8") pod "9c036120-5481-45e6-8575-5df04a770fd1" (UID: "9c036120-5481-45e6-8575-5df04a770fd1"). InnerVolumeSpecName "kube-api-access-sgng8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.160514 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c036120-5481-45e6-8575-5df04a770fd1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9c036120-5481-45e6-8575-5df04a770fd1" (UID: "9c036120-5481-45e6-8575-5df04a770fd1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.173968 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c036120-5481-45e6-8575-5df04a770fd1-inventory" (OuterVolumeSpecName: "inventory") pod "9c036120-5481-45e6-8575-5df04a770fd1" (UID: "9c036120-5481-45e6-8575-5df04a770fd1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.231355 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c036120-5481-45e6-8575-5df04a770fd1-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.231390 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgng8\" (UniqueName: \"kubernetes.io/projected/9c036120-5481-45e6-8575-5df04a770fd1-kube-api-access-sgng8\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.231401 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c036120-5481-45e6-8575-5df04a770fd1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.640845 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-spxhg" event={"ID":"9c036120-5481-45e6-8575-5df04a770fd1","Type":"ContainerDied","Data":"6a3b25be823055e067912b584459fbb354ef346222256a241808e921c348f914"} Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.640892 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a3b25be823055e067912b584459fbb354ef346222256a241808e921c348f914" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.640895 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-spxhg" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.712609 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-networker-ghd89"] Dec 05 09:24:08 crc kubenswrapper[4997]: E1205 09:24:08.713151 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c036120-5481-45e6-8575-5df04a770fd1" containerName="run-os-openstack-openstack-networker" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.713168 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c036120-5481-45e6-8575-5df04a770fd1" containerName="run-os-openstack-openstack-networker" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.713440 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c036120-5481-45e6-8575-5df04a770fd1" containerName="run-os-openstack-openstack-networker" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.714351 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-ghd89" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.716754 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-4542r" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.717152 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.725309 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-networker-ghd89"] Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.741123 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a7a8d6a-c2df-4df5-8079-695c04523bf9-inventory\") pod \"reboot-os-openstack-openstack-networker-ghd89\" (UID: \"8a7a8d6a-c2df-4df5-8079-695c04523bf9\") " pod="openstack/reboot-os-openstack-openstack-networker-ghd89" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.741963 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swllq\" (UniqueName: \"kubernetes.io/projected/8a7a8d6a-c2df-4df5-8079-695c04523bf9-kube-api-access-swllq\") pod \"reboot-os-openstack-openstack-networker-ghd89\" (UID: \"8a7a8d6a-c2df-4df5-8079-695c04523bf9\") " pod="openstack/reboot-os-openstack-openstack-networker-ghd89" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.742232 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a7a8d6a-c2df-4df5-8079-695c04523bf9-ssh-key\") pod \"reboot-os-openstack-openstack-networker-ghd89\" (UID: \"8a7a8d6a-c2df-4df5-8079-695c04523bf9\") " pod="openstack/reboot-os-openstack-openstack-networker-ghd89" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.844709 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a7a8d6a-c2df-4df5-8079-695c04523bf9-inventory\") pod \"reboot-os-openstack-openstack-networker-ghd89\" (UID: \"8a7a8d6a-c2df-4df5-8079-695c04523bf9\") " pod="openstack/reboot-os-openstack-openstack-networker-ghd89" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.844838 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swllq\" (UniqueName: \"kubernetes.io/projected/8a7a8d6a-c2df-4df5-8079-695c04523bf9-kube-api-access-swllq\") pod \"reboot-os-openstack-openstack-networker-ghd89\" (UID: \"8a7a8d6a-c2df-4df5-8079-695c04523bf9\") " pod="openstack/reboot-os-openstack-openstack-networker-ghd89" Dec 05 09:24:08 crc kubenswrapper[4997]: I1205 09:24:08.844941 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a7a8d6a-c2df-4df5-8079-695c04523bf9-ssh-key\") pod \"reboot-os-openstack-openstack-networker-ghd89\" (UID: \"8a7a8d6a-c2df-4df5-8079-695c04523bf9\") " pod="openstack/reboot-os-openstack-openstack-networker-ghd89" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.103206 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a7a8d6a-c2df-4df5-8079-695c04523bf9-ssh-key\") pod \"reboot-os-openstack-openstack-networker-ghd89\" (UID: \"8a7a8d6a-c2df-4df5-8079-695c04523bf9\") " pod="openstack/reboot-os-openstack-openstack-networker-ghd89" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.103251 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a7a8d6a-c2df-4df5-8079-695c04523bf9-inventory\") pod \"reboot-os-openstack-openstack-networker-ghd89\" (UID: \"8a7a8d6a-c2df-4df5-8079-695c04523bf9\") " pod="openstack/reboot-os-openstack-openstack-networker-ghd89" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.103822 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swllq\" (UniqueName: \"kubernetes.io/projected/8a7a8d6a-c2df-4df5-8079-695c04523bf9-kube-api-access-swllq\") pod \"reboot-os-openstack-openstack-networker-ghd89\" (UID: \"8a7a8d6a-c2df-4df5-8079-695c04523bf9\") " pod="openstack/reboot-os-openstack-openstack-networker-ghd89" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.209216 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.251608 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-ceph\") pod \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\" (UID: \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\") " Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.251698 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-inventory\") pod \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\" (UID: \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\") " Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.251772 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wk7s\" (UniqueName: \"kubernetes.io/projected/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-kube-api-access-8wk7s\") pod \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\" (UID: \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\") " Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.251926 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-ssh-key\") pod \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\" (UID: \"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4\") " Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.257421 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-ceph" (OuterVolumeSpecName: "ceph") pod "cfe38bed-b2b8-4d90-b74b-d82cea4b63a4" (UID: "cfe38bed-b2b8-4d90-b74b-d82cea4b63a4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.258498 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-kube-api-access-8wk7s" (OuterVolumeSpecName: "kube-api-access-8wk7s") pod "cfe38bed-b2b8-4d90-b74b-d82cea4b63a4" (UID: "cfe38bed-b2b8-4d90-b74b-d82cea4b63a4"). InnerVolumeSpecName "kube-api-access-8wk7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.285342 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-inventory" (OuterVolumeSpecName: "inventory") pod "cfe38bed-b2b8-4d90-b74b-d82cea4b63a4" (UID: "cfe38bed-b2b8-4d90-b74b-d82cea4b63a4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.290286 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cfe38bed-b2b8-4d90-b74b-d82cea4b63a4" (UID: "cfe38bed-b2b8-4d90-b74b-d82cea4b63a4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.336421 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-ghd89" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.354503 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.354551 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wk7s\" (UniqueName: \"kubernetes.io/projected/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-kube-api-access-8wk7s\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.354565 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.354578 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cfe38bed-b2b8-4d90-b74b-d82cea4b63a4-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.652631 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.652814 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-jsbwt" event={"ID":"cfe38bed-b2b8-4d90-b74b-d82cea4b63a4","Type":"ContainerDied","Data":"0d585084a3626ee981933050f86fc8ccae6fb1878959d3922a3418535f63d995"} Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.653098 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d585084a3626ee981933050f86fc8ccae6fb1878959d3922a3418535f63d995" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.725319 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-26zr2"] Dec 05 09:24:09 crc kubenswrapper[4997]: E1205 09:24:09.726084 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe38bed-b2b8-4d90-b74b-d82cea4b63a4" containerName="configure-os-openstack-openstack-cell1" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.726104 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe38bed-b2b8-4d90-b74b-d82cea4b63a4" containerName="configure-os-openstack-openstack-cell1" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.726326 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfe38bed-b2b8-4d90-b74b-d82cea4b63a4" containerName="configure-os-openstack-openstack-cell1" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.729355 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.732335 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.732766 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.763687 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-inventory-1\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.763735 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gckbd\" (UniqueName: \"kubernetes.io/projected/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-kube-api-access-gckbd\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.763915 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ssh-key-openstack-networker\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.763951 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.764069 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ceph\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.764300 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-inventory-0\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.773592 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-26zr2"] Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.866484 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-inventory-1\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.866555 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gckbd\" (UniqueName: \"kubernetes.io/projected/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-kube-api-access-gckbd\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.866642 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ssh-key-openstack-networker\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.866667 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.866725 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ceph\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.866778 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-inventory-0\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.879736 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ssh-key-openstack-networker\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.880399 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.882454 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-inventory-0\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.883519 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ceph\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.884009 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-inventory-1\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.897976 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-networker-ghd89"] Dec 05 09:24:09 crc kubenswrapper[4997]: I1205 09:24:09.927601 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gckbd\" (UniqueName: \"kubernetes.io/projected/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-kube-api-access-gckbd\") pod \"ssh-known-hosts-openstack-26zr2\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:10 crc kubenswrapper[4997]: I1205 09:24:10.049335 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:10 crc kubenswrapper[4997]: I1205 09:24:10.615854 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-26zr2"] Dec 05 09:24:10 crc kubenswrapper[4997]: I1205 09:24:10.677886 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-ghd89" event={"ID":"8a7a8d6a-c2df-4df5-8079-695c04523bf9","Type":"ContainerStarted","Data":"b809386bd786d610c543709d4b65a5aa101bc464791ba2ae33641aac0dedce56"} Dec 05 09:24:10 crc kubenswrapper[4997]: I1205 09:24:10.677937 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-ghd89" event={"ID":"8a7a8d6a-c2df-4df5-8079-695c04523bf9","Type":"ContainerStarted","Data":"1770682fadcd036d5554cae294772bf18d1574387f9e1a9662510963fd54e0c3"} Dec 05 09:24:10 crc kubenswrapper[4997]: I1205 09:24:10.680306 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-26zr2" event={"ID":"1fc2d23b-1a50-4016-81ab-9df0a6c8d021","Type":"ContainerStarted","Data":"35d87e6a2991d00f32ed03cd03ce350859e390da221447979f56a90e59acbcde"} Dec 05 09:24:10 crc kubenswrapper[4997]: I1205 09:24:10.705687 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-networker-ghd89" podStartSLOduration=2.2813365660000002 podStartE2EDuration="2.705662983s" podCreationTimestamp="2025-12-05 09:24:08 +0000 UTC" firstStartedPulling="2025-12-05 09:24:09.920462562 +0000 UTC m=+8950.449369823" lastFinishedPulling="2025-12-05 09:24:10.344788979 +0000 UTC m=+8950.873696240" observedRunningTime="2025-12-05 09:24:10.695378786 +0000 UTC m=+8951.224286057" watchObservedRunningTime="2025-12-05 09:24:10.705662983 +0000 UTC m=+8951.234570264" Dec 05 09:24:11 crc kubenswrapper[4997]: I1205 09:24:11.711684 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-26zr2" event={"ID":"1fc2d23b-1a50-4016-81ab-9df0a6c8d021","Type":"ContainerStarted","Data":"e1d4434556ac8ad50671a451dc27caa2d029e90f5a2fc8db87b30f3fb24c3807"} Dec 05 09:24:11 crc kubenswrapper[4997]: I1205 09:24:11.738333 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-26zr2" podStartSLOduration=2.334178514 podStartE2EDuration="2.738314497s" podCreationTimestamp="2025-12-05 09:24:09 +0000 UTC" firstStartedPulling="2025-12-05 09:24:10.623155255 +0000 UTC m=+8951.152062516" lastFinishedPulling="2025-12-05 09:24:11.027291228 +0000 UTC m=+8951.556198499" observedRunningTime="2025-12-05 09:24:11.729553839 +0000 UTC m=+8952.258461110" watchObservedRunningTime="2025-12-05 09:24:11.738314497 +0000 UTC m=+8952.267221758" Dec 05 09:24:14 crc kubenswrapper[4997]: I1205 09:24:14.750285 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:24:14 crc kubenswrapper[4997]: E1205 09:24:14.751141 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:24:24 crc kubenswrapper[4997]: I1205 09:24:24.840381 4997 generic.go:334] "Generic (PLEG): container finished" podID="8a7a8d6a-c2df-4df5-8079-695c04523bf9" containerID="b809386bd786d610c543709d4b65a5aa101bc464791ba2ae33641aac0dedce56" exitCode=0 Dec 05 09:24:24 crc kubenswrapper[4997]: I1205 09:24:24.840443 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-ghd89" event={"ID":"8a7a8d6a-c2df-4df5-8079-695c04523bf9","Type":"ContainerDied","Data":"b809386bd786d610c543709d4b65a5aa101bc464791ba2ae33641aac0dedce56"} Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.280480 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-ghd89" Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.423508 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a7a8d6a-c2df-4df5-8079-695c04523bf9-ssh-key\") pod \"8a7a8d6a-c2df-4df5-8079-695c04523bf9\" (UID: \"8a7a8d6a-c2df-4df5-8079-695c04523bf9\") " Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.423609 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swllq\" (UniqueName: \"kubernetes.io/projected/8a7a8d6a-c2df-4df5-8079-695c04523bf9-kube-api-access-swllq\") pod \"8a7a8d6a-c2df-4df5-8079-695c04523bf9\" (UID: \"8a7a8d6a-c2df-4df5-8079-695c04523bf9\") " Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.423859 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a7a8d6a-c2df-4df5-8079-695c04523bf9-inventory\") pod \"8a7a8d6a-c2df-4df5-8079-695c04523bf9\" (UID: \"8a7a8d6a-c2df-4df5-8079-695c04523bf9\") " Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.430451 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a7a8d6a-c2df-4df5-8079-695c04523bf9-kube-api-access-swllq" (OuterVolumeSpecName: "kube-api-access-swllq") pod "8a7a8d6a-c2df-4df5-8079-695c04523bf9" (UID: "8a7a8d6a-c2df-4df5-8079-695c04523bf9"). InnerVolumeSpecName "kube-api-access-swllq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.451716 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a7a8d6a-c2df-4df5-8079-695c04523bf9-inventory" (OuterVolumeSpecName: "inventory") pod "8a7a8d6a-c2df-4df5-8079-695c04523bf9" (UID: "8a7a8d6a-c2df-4df5-8079-695c04523bf9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.451775 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a7a8d6a-c2df-4df5-8079-695c04523bf9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8a7a8d6a-c2df-4df5-8079-695c04523bf9" (UID: "8a7a8d6a-c2df-4df5-8079-695c04523bf9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.528063 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a7a8d6a-c2df-4df5-8079-695c04523bf9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.528118 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swllq\" (UniqueName: \"kubernetes.io/projected/8a7a8d6a-c2df-4df5-8079-695c04523bf9-kube-api-access-swllq\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.528141 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a7a8d6a-c2df-4df5-8079-695c04523bf9-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.867104 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-ghd89" event={"ID":"8a7a8d6a-c2df-4df5-8079-695c04523bf9","Type":"ContainerDied","Data":"1770682fadcd036d5554cae294772bf18d1574387f9e1a9662510963fd54e0c3"} Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.867146 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1770682fadcd036d5554cae294772bf18d1574387f9e1a9662510963fd54e0c3" Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.867145 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-ghd89" Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.869294 4997 generic.go:334] "Generic (PLEG): container finished" podID="1fc2d23b-1a50-4016-81ab-9df0a6c8d021" containerID="e1d4434556ac8ad50671a451dc27caa2d029e90f5a2fc8db87b30f3fb24c3807" exitCode=0 Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.869328 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-26zr2" event={"ID":"1fc2d23b-1a50-4016-81ab-9df0a6c8d021","Type":"ContainerDied","Data":"e1d4434556ac8ad50671a451dc27caa2d029e90f5a2fc8db87b30f3fb24c3807"} Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.955870 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-networker-zzx85"] Dec 05 09:24:26 crc kubenswrapper[4997]: E1205 09:24:26.956289 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a7a8d6a-c2df-4df5-8079-695c04523bf9" containerName="reboot-os-openstack-openstack-networker" Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.956306 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a7a8d6a-c2df-4df5-8079-695c04523bf9" containerName="reboot-os-openstack-openstack-networker" Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.956523 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a7a8d6a-c2df-4df5-8079-695c04523bf9" containerName="reboot-os-openstack-openstack-networker" Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.957280 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.959350 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-4542r" Dec 05 09:24:26 crc kubenswrapper[4997]: I1205 09:24:26.972639 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-networker-zzx85"] Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.139357 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt8jr\" (UniqueName: \"kubernetes.io/projected/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-kube-api-access-rt8jr\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.139412 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-inventory\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.139443 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-ssh-key\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.139471 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.139714 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.139957 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.241886 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-inventory\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.241956 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-ssh-key\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.241989 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.242071 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.242180 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.242336 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt8jr\" (UniqueName: \"kubernetes.io/projected/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-kube-api-access-rt8jr\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.248394 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.248602 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.248760 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.250815 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-inventory\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.255550 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-ssh-key\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.262750 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt8jr\" (UniqueName: \"kubernetes.io/projected/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-kube-api-access-rt8jr\") pod \"install-certs-openstack-openstack-networker-zzx85\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:27 crc kubenswrapper[4997]: I1205 09:24:27.274977 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:28 crc kubenswrapper[4997]: I1205 09:24:28.590818 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-networker-zzx85"] Dec 05 09:24:28 crc kubenswrapper[4997]: I1205 09:24:28.748670 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:24:28 crc kubenswrapper[4997]: E1205 09:24:28.749114 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:24:28 crc kubenswrapper[4997]: I1205 09:24:28.887767 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-zzx85" event={"ID":"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e","Type":"ContainerStarted","Data":"ce54a0e99c40e406dde02483e6119fccd95a8e6676a466741fe609895afa4060"} Dec 05 09:24:28 crc kubenswrapper[4997]: I1205 09:24:28.889309 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-26zr2" event={"ID":"1fc2d23b-1a50-4016-81ab-9df0a6c8d021","Type":"ContainerDied","Data":"35d87e6a2991d00f32ed03cd03ce350859e390da221447979f56a90e59acbcde"} Dec 05 09:24:28 crc kubenswrapper[4997]: I1205 09:24:28.889343 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35d87e6a2991d00f32ed03cd03ce350859e390da221447979f56a90e59acbcde" Dec 05 09:24:28 crc kubenswrapper[4997]: I1205 09:24:28.917784 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.094156 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ceph\") pod \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.094232 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ssh-key-openstack-cell1\") pod \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.094262 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gckbd\" (UniqueName: \"kubernetes.io/projected/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-kube-api-access-gckbd\") pod \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.094355 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ssh-key-openstack-networker\") pod \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.094387 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-inventory-1\") pod \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.094533 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-inventory-0\") pod \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\" (UID: \"1fc2d23b-1a50-4016-81ab-9df0a6c8d021\") " Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.101507 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ceph" (OuterVolumeSpecName: "ceph") pod "1fc2d23b-1a50-4016-81ab-9df0a6c8d021" (UID: "1fc2d23b-1a50-4016-81ab-9df0a6c8d021"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.101758 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-kube-api-access-gckbd" (OuterVolumeSpecName: "kube-api-access-gckbd") pod "1fc2d23b-1a50-4016-81ab-9df0a6c8d021" (UID: "1fc2d23b-1a50-4016-81ab-9df0a6c8d021"). InnerVolumeSpecName "kube-api-access-gckbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.125728 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "1fc2d23b-1a50-4016-81ab-9df0a6c8d021" (UID: "1fc2d23b-1a50-4016-81ab-9df0a6c8d021"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.128756 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "1fc2d23b-1a50-4016-81ab-9df0a6c8d021" (UID: "1fc2d23b-1a50-4016-81ab-9df0a6c8d021"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.136395 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-inventory-1" (OuterVolumeSpecName: "inventory-1") pod "1fc2d23b-1a50-4016-81ab-9df0a6c8d021" (UID: "1fc2d23b-1a50-4016-81ab-9df0a6c8d021"). InnerVolumeSpecName "inventory-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.137011 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "1fc2d23b-1a50-4016-81ab-9df0a6c8d021" (UID: "1fc2d23b-1a50-4016-81ab-9df0a6c8d021"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.197858 4997 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.198166 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.198182 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.198196 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gckbd\" (UniqueName: \"kubernetes.io/projected/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-kube-api-access-gckbd\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.198205 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.198216 4997 reconciler_common.go:293] "Volume detached for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/1fc2d23b-1a50-4016-81ab-9df0a6c8d021-inventory-1\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.901597 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-zzx85" event={"ID":"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e","Type":"ContainerStarted","Data":"ec5d2c81c21c72c6488a2230059ab910c7d29efa07ab40f04989de790fa7d5a0"} Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.901658 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-26zr2" Dec 05 09:24:29 crc kubenswrapper[4997]: I1205 09:24:29.929325 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-networker-zzx85" podStartSLOduration=3.503723943 podStartE2EDuration="3.929291625s" podCreationTimestamp="2025-12-05 09:24:26 +0000 UTC" firstStartedPulling="2025-12-05 09:24:28.828039319 +0000 UTC m=+8969.356946580" lastFinishedPulling="2025-12-05 09:24:29.253607001 +0000 UTC m=+8969.782514262" observedRunningTime="2025-12-05 09:24:29.919495301 +0000 UTC m=+8970.448402582" watchObservedRunningTime="2025-12-05 09:24:29.929291625 +0000 UTC m=+8970.458198906" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.018311 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-x8tvh"] Dec 05 09:24:30 crc kubenswrapper[4997]: E1205 09:24:30.019171 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fc2d23b-1a50-4016-81ab-9df0a6c8d021" containerName="ssh-known-hosts-openstack" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.019199 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fc2d23b-1a50-4016-81ab-9df0a6c8d021" containerName="ssh-known-hosts-openstack" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.019481 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fc2d23b-1a50-4016-81ab-9df0a6c8d021" containerName="ssh-known-hosts-openstack" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.020844 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-x8tvh" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.024736 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.024976 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.045071 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-x8tvh"] Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.118901 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-ceph\") pod \"run-os-openstack-openstack-cell1-x8tvh\" (UID: \"fa02f03a-593b-447a-a94b-99484b6cd084\") " pod="openstack/run-os-openstack-openstack-cell1-x8tvh" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.119170 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kmsp\" (UniqueName: \"kubernetes.io/projected/fa02f03a-593b-447a-a94b-99484b6cd084-kube-api-access-9kmsp\") pod \"run-os-openstack-openstack-cell1-x8tvh\" (UID: \"fa02f03a-593b-447a-a94b-99484b6cd084\") " pod="openstack/run-os-openstack-openstack-cell1-x8tvh" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.119254 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-inventory\") pod \"run-os-openstack-openstack-cell1-x8tvh\" (UID: \"fa02f03a-593b-447a-a94b-99484b6cd084\") " pod="openstack/run-os-openstack-openstack-cell1-x8tvh" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.119288 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-ssh-key\") pod \"run-os-openstack-openstack-cell1-x8tvh\" (UID: \"fa02f03a-593b-447a-a94b-99484b6cd084\") " pod="openstack/run-os-openstack-openstack-cell1-x8tvh" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.221982 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kmsp\" (UniqueName: \"kubernetes.io/projected/fa02f03a-593b-447a-a94b-99484b6cd084-kube-api-access-9kmsp\") pod \"run-os-openstack-openstack-cell1-x8tvh\" (UID: \"fa02f03a-593b-447a-a94b-99484b6cd084\") " pod="openstack/run-os-openstack-openstack-cell1-x8tvh" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.222158 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-inventory\") pod \"run-os-openstack-openstack-cell1-x8tvh\" (UID: \"fa02f03a-593b-447a-a94b-99484b6cd084\") " pod="openstack/run-os-openstack-openstack-cell1-x8tvh" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.222232 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-ssh-key\") pod \"run-os-openstack-openstack-cell1-x8tvh\" (UID: \"fa02f03a-593b-447a-a94b-99484b6cd084\") " pod="openstack/run-os-openstack-openstack-cell1-x8tvh" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.222313 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-ceph\") pod \"run-os-openstack-openstack-cell1-x8tvh\" (UID: \"fa02f03a-593b-447a-a94b-99484b6cd084\") " pod="openstack/run-os-openstack-openstack-cell1-x8tvh" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.703535 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-inventory\") pod \"run-os-openstack-openstack-cell1-x8tvh\" (UID: \"fa02f03a-593b-447a-a94b-99484b6cd084\") " pod="openstack/run-os-openstack-openstack-cell1-x8tvh" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.704061 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-ssh-key\") pod \"run-os-openstack-openstack-cell1-x8tvh\" (UID: \"fa02f03a-593b-447a-a94b-99484b6cd084\") " pod="openstack/run-os-openstack-openstack-cell1-x8tvh" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.709531 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-ceph\") pod \"run-os-openstack-openstack-cell1-x8tvh\" (UID: \"fa02f03a-593b-447a-a94b-99484b6cd084\") " pod="openstack/run-os-openstack-openstack-cell1-x8tvh" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.710348 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kmsp\" (UniqueName: \"kubernetes.io/projected/fa02f03a-593b-447a-a94b-99484b6cd084-kube-api-access-9kmsp\") pod \"run-os-openstack-openstack-cell1-x8tvh\" (UID: \"fa02f03a-593b-447a-a94b-99484b6cd084\") " pod="openstack/run-os-openstack-openstack-cell1-x8tvh" Dec 05 09:24:30 crc kubenswrapper[4997]: I1205 09:24:30.954126 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-x8tvh" Dec 05 09:24:31 crc kubenswrapper[4997]: W1205 09:24:31.487037 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa02f03a_593b_447a_a94b_99484b6cd084.slice/crio-1db4426ac82c55978ab1925f80d648eba99bcd0983c6effa7cd65791a04ef206 WatchSource:0}: Error finding container 1db4426ac82c55978ab1925f80d648eba99bcd0983c6effa7cd65791a04ef206: Status 404 returned error can't find the container with id 1db4426ac82c55978ab1925f80d648eba99bcd0983c6effa7cd65791a04ef206 Dec 05 09:24:31 crc kubenswrapper[4997]: I1205 09:24:31.493727 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-x8tvh"] Dec 05 09:24:31 crc kubenswrapper[4997]: I1205 09:24:31.920412 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-x8tvh" event={"ID":"fa02f03a-593b-447a-a94b-99484b6cd084","Type":"ContainerStarted","Data":"1db4426ac82c55978ab1925f80d648eba99bcd0983c6effa7cd65791a04ef206"} Dec 05 09:24:32 crc kubenswrapper[4997]: I1205 09:24:32.931364 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-x8tvh" event={"ID":"fa02f03a-593b-447a-a94b-99484b6cd084","Type":"ContainerStarted","Data":"62ffed70b05acb67eee6d7cc31cc5c4a7e80a8849ce889425f80ffe7c04c461b"} Dec 05 09:24:32 crc kubenswrapper[4997]: I1205 09:24:32.950464 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-x8tvh" podStartSLOduration=3.529937297 podStartE2EDuration="3.950443011s" podCreationTimestamp="2025-12-05 09:24:29 +0000 UTC" firstStartedPulling="2025-12-05 09:24:31.490749026 +0000 UTC m=+8972.019656287" lastFinishedPulling="2025-12-05 09:24:31.91125472 +0000 UTC m=+8972.440162001" observedRunningTime="2025-12-05 09:24:32.948450916 +0000 UTC m=+8973.477358177" watchObservedRunningTime="2025-12-05 09:24:32.950443011 +0000 UTC m=+8973.479350272" Dec 05 09:24:39 crc kubenswrapper[4997]: I1205 09:24:39.755934 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:24:39 crc kubenswrapper[4997]: E1205 09:24:39.756750 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:24:41 crc kubenswrapper[4997]: I1205 09:24:41.017129 4997 generic.go:334] "Generic (PLEG): container finished" podID="5e5fc186-cf3d-444f-9f4d-3dc494faaa7e" containerID="ec5d2c81c21c72c6488a2230059ab910c7d29efa07ab40f04989de790fa7d5a0" exitCode=0 Dec 05 09:24:41 crc kubenswrapper[4997]: I1205 09:24:41.017190 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-zzx85" event={"ID":"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e","Type":"ContainerDied","Data":"ec5d2c81c21c72c6488a2230059ab910c7d29efa07ab40f04989de790fa7d5a0"} Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.464797 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.602217 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-inventory\") pod \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.602288 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-ssh-key\") pod \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.602427 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-neutron-metadata-combined-ca-bundle\") pod \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.602448 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-bootstrap-combined-ca-bundle\") pod \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.602671 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-ovn-combined-ca-bundle\") pod \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.602693 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt8jr\" (UniqueName: \"kubernetes.io/projected/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-kube-api-access-rt8jr\") pod \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\" (UID: \"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e\") " Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.608604 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "5e5fc186-cf3d-444f-9f4d-3dc494faaa7e" (UID: "5e5fc186-cf3d-444f-9f4d-3dc494faaa7e"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.608938 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "5e5fc186-cf3d-444f-9f4d-3dc494faaa7e" (UID: "5e5fc186-cf3d-444f-9f4d-3dc494faaa7e"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.625547 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "5e5fc186-cf3d-444f-9f4d-3dc494faaa7e" (UID: "5e5fc186-cf3d-444f-9f4d-3dc494faaa7e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.629472 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-kube-api-access-rt8jr" (OuterVolumeSpecName: "kube-api-access-rt8jr") pod "5e5fc186-cf3d-444f-9f4d-3dc494faaa7e" (UID: "5e5fc186-cf3d-444f-9f4d-3dc494faaa7e"). InnerVolumeSpecName "kube-api-access-rt8jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.635195 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-inventory" (OuterVolumeSpecName: "inventory") pod "5e5fc186-cf3d-444f-9f4d-3dc494faaa7e" (UID: "5e5fc186-cf3d-444f-9f4d-3dc494faaa7e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.647781 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5e5fc186-cf3d-444f-9f4d-3dc494faaa7e" (UID: "5e5fc186-cf3d-444f-9f4d-3dc494faaa7e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.705851 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.705882 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.705892 4997 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.705902 4997 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.705915 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt8jr\" (UniqueName: \"kubernetes.io/projected/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-kube-api-access-rt8jr\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:42 crc kubenswrapper[4997]: I1205 09:24:42.705924 4997 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5fc186-cf3d-444f-9f4d-3dc494faaa7e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.038411 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-zzx85" event={"ID":"5e5fc186-cf3d-444f-9f4d-3dc494faaa7e","Type":"ContainerDied","Data":"ce54a0e99c40e406dde02483e6119fccd95a8e6676a466741fe609895afa4060"} Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.038710 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce54a0e99c40e406dde02483e6119fccd95a8e6676a466741fe609895afa4060" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.038584 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-zzx85" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.111046 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-networker-9mbvl"] Dec 05 09:24:43 crc kubenswrapper[4997]: E1205 09:24:43.111848 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5fc186-cf3d-444f-9f4d-3dc494faaa7e" containerName="install-certs-openstack-openstack-networker" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.111878 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5fc186-cf3d-444f-9f4d-3dc494faaa7e" containerName="install-certs-openstack-openstack-networker" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.112227 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5fc186-cf3d-444f-9f4d-3dc494faaa7e" containerName="install-certs-openstack-openstack-networker" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.113375 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.118242 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-4542r" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.118443 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.118744 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.130035 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-networker-9mbvl"] Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.216334 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/51661db5-879f-4b6c-a563-ff1c84e47f0d-ovncontroller-config-0\") pod \"ovn-openstack-openstack-networker-9mbvl\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.216592 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-ssh-key\") pod \"ovn-openstack-openstack-networker-9mbvl\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.217041 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l64f9\" (UniqueName: \"kubernetes.io/projected/51661db5-879f-4b6c-a563-ff1c84e47f0d-kube-api-access-l64f9\") pod \"ovn-openstack-openstack-networker-9mbvl\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.217163 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-inventory\") pod \"ovn-openstack-openstack-networker-9mbvl\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.217223 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-networker-9mbvl\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.319410 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l64f9\" (UniqueName: \"kubernetes.io/projected/51661db5-879f-4b6c-a563-ff1c84e47f0d-kube-api-access-l64f9\") pod \"ovn-openstack-openstack-networker-9mbvl\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.319471 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-inventory\") pod \"ovn-openstack-openstack-networker-9mbvl\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.319508 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-networker-9mbvl\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.319584 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/51661db5-879f-4b6c-a563-ff1c84e47f0d-ovncontroller-config-0\") pod \"ovn-openstack-openstack-networker-9mbvl\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.320141 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-ssh-key\") pod \"ovn-openstack-openstack-networker-9mbvl\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.322165 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/51661db5-879f-4b6c-a563-ff1c84e47f0d-ovncontroller-config-0\") pod \"ovn-openstack-openstack-networker-9mbvl\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.324110 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-networker-9mbvl\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.324194 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-inventory\") pod \"ovn-openstack-openstack-networker-9mbvl\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.334274 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-ssh-key\") pod \"ovn-openstack-openstack-networker-9mbvl\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.338098 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l64f9\" (UniqueName: \"kubernetes.io/projected/51661db5-879f-4b6c-a563-ff1c84e47f0d-kube-api-access-l64f9\") pod \"ovn-openstack-openstack-networker-9mbvl\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:24:43 crc kubenswrapper[4997]: I1205 09:24:43.431196 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:24:44 crc kubenswrapper[4997]: I1205 09:24:44.048775 4997 generic.go:334] "Generic (PLEG): container finished" podID="fa02f03a-593b-447a-a94b-99484b6cd084" containerID="62ffed70b05acb67eee6d7cc31cc5c4a7e80a8849ce889425f80ffe7c04c461b" exitCode=0 Dec 05 09:24:44 crc kubenswrapper[4997]: I1205 09:24:44.048861 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-x8tvh" event={"ID":"fa02f03a-593b-447a-a94b-99484b6cd084","Type":"ContainerDied","Data":"62ffed70b05acb67eee6d7cc31cc5c4a7e80a8849ce889425f80ffe7c04c461b"} Dec 05 09:24:44 crc kubenswrapper[4997]: I1205 09:24:44.175051 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-networker-9mbvl"] Dec 05 09:24:45 crc kubenswrapper[4997]: I1205 09:24:45.059773 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-9mbvl" event={"ID":"51661db5-879f-4b6c-a563-ff1c84e47f0d","Type":"ContainerStarted","Data":"470e604c7fa4354aefd190c4b20aac805e6587b1e1bd51b423c8cbbc5e327562"} Dec 05 09:24:45 crc kubenswrapper[4997]: I1205 09:24:45.061394 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-9mbvl" event={"ID":"51661db5-879f-4b6c-a563-ff1c84e47f0d","Type":"ContainerStarted","Data":"b2e83ff9dd96f92b4edbba9039bff743c0b980760cea3f9a9020bb1069aeadf8"} Dec 05 09:24:45 crc kubenswrapper[4997]: I1205 09:24:45.099578 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-networker-9mbvl" podStartSLOduration=1.543565141 podStartE2EDuration="2.099531122s" podCreationTimestamp="2025-12-05 09:24:43 +0000 UTC" firstStartedPulling="2025-12-05 09:24:44.17765891 +0000 UTC m=+8984.706566171" lastFinishedPulling="2025-12-05 09:24:44.733624891 +0000 UTC m=+8985.262532152" observedRunningTime="2025-12-05 09:24:45.082959735 +0000 UTC m=+8985.611867016" watchObservedRunningTime="2025-12-05 09:24:45.099531122 +0000 UTC m=+8985.628438393" Dec 05 09:24:45 crc kubenswrapper[4997]: I1205 09:24:45.511116 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-x8tvh" Dec 05 09:24:45 crc kubenswrapper[4997]: I1205 09:24:45.668769 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-ceph\") pod \"fa02f03a-593b-447a-a94b-99484b6cd084\" (UID: \"fa02f03a-593b-447a-a94b-99484b6cd084\") " Dec 05 09:24:45 crc kubenswrapper[4997]: I1205 09:24:45.669147 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-ssh-key\") pod \"fa02f03a-593b-447a-a94b-99484b6cd084\" (UID: \"fa02f03a-593b-447a-a94b-99484b6cd084\") " Dec 05 09:24:45 crc kubenswrapper[4997]: I1205 09:24:45.670347 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kmsp\" (UniqueName: \"kubernetes.io/projected/fa02f03a-593b-447a-a94b-99484b6cd084-kube-api-access-9kmsp\") pod \"fa02f03a-593b-447a-a94b-99484b6cd084\" (UID: \"fa02f03a-593b-447a-a94b-99484b6cd084\") " Dec 05 09:24:45 crc kubenswrapper[4997]: I1205 09:24:45.670554 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-inventory\") pod \"fa02f03a-593b-447a-a94b-99484b6cd084\" (UID: \"fa02f03a-593b-447a-a94b-99484b6cd084\") " Dec 05 09:24:45 crc kubenswrapper[4997]: I1205 09:24:45.674092 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-ceph" (OuterVolumeSpecName: "ceph") pod "fa02f03a-593b-447a-a94b-99484b6cd084" (UID: "fa02f03a-593b-447a-a94b-99484b6cd084"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:45 crc kubenswrapper[4997]: I1205 09:24:45.674185 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa02f03a-593b-447a-a94b-99484b6cd084-kube-api-access-9kmsp" (OuterVolumeSpecName: "kube-api-access-9kmsp") pod "fa02f03a-593b-447a-a94b-99484b6cd084" (UID: "fa02f03a-593b-447a-a94b-99484b6cd084"). InnerVolumeSpecName "kube-api-access-9kmsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:24:45 crc kubenswrapper[4997]: I1205 09:24:45.698744 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fa02f03a-593b-447a-a94b-99484b6cd084" (UID: "fa02f03a-593b-447a-a94b-99484b6cd084"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:45 crc kubenswrapper[4997]: I1205 09:24:45.703221 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-inventory" (OuterVolumeSpecName: "inventory") pod "fa02f03a-593b-447a-a94b-99484b6cd084" (UID: "fa02f03a-593b-447a-a94b-99484b6cd084"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:24:45 crc kubenswrapper[4997]: I1205 09:24:45.772971 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kmsp\" (UniqueName: \"kubernetes.io/projected/fa02f03a-593b-447a-a94b-99484b6cd084-kube-api-access-9kmsp\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:45 crc kubenswrapper[4997]: I1205 09:24:45.773002 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:45 crc kubenswrapper[4997]: I1205 09:24:45.773011 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:45 crc kubenswrapper[4997]: I1205 09:24:45.773021 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa02f03a-593b-447a-a94b-99484b6cd084-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.072015 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-x8tvh" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.072047 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-x8tvh" event={"ID":"fa02f03a-593b-447a-a94b-99484b6cd084","Type":"ContainerDied","Data":"1db4426ac82c55978ab1925f80d648eba99bcd0983c6effa7cd65791a04ef206"} Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.073105 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1db4426ac82c55978ab1925f80d648eba99bcd0983c6effa7cd65791a04ef206" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.147592 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-bmkjw"] Dec 05 09:24:46 crc kubenswrapper[4997]: E1205 09:24:46.148490 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa02f03a-593b-447a-a94b-99484b6cd084" containerName="run-os-openstack-openstack-cell1" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.148513 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa02f03a-593b-447a-a94b-99484b6cd084" containerName="run-os-openstack-openstack-cell1" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.148836 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa02f03a-593b-447a-a94b-99484b6cd084" containerName="run-os-openstack-openstack-cell1" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.149799 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.152695 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.153264 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.159983 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-bmkjw"] Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.287519 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-inventory\") pod \"reboot-os-openstack-openstack-cell1-bmkjw\" (UID: \"6e5fc8d2-9421-49e7-9007-697047cddecd\") " pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.287863 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-ceph\") pod \"reboot-os-openstack-openstack-cell1-bmkjw\" (UID: \"6e5fc8d2-9421-49e7-9007-697047cddecd\") " pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.287914 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-bmkjw\" (UID: \"6e5fc8d2-9421-49e7-9007-697047cddecd\") " pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.287984 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rxzg\" (UniqueName: \"kubernetes.io/projected/6e5fc8d2-9421-49e7-9007-697047cddecd-kube-api-access-2rxzg\") pod \"reboot-os-openstack-openstack-cell1-bmkjw\" (UID: \"6e5fc8d2-9421-49e7-9007-697047cddecd\") " pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.390321 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rxzg\" (UniqueName: \"kubernetes.io/projected/6e5fc8d2-9421-49e7-9007-697047cddecd-kube-api-access-2rxzg\") pod \"reboot-os-openstack-openstack-cell1-bmkjw\" (UID: \"6e5fc8d2-9421-49e7-9007-697047cddecd\") " pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.390456 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-inventory\") pod \"reboot-os-openstack-openstack-cell1-bmkjw\" (UID: \"6e5fc8d2-9421-49e7-9007-697047cddecd\") " pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.390528 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-ceph\") pod \"reboot-os-openstack-openstack-cell1-bmkjw\" (UID: \"6e5fc8d2-9421-49e7-9007-697047cddecd\") " pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.390635 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-bmkjw\" (UID: \"6e5fc8d2-9421-49e7-9007-697047cddecd\") " pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.395296 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-bmkjw\" (UID: \"6e5fc8d2-9421-49e7-9007-697047cddecd\") " pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.395356 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-inventory\") pod \"reboot-os-openstack-openstack-cell1-bmkjw\" (UID: \"6e5fc8d2-9421-49e7-9007-697047cddecd\") " pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.395719 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-ceph\") pod \"reboot-os-openstack-openstack-cell1-bmkjw\" (UID: \"6e5fc8d2-9421-49e7-9007-697047cddecd\") " pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.410984 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rxzg\" (UniqueName: \"kubernetes.io/projected/6e5fc8d2-9421-49e7-9007-697047cddecd-kube-api-access-2rxzg\") pod \"reboot-os-openstack-openstack-cell1-bmkjw\" (UID: \"6e5fc8d2-9421-49e7-9007-697047cddecd\") " pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" Dec 05 09:24:46 crc kubenswrapper[4997]: I1205 09:24:46.489426 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" Dec 05 09:24:47 crc kubenswrapper[4997]: I1205 09:24:47.024126 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-bmkjw"] Dec 05 09:24:47 crc kubenswrapper[4997]: I1205 09:24:47.081929 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" event={"ID":"6e5fc8d2-9421-49e7-9007-697047cddecd","Type":"ContainerStarted","Data":"eb384c93609189005fa502ae61d0abac6238ac09c313e896bbf697e2a7dec681"} Dec 05 09:24:48 crc kubenswrapper[4997]: I1205 09:24:48.092197 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" event={"ID":"6e5fc8d2-9421-49e7-9007-697047cddecd","Type":"ContainerStarted","Data":"25e2109e753fad22130a1c840e2782385a3c9ab52d62fda3104c67a92034aef8"} Dec 05 09:24:48 crc kubenswrapper[4997]: I1205 09:24:48.114190 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" podStartSLOduration=1.5531480819999999 podStartE2EDuration="2.114172601s" podCreationTimestamp="2025-12-05 09:24:46 +0000 UTC" firstStartedPulling="2025-12-05 09:24:47.035117295 +0000 UTC m=+8987.564024556" lastFinishedPulling="2025-12-05 09:24:47.596141814 +0000 UTC m=+8988.125049075" observedRunningTime="2025-12-05 09:24:48.106624537 +0000 UTC m=+8988.635531788" watchObservedRunningTime="2025-12-05 09:24:48.114172601 +0000 UTC m=+8988.643079852" Dec 05 09:24:52 crc kubenswrapper[4997]: I1205 09:24:52.750923 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:24:52 crc kubenswrapper[4997]: E1205 09:24:52.751544 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:25:05 crc kubenswrapper[4997]: I1205 09:25:05.255028 4997 generic.go:334] "Generic (PLEG): container finished" podID="6e5fc8d2-9421-49e7-9007-697047cddecd" containerID="25e2109e753fad22130a1c840e2782385a3c9ab52d62fda3104c67a92034aef8" exitCode=0 Dec 05 09:25:05 crc kubenswrapper[4997]: I1205 09:25:05.255107 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" event={"ID":"6e5fc8d2-9421-49e7-9007-697047cddecd","Type":"ContainerDied","Data":"25e2109e753fad22130a1c840e2782385a3c9ab52d62fda3104c67a92034aef8"} Dec 05 09:25:06 crc kubenswrapper[4997]: I1205 09:25:06.726005 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" Dec 05 09:25:06 crc kubenswrapper[4997]: I1205 09:25:06.820606 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-ssh-key\") pod \"6e5fc8d2-9421-49e7-9007-697047cddecd\" (UID: \"6e5fc8d2-9421-49e7-9007-697047cddecd\") " Dec 05 09:25:06 crc kubenswrapper[4997]: I1205 09:25:06.820804 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-ceph\") pod \"6e5fc8d2-9421-49e7-9007-697047cddecd\" (UID: \"6e5fc8d2-9421-49e7-9007-697047cddecd\") " Dec 05 09:25:06 crc kubenswrapper[4997]: I1205 09:25:06.820983 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rxzg\" (UniqueName: \"kubernetes.io/projected/6e5fc8d2-9421-49e7-9007-697047cddecd-kube-api-access-2rxzg\") pod \"6e5fc8d2-9421-49e7-9007-697047cddecd\" (UID: \"6e5fc8d2-9421-49e7-9007-697047cddecd\") " Dec 05 09:25:06 crc kubenswrapper[4997]: I1205 09:25:06.821009 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-inventory\") pod \"6e5fc8d2-9421-49e7-9007-697047cddecd\" (UID: \"6e5fc8d2-9421-49e7-9007-697047cddecd\") " Dec 05 09:25:06 crc kubenswrapper[4997]: I1205 09:25:06.827748 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-ceph" (OuterVolumeSpecName: "ceph") pod "6e5fc8d2-9421-49e7-9007-697047cddecd" (UID: "6e5fc8d2-9421-49e7-9007-697047cddecd"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:25:06 crc kubenswrapper[4997]: I1205 09:25:06.827834 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e5fc8d2-9421-49e7-9007-697047cddecd-kube-api-access-2rxzg" (OuterVolumeSpecName: "kube-api-access-2rxzg") pod "6e5fc8d2-9421-49e7-9007-697047cddecd" (UID: "6e5fc8d2-9421-49e7-9007-697047cddecd"). InnerVolumeSpecName "kube-api-access-2rxzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:25:06 crc kubenswrapper[4997]: I1205 09:25:06.850521 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-inventory" (OuterVolumeSpecName: "inventory") pod "6e5fc8d2-9421-49e7-9007-697047cddecd" (UID: "6e5fc8d2-9421-49e7-9007-697047cddecd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:25:06 crc kubenswrapper[4997]: I1205 09:25:06.852058 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6e5fc8d2-9421-49e7-9007-697047cddecd" (UID: "6e5fc8d2-9421-49e7-9007-697047cddecd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:25:06 crc kubenswrapper[4997]: I1205 09:25:06.923870 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rxzg\" (UniqueName: \"kubernetes.io/projected/6e5fc8d2-9421-49e7-9007-697047cddecd-kube-api-access-2rxzg\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:06 crc kubenswrapper[4997]: I1205 09:25:06.924446 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:06 crc kubenswrapper[4997]: I1205 09:25:06.924523 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:06 crc kubenswrapper[4997]: I1205 09:25:06.924611 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6e5fc8d2-9421-49e7-9007-697047cddecd-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.279189 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" event={"ID":"6e5fc8d2-9421-49e7-9007-697047cddecd","Type":"ContainerDied","Data":"eb384c93609189005fa502ae61d0abac6238ac09c313e896bbf697e2a7dec681"} Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.279610 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb384c93609189005fa502ae61d0abac6238ac09c313e896bbf697e2a7dec681" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.279427 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-bmkjw" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.379550 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-2kzxz"] Dec 05 09:25:07 crc kubenswrapper[4997]: E1205 09:25:07.380091 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e5fc8d2-9421-49e7-9007-697047cddecd" containerName="reboot-os-openstack-openstack-cell1" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.380114 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e5fc8d2-9421-49e7-9007-697047cddecd" containerName="reboot-os-openstack-openstack-cell1" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.380373 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e5fc8d2-9421-49e7-9007-697047cddecd" containerName="reboot-os-openstack-openstack-cell1" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.381304 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.388810 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.392799 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.420750 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-2kzxz"] Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.434150 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.434203 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqwx9\" (UniqueName: \"kubernetes.io/projected/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-kube-api-access-gqwx9\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.434237 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.434271 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.434311 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-inventory\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.434364 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.434391 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ssh-key\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.434470 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.434521 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.434560 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ceph\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.434608 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.435700 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.538245 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.538310 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ceph\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.538356 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.538386 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.538482 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.538510 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqwx9\" (UniqueName: \"kubernetes.io/projected/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-kube-api-access-gqwx9\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.538539 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.538572 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.538628 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-inventory\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.538681 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.538726 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ssh-key\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.538791 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.545074 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.545292 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.545633 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.545654 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.546407 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-inventory\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.546850 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.548118 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ssh-key\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.548318 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.548823 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ceph\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.551380 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.552381 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.556481 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqwx9\" (UniqueName: \"kubernetes.io/projected/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-kube-api-access-gqwx9\") pod \"install-certs-openstack-openstack-cell1-2kzxz\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.711952 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:07 crc kubenswrapper[4997]: I1205 09:25:07.750022 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:25:07 crc kubenswrapper[4997]: E1205 09:25:07.750692 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:25:08 crc kubenswrapper[4997]: I1205 09:25:08.228125 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-2kzxz"] Dec 05 09:25:08 crc kubenswrapper[4997]: I1205 09:25:08.288736 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" event={"ID":"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57","Type":"ContainerStarted","Data":"3b90b261354c24f5e306d6e53c8cca1a3b63384de372228d754269f04b669ba9"} Dec 05 09:25:09 crc kubenswrapper[4997]: I1205 09:25:09.301216 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" event={"ID":"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57","Type":"ContainerStarted","Data":"14f8fe56370f8d28496a5958c6048f9ba636e7a4098d95c66b15c03f21d57759"} Dec 05 09:25:09 crc kubenswrapper[4997]: I1205 09:25:09.334011 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" podStartSLOduration=1.906704429 podStartE2EDuration="2.333990736s" podCreationTimestamp="2025-12-05 09:25:07 +0000 UTC" firstStartedPulling="2025-12-05 09:25:08.229809892 +0000 UTC m=+9008.758717153" lastFinishedPulling="2025-12-05 09:25:08.657096199 +0000 UTC m=+9009.186003460" observedRunningTime="2025-12-05 09:25:09.318987821 +0000 UTC m=+9009.847895102" watchObservedRunningTime="2025-12-05 09:25:09.333990736 +0000 UTC m=+9009.862897997" Dec 05 09:25:18 crc kubenswrapper[4997]: I1205 09:25:18.749276 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:25:18 crc kubenswrapper[4997]: E1205 09:25:18.750235 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:25:30 crc kubenswrapper[4997]: I1205 09:25:30.504589 4997 generic.go:334] "Generic (PLEG): container finished" podID="b47da422-d6ad-4a06-8f9d-f6e0d6c80d57" containerID="14f8fe56370f8d28496a5958c6048f9ba636e7a4098d95c66b15c03f21d57759" exitCode=0 Dec 05 09:25:30 crc kubenswrapper[4997]: I1205 09:25:30.504833 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" event={"ID":"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57","Type":"ContainerDied","Data":"14f8fe56370f8d28496a5958c6048f9ba636e7a4098d95c66b15c03f21d57759"} Dec 05 09:25:31 crc kubenswrapper[4997]: I1205 09:25:31.740953 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2w7hm"] Dec 05 09:25:31 crc kubenswrapper[4997]: I1205 09:25:31.743514 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2w7hm" Dec 05 09:25:31 crc kubenswrapper[4997]: I1205 09:25:31.750604 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:25:31 crc kubenswrapper[4997]: E1205 09:25:31.750943 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:25:31 crc kubenswrapper[4997]: I1205 09:25:31.773224 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2w7hm"] Dec 05 09:25:31 crc kubenswrapper[4997]: I1205 09:25:31.773904 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-utilities\") pod \"redhat-operators-2w7hm\" (UID: \"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a\") " pod="openshift-marketplace/redhat-operators-2w7hm" Dec 05 09:25:31 crc kubenswrapper[4997]: I1205 09:25:31.774148 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-catalog-content\") pod \"redhat-operators-2w7hm\" (UID: \"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a\") " pod="openshift-marketplace/redhat-operators-2w7hm" Dec 05 09:25:31 crc kubenswrapper[4997]: I1205 09:25:31.774254 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v87nv\" (UniqueName: \"kubernetes.io/projected/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-kube-api-access-v87nv\") pod \"redhat-operators-2w7hm\" (UID: \"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a\") " pod="openshift-marketplace/redhat-operators-2w7hm" Dec 05 09:25:31 crc kubenswrapper[4997]: I1205 09:25:31.877104 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-catalog-content\") pod \"redhat-operators-2w7hm\" (UID: \"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a\") " pod="openshift-marketplace/redhat-operators-2w7hm" Dec 05 09:25:31 crc kubenswrapper[4997]: I1205 09:25:31.877224 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v87nv\" (UniqueName: \"kubernetes.io/projected/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-kube-api-access-v87nv\") pod \"redhat-operators-2w7hm\" (UID: \"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a\") " pod="openshift-marketplace/redhat-operators-2w7hm" Dec 05 09:25:31 crc kubenswrapper[4997]: I1205 09:25:31.877305 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-utilities\") pod \"redhat-operators-2w7hm\" (UID: \"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a\") " pod="openshift-marketplace/redhat-operators-2w7hm" Dec 05 09:25:31 crc kubenswrapper[4997]: I1205 09:25:31.878559 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-catalog-content\") pod \"redhat-operators-2w7hm\" (UID: \"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a\") " pod="openshift-marketplace/redhat-operators-2w7hm" Dec 05 09:25:31 crc kubenswrapper[4997]: I1205 09:25:31.878918 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-utilities\") pod \"redhat-operators-2w7hm\" (UID: \"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a\") " pod="openshift-marketplace/redhat-operators-2w7hm" Dec 05 09:25:31 crc kubenswrapper[4997]: I1205 09:25:31.903002 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v87nv\" (UniqueName: \"kubernetes.io/projected/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-kube-api-access-v87nv\") pod \"redhat-operators-2w7hm\" (UID: \"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a\") " pod="openshift-marketplace/redhat-operators-2w7hm" Dec 05 09:25:31 crc kubenswrapper[4997]: I1205 09:25:31.997037 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.080594 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-libvirt-combined-ca-bundle\") pod \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.080722 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-sriov-combined-ca-bundle\") pod \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.080799 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqwx9\" (UniqueName: \"kubernetes.io/projected/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-kube-api-access-gqwx9\") pod \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.080829 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ceph\") pod \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.080856 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-inventory\") pod \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.080897 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ovn-combined-ca-bundle\") pod \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.080956 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-dhcp-combined-ca-bundle\") pod \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.080977 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ssh-key\") pod \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.083997 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-metadata-combined-ca-bundle\") pod \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.084080 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-telemetry-combined-ca-bundle\") pod \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.084131 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-nova-combined-ca-bundle\") pod \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.084170 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-bootstrap-combined-ca-bundle\") pod \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\" (UID: \"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57\") " Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.086945 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57" (UID: "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.087221 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2w7hm" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.087764 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ceph" (OuterVolumeSpecName: "ceph") pod "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57" (UID: "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.089538 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57" (UID: "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.089571 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57" (UID: "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.091018 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-kube-api-access-gqwx9" (OuterVolumeSpecName: "kube-api-access-gqwx9") pod "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57" (UID: "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57"). InnerVolumeSpecName "kube-api-access-gqwx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.095076 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57" (UID: "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.095745 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57" (UID: "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.096786 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57" (UID: "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.108416 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57" (UID: "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.109100 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57" (UID: "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.154786 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-inventory" (OuterVolumeSpecName: "inventory") pod "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57" (UID: "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.159137 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57" (UID: "b47da422-d6ad-4a06-8f9d-f6e0d6c80d57"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.187154 4997 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.187200 4997 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.187214 4997 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.187226 4997 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.187239 4997 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.187253 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqwx9\" (UniqueName: \"kubernetes.io/projected/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-kube-api-access-gqwx9\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.187268 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.187280 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.187290 4997 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.187302 4997 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.187314 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.187329 4997 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47da422-d6ad-4a06-8f9d-f6e0d6c80d57-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.526940 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" event={"ID":"b47da422-d6ad-4a06-8f9d-f6e0d6c80d57","Type":"ContainerDied","Data":"3b90b261354c24f5e306d6e53c8cca1a3b63384de372228d754269f04b669ba9"} Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.527326 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b90b261354c24f5e306d6e53c8cca1a3b63384de372228d754269f04b669ba9" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.527012 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-2kzxz" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.625541 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-cfqnz"] Dec 05 09:25:32 crc kubenswrapper[4997]: E1205 09:25:32.625990 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b47da422-d6ad-4a06-8f9d-f6e0d6c80d57" containerName="install-certs-openstack-openstack-cell1" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.626008 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b47da422-d6ad-4a06-8f9d-f6e0d6c80d57" containerName="install-certs-openstack-openstack-cell1" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.626240 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="b47da422-d6ad-4a06-8f9d-f6e0d6c80d57" containerName="install-certs-openstack-openstack-cell1" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.627038 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.630500 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.630763 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.650054 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-cfqnz"] Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.665286 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2w7hm"] Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.700998 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbxk9\" (UniqueName: \"kubernetes.io/projected/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-kube-api-access-jbxk9\") pod \"ceph-client-openstack-openstack-cell1-cfqnz\" (UID: \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\") " pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.701056 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-ceph\") pod \"ceph-client-openstack-openstack-cell1-cfqnz\" (UID: \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\") " pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.701108 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-inventory\") pod \"ceph-client-openstack-openstack-cell1-cfqnz\" (UID: \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\") " pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.701184 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-cfqnz\" (UID: \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\") " pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.804181 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbxk9\" (UniqueName: \"kubernetes.io/projected/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-kube-api-access-jbxk9\") pod \"ceph-client-openstack-openstack-cell1-cfqnz\" (UID: \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\") " pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.804848 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-ceph\") pod \"ceph-client-openstack-openstack-cell1-cfqnz\" (UID: \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\") " pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.804972 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-inventory\") pod \"ceph-client-openstack-openstack-cell1-cfqnz\" (UID: \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\") " pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.805110 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-cfqnz\" (UID: \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\") " pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.813840 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-inventory\") pod \"ceph-client-openstack-openstack-cell1-cfqnz\" (UID: \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\") " pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.816339 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-cfqnz\" (UID: \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\") " pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.818519 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-ceph\") pod \"ceph-client-openstack-openstack-cell1-cfqnz\" (UID: \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\") " pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" Dec 05 09:25:32 crc kubenswrapper[4997]: I1205 09:25:32.825202 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbxk9\" (UniqueName: \"kubernetes.io/projected/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-kube-api-access-jbxk9\") pod \"ceph-client-openstack-openstack-cell1-cfqnz\" (UID: \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\") " pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" Dec 05 09:25:33 crc kubenswrapper[4997]: I1205 09:25:33.005970 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" Dec 05 09:25:33 crc kubenswrapper[4997]: I1205 09:25:33.540888 4997 generic.go:334] "Generic (PLEG): container finished" podID="ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a" containerID="3d09f3187cca7690dbfa1af052f061e0383ac3f1d93011b385c265a1354e7f51" exitCode=0 Dec 05 09:25:33 crc kubenswrapper[4997]: I1205 09:25:33.540957 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2w7hm" event={"ID":"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a","Type":"ContainerDied","Data":"3d09f3187cca7690dbfa1af052f061e0383ac3f1d93011b385c265a1354e7f51"} Dec 05 09:25:33 crc kubenswrapper[4997]: I1205 09:25:33.541303 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2w7hm" event={"ID":"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a","Type":"ContainerStarted","Data":"9bcba13041912ceab1497c12e1fc0e36a2757fccb0fedebf30a1cd65b3f29290"} Dec 05 09:25:33 crc kubenswrapper[4997]: I1205 09:25:33.930798 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-cfqnz"] Dec 05 09:25:34 crc kubenswrapper[4997]: I1205 09:25:34.554057 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" event={"ID":"f8efbd91-b09a-4dac-bf7d-b0b0433023ff","Type":"ContainerStarted","Data":"37282f4ba885e6f18af79b7b770dfc26705e665efc15fdc0c9e3eb4f6ab439e2"} Dec 05 09:25:35 crc kubenswrapper[4997]: I1205 09:25:35.566535 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2w7hm" event={"ID":"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a","Type":"ContainerStarted","Data":"8bfa02d026ab3bbe68ef8d3ebbe6091122e663647f85ef256246151ef1f29237"} Dec 05 09:25:35 crc kubenswrapper[4997]: I1205 09:25:35.568532 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" event={"ID":"f8efbd91-b09a-4dac-bf7d-b0b0433023ff","Type":"ContainerStarted","Data":"73ba391f5a0b1beaa1a5a0301c3238cbb27d88151909d83d8123eaaa5285d61e"} Dec 05 09:25:35 crc kubenswrapper[4997]: I1205 09:25:35.612479 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" podStartSLOduration=3.199454147 podStartE2EDuration="3.612457739s" podCreationTimestamp="2025-12-05 09:25:32 +0000 UTC" firstStartedPulling="2025-12-05 09:25:33.9340801 +0000 UTC m=+9034.462987361" lastFinishedPulling="2025-12-05 09:25:34.347083692 +0000 UTC m=+9034.875990953" observedRunningTime="2025-12-05 09:25:35.611827761 +0000 UTC m=+9036.140735042" watchObservedRunningTime="2025-12-05 09:25:35.612457739 +0000 UTC m=+9036.141365000" Dec 05 09:25:37 crc kubenswrapper[4997]: I1205 09:25:37.594845 4997 generic.go:334] "Generic (PLEG): container finished" podID="ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a" containerID="8bfa02d026ab3bbe68ef8d3ebbe6091122e663647f85ef256246151ef1f29237" exitCode=0 Dec 05 09:25:37 crc kubenswrapper[4997]: I1205 09:25:37.594944 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2w7hm" event={"ID":"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a","Type":"ContainerDied","Data":"8bfa02d026ab3bbe68ef8d3ebbe6091122e663647f85ef256246151ef1f29237"} Dec 05 09:25:39 crc kubenswrapper[4997]: I1205 09:25:39.619314 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2w7hm" event={"ID":"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a","Type":"ContainerStarted","Data":"c3d0dd7f25e792fdacb878d98afc0c7edfd1af4f0c1c57b8e9bbcc60d260b402"} Dec 05 09:25:39 crc kubenswrapper[4997]: I1205 09:25:39.648050 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2w7hm" podStartSLOduration=3.661984485 podStartE2EDuration="8.648031555s" podCreationTimestamp="2025-12-05 09:25:31 +0000 UTC" firstStartedPulling="2025-12-05 09:25:33.546362921 +0000 UTC m=+9034.075270182" lastFinishedPulling="2025-12-05 09:25:38.532409991 +0000 UTC m=+9039.061317252" observedRunningTime="2025-12-05 09:25:39.640472111 +0000 UTC m=+9040.169379382" watchObservedRunningTime="2025-12-05 09:25:39.648031555 +0000 UTC m=+9040.176938816" Dec 05 09:25:40 crc kubenswrapper[4997]: I1205 09:25:40.629060 4997 generic.go:334] "Generic (PLEG): container finished" podID="f8efbd91-b09a-4dac-bf7d-b0b0433023ff" containerID="73ba391f5a0b1beaa1a5a0301c3238cbb27d88151909d83d8123eaaa5285d61e" exitCode=0 Dec 05 09:25:40 crc kubenswrapper[4997]: I1205 09:25:40.629132 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" event={"ID":"f8efbd91-b09a-4dac-bf7d-b0b0433023ff","Type":"ContainerDied","Data":"73ba391f5a0b1beaa1a5a0301c3238cbb27d88151909d83d8123eaaa5285d61e"} Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.088334 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2w7hm" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.089149 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2w7hm" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.526097 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.626647 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-ssh-key\") pod \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\" (UID: \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\") " Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.627006 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-ceph\") pod \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\" (UID: \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\") " Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.627085 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-inventory\") pod \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\" (UID: \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\") " Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.627310 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbxk9\" (UniqueName: \"kubernetes.io/projected/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-kube-api-access-jbxk9\") pod \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\" (UID: \"f8efbd91-b09a-4dac-bf7d-b0b0433023ff\") " Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.635460 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-ceph" (OuterVolumeSpecName: "ceph") pod "f8efbd91-b09a-4dac-bf7d-b0b0433023ff" (UID: "f8efbd91-b09a-4dac-bf7d-b0b0433023ff"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.639787 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-kube-api-access-jbxk9" (OuterVolumeSpecName: "kube-api-access-jbxk9") pod "f8efbd91-b09a-4dac-bf7d-b0b0433023ff" (UID: "f8efbd91-b09a-4dac-bf7d-b0b0433023ff"). InnerVolumeSpecName "kube-api-access-jbxk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.664436 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" event={"ID":"f8efbd91-b09a-4dac-bf7d-b0b0433023ff","Type":"ContainerDied","Data":"37282f4ba885e6f18af79b7b770dfc26705e665efc15fdc0c9e3eb4f6ab439e2"} Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.664494 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37282f4ba885e6f18af79b7b770dfc26705e665efc15fdc0c9e3eb4f6ab439e2" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.664559 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-cfqnz" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.666584 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f8efbd91-b09a-4dac-bf7d-b0b0433023ff" (UID: "f8efbd91-b09a-4dac-bf7d-b0b0433023ff"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.673459 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-inventory" (OuterVolumeSpecName: "inventory") pod "f8efbd91-b09a-4dac-bf7d-b0b0433023ff" (UID: "f8efbd91-b09a-4dac-bf7d-b0b0433023ff"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.730867 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbxk9\" (UniqueName: \"kubernetes.io/projected/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-kube-api-access-jbxk9\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.731095 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.731207 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.731274 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8efbd91-b09a-4dac-bf7d-b0b0433023ff-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.742639 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-9dg4x"] Dec 05 09:25:42 crc kubenswrapper[4997]: E1205 09:25:42.743153 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8efbd91-b09a-4dac-bf7d-b0b0433023ff" containerName="ceph-client-openstack-openstack-cell1" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.743167 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8efbd91-b09a-4dac-bf7d-b0b0433023ff" containerName="ceph-client-openstack-openstack-cell1" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.743476 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8efbd91-b09a-4dac-bf7d-b0b0433023ff" containerName="ceph-client-openstack-openstack-cell1" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.744288 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.756413 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-9dg4x"] Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.833210 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.833266 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-inventory\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.833369 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.833442 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99zpp\" (UniqueName: \"kubernetes.io/projected/d9179dc6-dc8e-493d-8120-27d7a3e723e6-kube-api-access-99zpp\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.833559 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ceph\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.833651 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ssh-key\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.935225 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.935881 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-inventory\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.935943 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.935985 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99zpp\" (UniqueName: \"kubernetes.io/projected/d9179dc6-dc8e-493d-8120-27d7a3e723e6-kube-api-access-99zpp\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.936049 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ceph\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.936370 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ssh-key\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.936880 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.944492 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ssh-key\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.944787 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.945903 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ceph\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.945941 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-inventory\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:42 crc kubenswrapper[4997]: I1205 09:25:42.956078 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99zpp\" (UniqueName: \"kubernetes.io/projected/d9179dc6-dc8e-493d-8120-27d7a3e723e6-kube-api-access-99zpp\") pod \"ovn-openstack-openstack-cell1-9dg4x\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:43 crc kubenswrapper[4997]: I1205 09:25:43.071461 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:25:43 crc kubenswrapper[4997]: I1205 09:25:43.420579 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-9dg4x"] Dec 05 09:25:43 crc kubenswrapper[4997]: W1205 09:25:43.423800 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9179dc6_dc8e_493d_8120_27d7a3e723e6.slice/crio-88a3c7fd29d72ffd458c79cc908390428fde9f4a4f8fc77dbf4377515a14cb7d WatchSource:0}: Error finding container 88a3c7fd29d72ffd458c79cc908390428fde9f4a4f8fc77dbf4377515a14cb7d: Status 404 returned error can't find the container with id 88a3c7fd29d72ffd458c79cc908390428fde9f4a4f8fc77dbf4377515a14cb7d Dec 05 09:25:43 crc kubenswrapper[4997]: I1205 09:25:43.467169 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2w7hm" podUID="ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a" containerName="registry-server" probeResult="failure" output=< Dec 05 09:25:43 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 09:25:43 crc kubenswrapper[4997]: > Dec 05 09:25:43 crc kubenswrapper[4997]: I1205 09:25:43.674568 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-9dg4x" event={"ID":"d9179dc6-dc8e-493d-8120-27d7a3e723e6","Type":"ContainerStarted","Data":"88a3c7fd29d72ffd458c79cc908390428fde9f4a4f8fc77dbf4377515a14cb7d"} Dec 05 09:25:43 crc kubenswrapper[4997]: I1205 09:25:43.750214 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:25:43 crc kubenswrapper[4997]: E1205 09:25:43.750511 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:25:44 crc kubenswrapper[4997]: I1205 09:25:44.688084 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-9dg4x" event={"ID":"d9179dc6-dc8e-493d-8120-27d7a3e723e6","Type":"ContainerStarted","Data":"494e2eb0d8ca62edf890b2d719a5864d9ff4280c222b83f0c4f768232885c51e"} Dec 05 09:25:44 crc kubenswrapper[4997]: I1205 09:25:44.708520 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-9dg4x" podStartSLOduration=2.290856827 podStartE2EDuration="2.708499114s" podCreationTimestamp="2025-12-05 09:25:42 +0000 UTC" firstStartedPulling="2025-12-05 09:25:43.429365806 +0000 UTC m=+9043.958273067" lastFinishedPulling="2025-12-05 09:25:43.847008093 +0000 UTC m=+9044.375915354" observedRunningTime="2025-12-05 09:25:44.708108634 +0000 UTC m=+9045.237015895" watchObservedRunningTime="2025-12-05 09:25:44.708499114 +0000 UTC m=+9045.237406375" Dec 05 09:25:52 crc kubenswrapper[4997]: I1205 09:25:52.135007 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2w7hm" Dec 05 09:25:52 crc kubenswrapper[4997]: I1205 09:25:52.200100 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2w7hm" Dec 05 09:25:52 crc kubenswrapper[4997]: I1205 09:25:52.387852 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2w7hm"] Dec 05 09:25:53 crc kubenswrapper[4997]: I1205 09:25:53.776394 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2w7hm" podUID="ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a" containerName="registry-server" containerID="cri-o://c3d0dd7f25e792fdacb878d98afc0c7edfd1af4f0c1c57b8e9bbcc60d260b402" gracePeriod=2 Dec 05 09:25:54 crc kubenswrapper[4997]: I1205 09:25:54.784142 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2w7hm" Dec 05 09:25:54 crc kubenswrapper[4997]: I1205 09:25:54.789849 4997 generic.go:334] "Generic (PLEG): container finished" podID="ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a" containerID="c3d0dd7f25e792fdacb878d98afc0c7edfd1af4f0c1c57b8e9bbcc60d260b402" exitCode=0 Dec 05 09:25:54 crc kubenswrapper[4997]: I1205 09:25:54.789890 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2w7hm" event={"ID":"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a","Type":"ContainerDied","Data":"c3d0dd7f25e792fdacb878d98afc0c7edfd1af4f0c1c57b8e9bbcc60d260b402"} Dec 05 09:25:54 crc kubenswrapper[4997]: I1205 09:25:54.789920 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2w7hm" event={"ID":"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a","Type":"ContainerDied","Data":"9bcba13041912ceab1497c12e1fc0e36a2757fccb0fedebf30a1cd65b3f29290"} Dec 05 09:25:54 crc kubenswrapper[4997]: I1205 09:25:54.789938 4997 scope.go:117] "RemoveContainer" containerID="c3d0dd7f25e792fdacb878d98afc0c7edfd1af4f0c1c57b8e9bbcc60d260b402" Dec 05 09:25:54 crc kubenswrapper[4997]: I1205 09:25:54.789980 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2w7hm" Dec 05 09:25:54 crc kubenswrapper[4997]: I1205 09:25:54.816401 4997 scope.go:117] "RemoveContainer" containerID="8bfa02d026ab3bbe68ef8d3ebbe6091122e663647f85ef256246151ef1f29237" Dec 05 09:25:54 crc kubenswrapper[4997]: I1205 09:25:54.892152 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-utilities\") pod \"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a\" (UID: \"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a\") " Dec 05 09:25:54 crc kubenswrapper[4997]: I1205 09:25:54.892342 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-catalog-content\") pod \"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a\" (UID: \"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a\") " Dec 05 09:25:54 crc kubenswrapper[4997]: I1205 09:25:54.892429 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v87nv\" (UniqueName: \"kubernetes.io/projected/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-kube-api-access-v87nv\") pod \"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a\" (UID: \"ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a\") " Dec 05 09:25:54 crc kubenswrapper[4997]: I1205 09:25:54.892942 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-utilities" (OuterVolumeSpecName: "utilities") pod "ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a" (UID: "ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:25:54 crc kubenswrapper[4997]: I1205 09:25:54.893279 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:54 crc kubenswrapper[4997]: I1205 09:25:54.992955 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a" (UID: "ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:25:54 crc kubenswrapper[4997]: I1205 09:25:54.994606 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:55 crc kubenswrapper[4997]: I1205 09:25:55.205756 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-kube-api-access-v87nv" (OuterVolumeSpecName: "kube-api-access-v87nv") pod "ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a" (UID: "ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a"). InnerVolumeSpecName "kube-api-access-v87nv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:25:55 crc kubenswrapper[4997]: I1205 09:25:55.219895 4997 scope.go:117] "RemoveContainer" containerID="3d09f3187cca7690dbfa1af052f061e0383ac3f1d93011b385c265a1354e7f51" Dec 05 09:25:55 crc kubenswrapper[4997]: I1205 09:25:55.299698 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v87nv\" (UniqueName: \"kubernetes.io/projected/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a-kube-api-access-v87nv\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:55 crc kubenswrapper[4997]: I1205 09:25:55.309182 4997 scope.go:117] "RemoveContainer" containerID="c3d0dd7f25e792fdacb878d98afc0c7edfd1af4f0c1c57b8e9bbcc60d260b402" Dec 05 09:25:55 crc kubenswrapper[4997]: E1205 09:25:55.309665 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3d0dd7f25e792fdacb878d98afc0c7edfd1af4f0c1c57b8e9bbcc60d260b402\": container with ID starting with c3d0dd7f25e792fdacb878d98afc0c7edfd1af4f0c1c57b8e9bbcc60d260b402 not found: ID does not exist" containerID="c3d0dd7f25e792fdacb878d98afc0c7edfd1af4f0c1c57b8e9bbcc60d260b402" Dec 05 09:25:55 crc kubenswrapper[4997]: I1205 09:25:55.309706 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3d0dd7f25e792fdacb878d98afc0c7edfd1af4f0c1c57b8e9bbcc60d260b402"} err="failed to get container status \"c3d0dd7f25e792fdacb878d98afc0c7edfd1af4f0c1c57b8e9bbcc60d260b402\": rpc error: code = NotFound desc = could not find container \"c3d0dd7f25e792fdacb878d98afc0c7edfd1af4f0c1c57b8e9bbcc60d260b402\": container with ID starting with c3d0dd7f25e792fdacb878d98afc0c7edfd1af4f0c1c57b8e9bbcc60d260b402 not found: ID does not exist" Dec 05 09:25:55 crc kubenswrapper[4997]: I1205 09:25:55.309731 4997 scope.go:117] "RemoveContainer" containerID="8bfa02d026ab3bbe68ef8d3ebbe6091122e663647f85ef256246151ef1f29237" Dec 05 09:25:55 crc kubenswrapper[4997]: E1205 09:25:55.310198 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bfa02d026ab3bbe68ef8d3ebbe6091122e663647f85ef256246151ef1f29237\": container with ID starting with 8bfa02d026ab3bbe68ef8d3ebbe6091122e663647f85ef256246151ef1f29237 not found: ID does not exist" containerID="8bfa02d026ab3bbe68ef8d3ebbe6091122e663647f85ef256246151ef1f29237" Dec 05 09:25:55 crc kubenswrapper[4997]: I1205 09:25:55.310226 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bfa02d026ab3bbe68ef8d3ebbe6091122e663647f85ef256246151ef1f29237"} err="failed to get container status \"8bfa02d026ab3bbe68ef8d3ebbe6091122e663647f85ef256246151ef1f29237\": rpc error: code = NotFound desc = could not find container \"8bfa02d026ab3bbe68ef8d3ebbe6091122e663647f85ef256246151ef1f29237\": container with ID starting with 8bfa02d026ab3bbe68ef8d3ebbe6091122e663647f85ef256246151ef1f29237 not found: ID does not exist" Dec 05 09:25:55 crc kubenswrapper[4997]: I1205 09:25:55.310244 4997 scope.go:117] "RemoveContainer" containerID="3d09f3187cca7690dbfa1af052f061e0383ac3f1d93011b385c265a1354e7f51" Dec 05 09:25:55 crc kubenswrapper[4997]: E1205 09:25:55.310825 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d09f3187cca7690dbfa1af052f061e0383ac3f1d93011b385c265a1354e7f51\": container with ID starting with 3d09f3187cca7690dbfa1af052f061e0383ac3f1d93011b385c265a1354e7f51 not found: ID does not exist" containerID="3d09f3187cca7690dbfa1af052f061e0383ac3f1d93011b385c265a1354e7f51" Dec 05 09:25:55 crc kubenswrapper[4997]: I1205 09:25:55.310874 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d09f3187cca7690dbfa1af052f061e0383ac3f1d93011b385c265a1354e7f51"} err="failed to get container status \"3d09f3187cca7690dbfa1af052f061e0383ac3f1d93011b385c265a1354e7f51\": rpc error: code = NotFound desc = could not find container \"3d09f3187cca7690dbfa1af052f061e0383ac3f1d93011b385c265a1354e7f51\": container with ID starting with 3d09f3187cca7690dbfa1af052f061e0383ac3f1d93011b385c265a1354e7f51 not found: ID does not exist" Dec 05 09:25:55 crc kubenswrapper[4997]: I1205 09:25:55.462063 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2w7hm"] Dec 05 09:25:55 crc kubenswrapper[4997]: I1205 09:25:55.473166 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2w7hm"] Dec 05 09:25:55 crc kubenswrapper[4997]: I1205 09:25:55.767244 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a" path="/var/lib/kubelet/pods/ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a/volumes" Dec 05 09:25:57 crc kubenswrapper[4997]: I1205 09:25:57.751221 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:25:57 crc kubenswrapper[4997]: E1205 09:25:57.751846 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:26:07 crc kubenswrapper[4997]: I1205 09:26:07.922718 4997 generic.go:334] "Generic (PLEG): container finished" podID="51661db5-879f-4b6c-a563-ff1c84e47f0d" containerID="470e604c7fa4354aefd190c4b20aac805e6587b1e1bd51b423c8cbbc5e327562" exitCode=0 Dec 05 09:26:07 crc kubenswrapper[4997]: I1205 09:26:07.922808 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-9mbvl" event={"ID":"51661db5-879f-4b6c-a563-ff1c84e47f0d","Type":"ContainerDied","Data":"470e604c7fa4354aefd190c4b20aac805e6587b1e1bd51b423c8cbbc5e327562"} Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.448774 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.552226 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-inventory\") pod \"51661db5-879f-4b6c-a563-ff1c84e47f0d\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.552351 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l64f9\" (UniqueName: \"kubernetes.io/projected/51661db5-879f-4b6c-a563-ff1c84e47f0d-kube-api-access-l64f9\") pod \"51661db5-879f-4b6c-a563-ff1c84e47f0d\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.552411 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-ssh-key\") pod \"51661db5-879f-4b6c-a563-ff1c84e47f0d\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.552450 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-ovn-combined-ca-bundle\") pod \"51661db5-879f-4b6c-a563-ff1c84e47f0d\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.552647 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/51661db5-879f-4b6c-a563-ff1c84e47f0d-ovncontroller-config-0\") pod \"51661db5-879f-4b6c-a563-ff1c84e47f0d\" (UID: \"51661db5-879f-4b6c-a563-ff1c84e47f0d\") " Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.558824 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "51661db5-879f-4b6c-a563-ff1c84e47f0d" (UID: "51661db5-879f-4b6c-a563-ff1c84e47f0d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.560274 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51661db5-879f-4b6c-a563-ff1c84e47f0d-kube-api-access-l64f9" (OuterVolumeSpecName: "kube-api-access-l64f9") pod "51661db5-879f-4b6c-a563-ff1c84e47f0d" (UID: "51661db5-879f-4b6c-a563-ff1c84e47f0d"). InnerVolumeSpecName "kube-api-access-l64f9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.585147 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51661db5-879f-4b6c-a563-ff1c84e47f0d-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "51661db5-879f-4b6c-a563-ff1c84e47f0d" (UID: "51661db5-879f-4b6c-a563-ff1c84e47f0d"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.588866 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-inventory" (OuterVolumeSpecName: "inventory") pod "51661db5-879f-4b6c-a563-ff1c84e47f0d" (UID: "51661db5-879f-4b6c-a563-ff1c84e47f0d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.591341 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "51661db5-879f-4b6c-a563-ff1c84e47f0d" (UID: "51661db5-879f-4b6c-a563-ff1c84e47f0d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.656204 4997 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/51661db5-879f-4b6c-a563-ff1c84e47f0d-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.656255 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.656269 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l64f9\" (UniqueName: \"kubernetes.io/projected/51661db5-879f-4b6c-a563-ff1c84e47f0d-kube-api-access-l64f9\") on node \"crc\" DevicePath \"\"" Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.656280 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.656296 4997 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51661db5-879f-4b6c-a563-ff1c84e47f0d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.756829 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:26:09 crc kubenswrapper[4997]: E1205 09:26:09.757116 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.958830 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-9mbvl" event={"ID":"51661db5-879f-4b6c-a563-ff1c84e47f0d","Type":"ContainerDied","Data":"b2e83ff9dd96f92b4edbba9039bff743c0b980760cea3f9a9020bb1069aeadf8"} Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.958895 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-9mbvl" Dec 05 09:26:09 crc kubenswrapper[4997]: I1205 09:26:09.958915 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2e83ff9dd96f92b4edbba9039bff743c0b980760cea3f9a9020bb1069aeadf8" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.047333 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-networker-n7m22"] Dec 05 09:26:10 crc kubenswrapper[4997]: E1205 09:26:10.047795 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a" containerName="extract-utilities" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.047815 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a" containerName="extract-utilities" Dec 05 09:26:10 crc kubenswrapper[4997]: E1205 09:26:10.047842 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a" containerName="extract-content" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.047849 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a" containerName="extract-content" Dec 05 09:26:10 crc kubenswrapper[4997]: E1205 09:26:10.047868 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51661db5-879f-4b6c-a563-ff1c84e47f0d" containerName="ovn-openstack-openstack-networker" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.047876 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="51661db5-879f-4b6c-a563-ff1c84e47f0d" containerName="ovn-openstack-openstack-networker" Dec 05 09:26:10 crc kubenswrapper[4997]: E1205 09:26:10.047906 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a" containerName="registry-server" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.047914 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a" containerName="registry-server" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.048347 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="51661db5-879f-4b6c-a563-ff1c84e47f0d" containerName="ovn-openstack-openstack-networker" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.048371 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebf3ed66-e553-4722-a8d4-ed0a5ff3da4a" containerName="registry-server" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.049636 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.060946 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.061290 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.061464 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-4542r" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.061634 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.067878 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-networker-n7m22"] Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.172110 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnt86\" (UniqueName: \"kubernetes.io/projected/ba5f43cb-d3c6-49e8-9f83-551cd860f736-kube-api-access-pnt86\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.172219 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.172253 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-inventory\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.172453 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-ssh-key\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.172553 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.172723 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.275989 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnt86\" (UniqueName: \"kubernetes.io/projected/ba5f43cb-d3c6-49e8-9f83-551cd860f736-kube-api-access-pnt86\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.276282 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.276304 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-inventory\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.276364 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-ssh-key\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.276392 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.276415 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.280933 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-ssh-key\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.281602 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.281968 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-inventory\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.283086 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.283550 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.298278 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnt86\" (UniqueName: \"kubernetes.io/projected/ba5f43cb-d3c6-49e8-9f83-551cd860f736-kube-api-access-pnt86\") pod \"neutron-metadata-openstack-openstack-networker-n7m22\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.376740 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.922574 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-networker-n7m22"] Dec 05 09:26:10 crc kubenswrapper[4997]: I1205 09:26:10.971925 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" event={"ID":"ba5f43cb-d3c6-49e8-9f83-551cd860f736","Type":"ContainerStarted","Data":"5d42d47b57f6a38dfc54745654286d9e84b44182c74a080553c704a98c3e4dfe"} Dec 05 09:26:11 crc kubenswrapper[4997]: I1205 09:26:11.983526 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" event={"ID":"ba5f43cb-d3c6-49e8-9f83-551cd860f736","Type":"ContainerStarted","Data":"3907d93a4d2142741ad4e478fb18413e7a2b70f798d5e7d29f2d937d8649f377"} Dec 05 09:26:12 crc kubenswrapper[4997]: I1205 09:26:12.005004 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" podStartSLOduration=1.580687789 podStartE2EDuration="2.004986216s" podCreationTimestamp="2025-12-05 09:26:10 +0000 UTC" firstStartedPulling="2025-12-05 09:26:10.923795201 +0000 UTC m=+9071.452702462" lastFinishedPulling="2025-12-05 09:26:11.348093628 +0000 UTC m=+9071.877000889" observedRunningTime="2025-12-05 09:26:12.000641219 +0000 UTC m=+9072.529548500" watchObservedRunningTime="2025-12-05 09:26:12.004986216 +0000 UTC m=+9072.533893477" Dec 05 09:26:20 crc kubenswrapper[4997]: I1205 09:26:20.749631 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:26:20 crc kubenswrapper[4997]: E1205 09:26:20.750767 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:26:33 crc kubenswrapper[4997]: I1205 09:26:33.749693 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:26:33 crc kubenswrapper[4997]: E1205 09:26:33.750785 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:26:48 crc kubenswrapper[4997]: I1205 09:26:48.749014 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:26:48 crc kubenswrapper[4997]: E1205 09:26:48.749804 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:26:58 crc kubenswrapper[4997]: I1205 09:26:58.499211 4997 generic.go:334] "Generic (PLEG): container finished" podID="d9179dc6-dc8e-493d-8120-27d7a3e723e6" containerID="494e2eb0d8ca62edf890b2d719a5864d9ff4280c222b83f0c4f768232885c51e" exitCode=0 Dec 05 09:26:58 crc kubenswrapper[4997]: I1205 09:26:58.499318 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-9dg4x" event={"ID":"d9179dc6-dc8e-493d-8120-27d7a3e723e6","Type":"ContainerDied","Data":"494e2eb0d8ca62edf890b2d719a5864d9ff4280c222b83f0c4f768232885c51e"} Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.046535 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.106352 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99zpp\" (UniqueName: \"kubernetes.io/projected/d9179dc6-dc8e-493d-8120-27d7a3e723e6-kube-api-access-99zpp\") pod \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.106533 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ceph\") pod \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.106559 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ssh-key\") pod \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.106601 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ovncontroller-config-0\") pod \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.106806 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ovn-combined-ca-bundle\") pod \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.106848 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-inventory\") pod \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\" (UID: \"d9179dc6-dc8e-493d-8120-27d7a3e723e6\") " Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.113704 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ceph" (OuterVolumeSpecName: "ceph") pod "d9179dc6-dc8e-493d-8120-27d7a3e723e6" (UID: "d9179dc6-dc8e-493d-8120-27d7a3e723e6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.113791 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d9179dc6-dc8e-493d-8120-27d7a3e723e6" (UID: "d9179dc6-dc8e-493d-8120-27d7a3e723e6"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.114324 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9179dc6-dc8e-493d-8120-27d7a3e723e6-kube-api-access-99zpp" (OuterVolumeSpecName: "kube-api-access-99zpp") pod "d9179dc6-dc8e-493d-8120-27d7a3e723e6" (UID: "d9179dc6-dc8e-493d-8120-27d7a3e723e6"). InnerVolumeSpecName "kube-api-access-99zpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.136798 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d9179dc6-dc8e-493d-8120-27d7a3e723e6" (UID: "d9179dc6-dc8e-493d-8120-27d7a3e723e6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.143878 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "d9179dc6-dc8e-493d-8120-27d7a3e723e6" (UID: "d9179dc6-dc8e-493d-8120-27d7a3e723e6"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.146143 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-inventory" (OuterVolumeSpecName: "inventory") pod "d9179dc6-dc8e-493d-8120-27d7a3e723e6" (UID: "d9179dc6-dc8e-493d-8120-27d7a3e723e6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.213332 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.213383 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99zpp\" (UniqueName: \"kubernetes.io/projected/d9179dc6-dc8e-493d-8120-27d7a3e723e6-kube-api-access-99zpp\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.213398 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.213410 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.213422 4997 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.213436 4997 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9179dc6-dc8e-493d-8120-27d7a3e723e6-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.521009 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-9dg4x" event={"ID":"d9179dc6-dc8e-493d-8120-27d7a3e723e6","Type":"ContainerDied","Data":"88a3c7fd29d72ffd458c79cc908390428fde9f4a4f8fc77dbf4377515a14cb7d"} Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.521323 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88a3c7fd29d72ffd458c79cc908390428fde9f4a4f8fc77dbf4377515a14cb7d" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.521107 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-9dg4x" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.650661 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-v7mv9"] Dec 05 09:27:00 crc kubenswrapper[4997]: E1205 09:27:00.651115 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9179dc6-dc8e-493d-8120-27d7a3e723e6" containerName="ovn-openstack-openstack-cell1" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.651127 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9179dc6-dc8e-493d-8120-27d7a3e723e6" containerName="ovn-openstack-openstack-cell1" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.651389 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9179dc6-dc8e-493d-8120-27d7a3e723e6" containerName="ovn-openstack-openstack-cell1" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.652269 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.654738 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.655057 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.685668 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-v7mv9"] Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.724305 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.724361 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.724397 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g5bn\" (UniqueName: \"kubernetes.io/projected/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-kube-api-access-6g5bn\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.724652 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.724785 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.724864 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.724907 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.749605 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:27:00 crc kubenswrapper[4997]: E1205 09:27:00.749924 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.827054 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.827109 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.827163 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g5bn\" (UniqueName: \"kubernetes.io/projected/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-kube-api-access-6g5bn\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.827261 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.827349 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.827402 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.827429 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.832400 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.838517 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.850717 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.851027 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.851277 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.852102 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g5bn\" (UniqueName: \"kubernetes.io/projected/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-kube-api-access-6g5bn\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.862559 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-v7mv9\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:00 crc kubenswrapper[4997]: I1205 09:27:00.980393 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:27:01 crc kubenswrapper[4997]: I1205 09:27:01.525319 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-v7mv9"] Dec 05 09:27:01 crc kubenswrapper[4997]: I1205 09:27:01.526212 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:27:01 crc kubenswrapper[4997]: I1205 09:27:01.534702 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" event={"ID":"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4","Type":"ContainerStarted","Data":"89be7a93b7e503f3c3a4afa07d12d20dc3a35e89e53cc66aa9092561a98b8651"} Dec 05 09:27:02 crc kubenswrapper[4997]: I1205 09:27:02.548848 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" event={"ID":"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4","Type":"ContainerStarted","Data":"3fce8cd787a4871150309f656b7d97f26f49f2f884c799cfc14f15ab729ac81c"} Dec 05 09:27:02 crc kubenswrapper[4997]: I1205 09:27:02.578304 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" podStartSLOduration=2.048148614 podStartE2EDuration="2.578272039s" podCreationTimestamp="2025-12-05 09:27:00 +0000 UTC" firstStartedPulling="2025-12-05 09:27:01.525920613 +0000 UTC m=+9122.054827874" lastFinishedPulling="2025-12-05 09:27:02.056044038 +0000 UTC m=+9122.584951299" observedRunningTime="2025-12-05 09:27:02.567428746 +0000 UTC m=+9123.096336027" watchObservedRunningTime="2025-12-05 09:27:02.578272039 +0000 UTC m=+9123.107179310" Dec 05 09:27:14 crc kubenswrapper[4997]: I1205 09:27:14.223729 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p84gm"] Dec 05 09:27:14 crc kubenswrapper[4997]: I1205 09:27:14.227024 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p84gm" Dec 05 09:27:14 crc kubenswrapper[4997]: I1205 09:27:14.257679 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p84gm"] Dec 05 09:27:14 crc kubenswrapper[4997]: I1205 09:27:14.323635 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nq5v\" (UniqueName: \"kubernetes.io/projected/1394f200-4e7e-4cfa-8f47-c356b385137d-kube-api-access-9nq5v\") pod \"certified-operators-p84gm\" (UID: \"1394f200-4e7e-4cfa-8f47-c356b385137d\") " pod="openshift-marketplace/certified-operators-p84gm" Dec 05 09:27:14 crc kubenswrapper[4997]: I1205 09:27:14.323787 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1394f200-4e7e-4cfa-8f47-c356b385137d-utilities\") pod \"certified-operators-p84gm\" (UID: \"1394f200-4e7e-4cfa-8f47-c356b385137d\") " pod="openshift-marketplace/certified-operators-p84gm" Dec 05 09:27:14 crc kubenswrapper[4997]: I1205 09:27:14.323850 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1394f200-4e7e-4cfa-8f47-c356b385137d-catalog-content\") pod \"certified-operators-p84gm\" (UID: \"1394f200-4e7e-4cfa-8f47-c356b385137d\") " pod="openshift-marketplace/certified-operators-p84gm" Dec 05 09:27:14 crc kubenswrapper[4997]: I1205 09:27:14.427389 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1394f200-4e7e-4cfa-8f47-c356b385137d-catalog-content\") pod \"certified-operators-p84gm\" (UID: \"1394f200-4e7e-4cfa-8f47-c356b385137d\") " pod="openshift-marketplace/certified-operators-p84gm" Dec 05 09:27:14 crc kubenswrapper[4997]: I1205 09:27:14.427569 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nq5v\" (UniqueName: \"kubernetes.io/projected/1394f200-4e7e-4cfa-8f47-c356b385137d-kube-api-access-9nq5v\") pod \"certified-operators-p84gm\" (UID: \"1394f200-4e7e-4cfa-8f47-c356b385137d\") " pod="openshift-marketplace/certified-operators-p84gm" Dec 05 09:27:14 crc kubenswrapper[4997]: I1205 09:27:14.427751 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1394f200-4e7e-4cfa-8f47-c356b385137d-utilities\") pod \"certified-operators-p84gm\" (UID: \"1394f200-4e7e-4cfa-8f47-c356b385137d\") " pod="openshift-marketplace/certified-operators-p84gm" Dec 05 09:27:14 crc kubenswrapper[4997]: I1205 09:27:14.428048 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1394f200-4e7e-4cfa-8f47-c356b385137d-catalog-content\") pod \"certified-operators-p84gm\" (UID: \"1394f200-4e7e-4cfa-8f47-c356b385137d\") " pod="openshift-marketplace/certified-operators-p84gm" Dec 05 09:27:14 crc kubenswrapper[4997]: I1205 09:27:14.428438 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1394f200-4e7e-4cfa-8f47-c356b385137d-utilities\") pod \"certified-operators-p84gm\" (UID: \"1394f200-4e7e-4cfa-8f47-c356b385137d\") " pod="openshift-marketplace/certified-operators-p84gm" Dec 05 09:27:14 crc kubenswrapper[4997]: I1205 09:27:14.457227 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nq5v\" (UniqueName: \"kubernetes.io/projected/1394f200-4e7e-4cfa-8f47-c356b385137d-kube-api-access-9nq5v\") pod \"certified-operators-p84gm\" (UID: \"1394f200-4e7e-4cfa-8f47-c356b385137d\") " pod="openshift-marketplace/certified-operators-p84gm" Dec 05 09:27:14 crc kubenswrapper[4997]: I1205 09:27:14.552168 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p84gm" Dec 05 09:27:15 crc kubenswrapper[4997]: I1205 09:27:15.142785 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p84gm"] Dec 05 09:27:15 crc kubenswrapper[4997]: I1205 09:27:15.726439 4997 generic.go:334] "Generic (PLEG): container finished" podID="ba5f43cb-d3c6-49e8-9f83-551cd860f736" containerID="3907d93a4d2142741ad4e478fb18413e7a2b70f798d5e7d29f2d937d8649f377" exitCode=0 Dec 05 09:27:15 crc kubenswrapper[4997]: I1205 09:27:15.726543 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" event={"ID":"ba5f43cb-d3c6-49e8-9f83-551cd860f736","Type":"ContainerDied","Data":"3907d93a4d2142741ad4e478fb18413e7a2b70f798d5e7d29f2d937d8649f377"} Dec 05 09:27:15 crc kubenswrapper[4997]: I1205 09:27:15.728436 4997 generic.go:334] "Generic (PLEG): container finished" podID="1394f200-4e7e-4cfa-8f47-c356b385137d" containerID="3098f54888196e57babf2d2d3c50e94e19d47b0588431d393838ba822f1c3db1" exitCode=0 Dec 05 09:27:15 crc kubenswrapper[4997]: I1205 09:27:15.728467 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p84gm" event={"ID":"1394f200-4e7e-4cfa-8f47-c356b385137d","Type":"ContainerDied","Data":"3098f54888196e57babf2d2d3c50e94e19d47b0588431d393838ba822f1c3db1"} Dec 05 09:27:15 crc kubenswrapper[4997]: I1205 09:27:15.728496 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p84gm" event={"ID":"1394f200-4e7e-4cfa-8f47-c356b385137d","Type":"ContainerStarted","Data":"fd32e36d16b6b959eacc7f21c97c63e5f3ec41ce5dd103de85de2a2ec18bb009"} Dec 05 09:27:15 crc kubenswrapper[4997]: I1205 09:27:15.749295 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:27:15 crc kubenswrapper[4997]: E1205 09:27:15.749597 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:27:16 crc kubenswrapper[4997]: I1205 09:27:16.756356 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p84gm" event={"ID":"1394f200-4e7e-4cfa-8f47-c356b385137d","Type":"ContainerStarted","Data":"aca20c0f697ed763bf6bae586e2ee2030c1562eb4b8d3d216b18936134298fd7"} Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.275329 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.305589 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-neutron-metadata-combined-ca-bundle\") pod \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.305795 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-neutron-ovn-metadata-agent-neutron-config-0\") pod \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.305852 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-inventory\") pod \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.305899 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-nova-metadata-neutron-config-0\") pod \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.306135 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnt86\" (UniqueName: \"kubernetes.io/projected/ba5f43cb-d3c6-49e8-9f83-551cd860f736-kube-api-access-pnt86\") pod \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.306160 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-ssh-key\") pod \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\" (UID: \"ba5f43cb-d3c6-49e8-9f83-551cd860f736\") " Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.316910 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "ba5f43cb-d3c6-49e8-9f83-551cd860f736" (UID: "ba5f43cb-d3c6-49e8-9f83-551cd860f736"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.317153 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba5f43cb-d3c6-49e8-9f83-551cd860f736-kube-api-access-pnt86" (OuterVolumeSpecName: "kube-api-access-pnt86") pod "ba5f43cb-d3c6-49e8-9f83-551cd860f736" (UID: "ba5f43cb-d3c6-49e8-9f83-551cd860f736"). InnerVolumeSpecName "kube-api-access-pnt86". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.342184 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-inventory" (OuterVolumeSpecName: "inventory") pod "ba5f43cb-d3c6-49e8-9f83-551cd860f736" (UID: "ba5f43cb-d3c6-49e8-9f83-551cd860f736"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.360344 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "ba5f43cb-d3c6-49e8-9f83-551cd860f736" (UID: "ba5f43cb-d3c6-49e8-9f83-551cd860f736"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.364052 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ba5f43cb-d3c6-49e8-9f83-551cd860f736" (UID: "ba5f43cb-d3c6-49e8-9f83-551cd860f736"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.371745 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "ba5f43cb-d3c6-49e8-9f83-551cd860f736" (UID: "ba5f43cb-d3c6-49e8-9f83-551cd860f736"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.410479 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnt86\" (UniqueName: \"kubernetes.io/projected/ba5f43cb-d3c6-49e8-9f83-551cd860f736-kube-api-access-pnt86\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.410725 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.410754 4997 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.410782 4997 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.410802 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.410816 4997 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ba5f43cb-d3c6-49e8-9f83-551cd860f736-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.766471 4997 generic.go:334] "Generic (PLEG): container finished" podID="1394f200-4e7e-4cfa-8f47-c356b385137d" containerID="aca20c0f697ed763bf6bae586e2ee2030c1562eb4b8d3d216b18936134298fd7" exitCode=0 Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.766530 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p84gm" event={"ID":"1394f200-4e7e-4cfa-8f47-c356b385137d","Type":"ContainerDied","Data":"aca20c0f697ed763bf6bae586e2ee2030c1562eb4b8d3d216b18936134298fd7"} Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.768466 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" event={"ID":"ba5f43cb-d3c6-49e8-9f83-551cd860f736","Type":"ContainerDied","Data":"5d42d47b57f6a38dfc54745654286d9e84b44182c74a080553c704a98c3e4dfe"} Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.768493 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d42d47b57f6a38dfc54745654286d9e84b44182c74a080553c704a98c3e4dfe" Dec 05 09:27:17 crc kubenswrapper[4997]: I1205 09:27:17.768543 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-n7m22" Dec 05 09:27:18 crc kubenswrapper[4997]: I1205 09:27:18.781924 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p84gm" event={"ID":"1394f200-4e7e-4cfa-8f47-c356b385137d","Type":"ContainerStarted","Data":"5ee114df71a12614e64c132c8df043bb031687cc8dfb7d0dbde43c88678641d9"} Dec 05 09:27:18 crc kubenswrapper[4997]: I1205 09:27:18.808256 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p84gm" podStartSLOduration=2.221652047 podStartE2EDuration="4.808238019s" podCreationTimestamp="2025-12-05 09:27:14 +0000 UTC" firstStartedPulling="2025-12-05 09:27:15.7306671 +0000 UTC m=+9136.259574401" lastFinishedPulling="2025-12-05 09:27:18.317253112 +0000 UTC m=+9138.846160373" observedRunningTime="2025-12-05 09:27:18.798913478 +0000 UTC m=+9139.327820749" watchObservedRunningTime="2025-12-05 09:27:18.808238019 +0000 UTC m=+9139.337145280" Dec 05 09:27:24 crc kubenswrapper[4997]: I1205 09:27:24.552412 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p84gm" Dec 05 09:27:24 crc kubenswrapper[4997]: I1205 09:27:24.553036 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p84gm" Dec 05 09:27:24 crc kubenswrapper[4997]: I1205 09:27:24.620189 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p84gm" Dec 05 09:27:24 crc kubenswrapper[4997]: I1205 09:27:24.892635 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p84gm" Dec 05 09:27:24 crc kubenswrapper[4997]: I1205 09:27:24.950283 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p84gm"] Dec 05 09:27:26 crc kubenswrapper[4997]: I1205 09:27:26.881996 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p84gm" podUID="1394f200-4e7e-4cfa-8f47-c356b385137d" containerName="registry-server" containerID="cri-o://5ee114df71a12614e64c132c8df043bb031687cc8dfb7d0dbde43c88678641d9" gracePeriod=2 Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.409594 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p84gm" Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.531717 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1394f200-4e7e-4cfa-8f47-c356b385137d-utilities\") pod \"1394f200-4e7e-4cfa-8f47-c356b385137d\" (UID: \"1394f200-4e7e-4cfa-8f47-c356b385137d\") " Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.532306 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1394f200-4e7e-4cfa-8f47-c356b385137d-catalog-content\") pod \"1394f200-4e7e-4cfa-8f47-c356b385137d\" (UID: \"1394f200-4e7e-4cfa-8f47-c356b385137d\") " Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.532688 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1394f200-4e7e-4cfa-8f47-c356b385137d-utilities" (OuterVolumeSpecName: "utilities") pod "1394f200-4e7e-4cfa-8f47-c356b385137d" (UID: "1394f200-4e7e-4cfa-8f47-c356b385137d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.540964 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nq5v\" (UniqueName: \"kubernetes.io/projected/1394f200-4e7e-4cfa-8f47-c356b385137d-kube-api-access-9nq5v\") pod \"1394f200-4e7e-4cfa-8f47-c356b385137d\" (UID: \"1394f200-4e7e-4cfa-8f47-c356b385137d\") " Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.541893 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1394f200-4e7e-4cfa-8f47-c356b385137d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.546844 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1394f200-4e7e-4cfa-8f47-c356b385137d-kube-api-access-9nq5v" (OuterVolumeSpecName: "kube-api-access-9nq5v") pod "1394f200-4e7e-4cfa-8f47-c356b385137d" (UID: "1394f200-4e7e-4cfa-8f47-c356b385137d"). InnerVolumeSpecName "kube-api-access-9nq5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.588066 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1394f200-4e7e-4cfa-8f47-c356b385137d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1394f200-4e7e-4cfa-8f47-c356b385137d" (UID: "1394f200-4e7e-4cfa-8f47-c356b385137d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.643336 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1394f200-4e7e-4cfa-8f47-c356b385137d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.643370 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nq5v\" (UniqueName: \"kubernetes.io/projected/1394f200-4e7e-4cfa-8f47-c356b385137d-kube-api-access-9nq5v\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.897314 4997 generic.go:334] "Generic (PLEG): container finished" podID="1394f200-4e7e-4cfa-8f47-c356b385137d" containerID="5ee114df71a12614e64c132c8df043bb031687cc8dfb7d0dbde43c88678641d9" exitCode=0 Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.897361 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p84gm" event={"ID":"1394f200-4e7e-4cfa-8f47-c356b385137d","Type":"ContainerDied","Data":"5ee114df71a12614e64c132c8df043bb031687cc8dfb7d0dbde43c88678641d9"} Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.897388 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p84gm" event={"ID":"1394f200-4e7e-4cfa-8f47-c356b385137d","Type":"ContainerDied","Data":"fd32e36d16b6b959eacc7f21c97c63e5f3ec41ce5dd103de85de2a2ec18bb009"} Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.897387 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p84gm" Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.897405 4997 scope.go:117] "RemoveContainer" containerID="5ee114df71a12614e64c132c8df043bb031687cc8dfb7d0dbde43c88678641d9" Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.928586 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p84gm"] Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.936099 4997 scope.go:117] "RemoveContainer" containerID="aca20c0f697ed763bf6bae586e2ee2030c1562eb4b8d3d216b18936134298fd7" Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.941769 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p84gm"] Dec 05 09:27:27 crc kubenswrapper[4997]: I1205 09:27:27.965960 4997 scope.go:117] "RemoveContainer" containerID="3098f54888196e57babf2d2d3c50e94e19d47b0588431d393838ba822f1c3db1" Dec 05 09:27:28 crc kubenswrapper[4997]: I1205 09:27:28.022499 4997 scope.go:117] "RemoveContainer" containerID="5ee114df71a12614e64c132c8df043bb031687cc8dfb7d0dbde43c88678641d9" Dec 05 09:27:28 crc kubenswrapper[4997]: E1205 09:27:28.023296 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ee114df71a12614e64c132c8df043bb031687cc8dfb7d0dbde43c88678641d9\": container with ID starting with 5ee114df71a12614e64c132c8df043bb031687cc8dfb7d0dbde43c88678641d9 not found: ID does not exist" containerID="5ee114df71a12614e64c132c8df043bb031687cc8dfb7d0dbde43c88678641d9" Dec 05 09:27:28 crc kubenswrapper[4997]: I1205 09:27:28.023352 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ee114df71a12614e64c132c8df043bb031687cc8dfb7d0dbde43c88678641d9"} err="failed to get container status \"5ee114df71a12614e64c132c8df043bb031687cc8dfb7d0dbde43c88678641d9\": rpc error: code = NotFound desc = could not find container \"5ee114df71a12614e64c132c8df043bb031687cc8dfb7d0dbde43c88678641d9\": container with ID starting with 5ee114df71a12614e64c132c8df043bb031687cc8dfb7d0dbde43c88678641d9 not found: ID does not exist" Dec 05 09:27:28 crc kubenswrapper[4997]: I1205 09:27:28.023427 4997 scope.go:117] "RemoveContainer" containerID="aca20c0f697ed763bf6bae586e2ee2030c1562eb4b8d3d216b18936134298fd7" Dec 05 09:27:28 crc kubenswrapper[4997]: E1205 09:27:28.023942 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aca20c0f697ed763bf6bae586e2ee2030c1562eb4b8d3d216b18936134298fd7\": container with ID starting with aca20c0f697ed763bf6bae586e2ee2030c1562eb4b8d3d216b18936134298fd7 not found: ID does not exist" containerID="aca20c0f697ed763bf6bae586e2ee2030c1562eb4b8d3d216b18936134298fd7" Dec 05 09:27:28 crc kubenswrapper[4997]: I1205 09:27:28.023975 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aca20c0f697ed763bf6bae586e2ee2030c1562eb4b8d3d216b18936134298fd7"} err="failed to get container status \"aca20c0f697ed763bf6bae586e2ee2030c1562eb4b8d3d216b18936134298fd7\": rpc error: code = NotFound desc = could not find container \"aca20c0f697ed763bf6bae586e2ee2030c1562eb4b8d3d216b18936134298fd7\": container with ID starting with aca20c0f697ed763bf6bae586e2ee2030c1562eb4b8d3d216b18936134298fd7 not found: ID does not exist" Dec 05 09:27:28 crc kubenswrapper[4997]: I1205 09:27:28.023997 4997 scope.go:117] "RemoveContainer" containerID="3098f54888196e57babf2d2d3c50e94e19d47b0588431d393838ba822f1c3db1" Dec 05 09:27:28 crc kubenswrapper[4997]: E1205 09:27:28.024412 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3098f54888196e57babf2d2d3c50e94e19d47b0588431d393838ba822f1c3db1\": container with ID starting with 3098f54888196e57babf2d2d3c50e94e19d47b0588431d393838ba822f1c3db1 not found: ID does not exist" containerID="3098f54888196e57babf2d2d3c50e94e19d47b0588431d393838ba822f1c3db1" Dec 05 09:27:28 crc kubenswrapper[4997]: I1205 09:27:28.024497 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3098f54888196e57babf2d2d3c50e94e19d47b0588431d393838ba822f1c3db1"} err="failed to get container status \"3098f54888196e57babf2d2d3c50e94e19d47b0588431d393838ba822f1c3db1\": rpc error: code = NotFound desc = could not find container \"3098f54888196e57babf2d2d3c50e94e19d47b0588431d393838ba822f1c3db1\": container with ID starting with 3098f54888196e57babf2d2d3c50e94e19d47b0588431d393838ba822f1c3db1 not found: ID does not exist" Dec 05 09:27:29 crc kubenswrapper[4997]: I1205 09:27:29.763490 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1394f200-4e7e-4cfa-8f47-c356b385137d" path="/var/lib/kubelet/pods/1394f200-4e7e-4cfa-8f47-c356b385137d/volumes" Dec 05 09:27:30 crc kubenswrapper[4997]: I1205 09:27:30.750126 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:27:30 crc kubenswrapper[4997]: E1205 09:27:30.750526 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:27:45 crc kubenswrapper[4997]: I1205 09:27:45.749147 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:27:45 crc kubenswrapper[4997]: E1205 09:27:45.749951 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:27:56 crc kubenswrapper[4997]: I1205 09:27:56.748915 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:27:57 crc kubenswrapper[4997]: I1205 09:27:57.196195 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"6ce1c08e479b9a4b91cc9213507822cb7d19ab1ed1dfebf53f033a419412ffa0"} Dec 05 09:28:04 crc kubenswrapper[4997]: I1205 09:28:04.266129 4997 generic.go:334] "Generic (PLEG): container finished" podID="4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4" containerID="3fce8cd787a4871150309f656b7d97f26f49f2f884c799cfc14f15ab729ac81c" exitCode=0 Dec 05 09:28:04 crc kubenswrapper[4997]: I1205 09:28:04.266214 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" event={"ID":"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4","Type":"ContainerDied","Data":"3fce8cd787a4871150309f656b7d97f26f49f2f884c799cfc14f15ab729ac81c"} Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.760268 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.874533 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-nova-metadata-neutron-config-0\") pod \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.874599 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g5bn\" (UniqueName: \"kubernetes.io/projected/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-kube-api-access-6g5bn\") pod \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.874760 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-neutron-ovn-metadata-agent-neutron-config-0\") pod \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.874848 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-inventory\") pod \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.874869 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-neutron-metadata-combined-ca-bundle\") pod \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.874933 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-ssh-key\") pod \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.874964 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-ceph\") pod \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\" (UID: \"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4\") " Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.881751 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-kube-api-access-6g5bn" (OuterVolumeSpecName: "kube-api-access-6g5bn") pod "4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4" (UID: "4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4"). InnerVolumeSpecName "kube-api-access-6g5bn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.883121 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-ceph" (OuterVolumeSpecName: "ceph") pod "4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4" (UID: "4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.883933 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4" (UID: "4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.915981 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-inventory" (OuterVolumeSpecName: "inventory") pod "4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4" (UID: "4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.918925 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4" (UID: "4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.919831 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4" (UID: "4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.948186 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4" (UID: "4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.977704 4997 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.977745 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.977759 4997 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.977772 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.977783 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.977795 4997 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:05 crc kubenswrapper[4997]: I1205 09:28:05.977808 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g5bn\" (UniqueName: \"kubernetes.io/projected/4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4-kube-api-access-6g5bn\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.290274 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" event={"ID":"4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4","Type":"ContainerDied","Data":"89be7a93b7e503f3c3a4afa07d12d20dc3a35e89e53cc66aa9092561a98b8651"} Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.290688 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89be7a93b7e503f3c3a4afa07d12d20dc3a35e89e53cc66aa9092561a98b8651" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.290360 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-v7mv9" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.396059 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-5q88v"] Dec 05 09:28:06 crc kubenswrapper[4997]: E1205 09:28:06.396468 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5f43cb-d3c6-49e8-9f83-551cd860f736" containerName="neutron-metadata-openstack-openstack-networker" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.396485 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5f43cb-d3c6-49e8-9f83-551cd860f736" containerName="neutron-metadata-openstack-openstack-networker" Dec 05 09:28:06 crc kubenswrapper[4997]: E1205 09:28:06.396500 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1394f200-4e7e-4cfa-8f47-c356b385137d" containerName="extract-utilities" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.396508 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="1394f200-4e7e-4cfa-8f47-c356b385137d" containerName="extract-utilities" Dec 05 09:28:06 crc kubenswrapper[4997]: E1205 09:28:06.396527 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1394f200-4e7e-4cfa-8f47-c356b385137d" containerName="registry-server" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.396534 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="1394f200-4e7e-4cfa-8f47-c356b385137d" containerName="registry-server" Dec 05 09:28:06 crc kubenswrapper[4997]: E1205 09:28:06.396547 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4" containerName="neutron-metadata-openstack-openstack-cell1" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.396554 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4" containerName="neutron-metadata-openstack-openstack-cell1" Dec 05 09:28:06 crc kubenswrapper[4997]: E1205 09:28:06.396572 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1394f200-4e7e-4cfa-8f47-c356b385137d" containerName="extract-content" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.396581 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="1394f200-4e7e-4cfa-8f47-c356b385137d" containerName="extract-content" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.396789 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4" containerName="neutron-metadata-openstack-openstack-cell1" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.396810 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba5f43cb-d3c6-49e8-9f83-551cd860f736" containerName="neutron-metadata-openstack-openstack-networker" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.396835 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="1394f200-4e7e-4cfa-8f47-c356b385137d" containerName="registry-server" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.397475 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.402124 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.402225 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.402705 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.402893 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.403069 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.413831 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-5q88v"] Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.487907 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-inventory\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.488009 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-ssh-key\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.488051 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.488096 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm2gs\" (UniqueName: \"kubernetes.io/projected/93e1015b-7d40-4ad7-a1b8-78f9a243223e-kube-api-access-sm2gs\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.488141 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-ceph\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.488182 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.589859 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-inventory\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.590216 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-ssh-key\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.590335 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.590472 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm2gs\" (UniqueName: \"kubernetes.io/projected/93e1015b-7d40-4ad7-a1b8-78f9a243223e-kube-api-access-sm2gs\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.590638 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-ceph\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.590747 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.595992 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-inventory\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.596688 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.596832 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-ssh-key\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.596884 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-ceph\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.598783 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.615028 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm2gs\" (UniqueName: \"kubernetes.io/projected/93e1015b-7d40-4ad7-a1b8-78f9a243223e-kube-api-access-sm2gs\") pod \"libvirt-openstack-openstack-cell1-5q88v\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:06 crc kubenswrapper[4997]: I1205 09:28:06.716930 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:28:07 crc kubenswrapper[4997]: I1205 09:28:07.269555 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-5q88v"] Dec 05 09:28:07 crc kubenswrapper[4997]: I1205 09:28:07.305476 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-5q88v" event={"ID":"93e1015b-7d40-4ad7-a1b8-78f9a243223e","Type":"ContainerStarted","Data":"0597bea7eeab354f61cc6ac9061c6c68c533c821490450bf057d013d972348ff"} Dec 05 09:28:11 crc kubenswrapper[4997]: I1205 09:28:11.346304 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-5q88v" event={"ID":"93e1015b-7d40-4ad7-a1b8-78f9a243223e","Type":"ContainerStarted","Data":"0a2635e13f6a5da24bbdf19fd436c8b280601d819d12ccb08a23c355c15bdadd"} Dec 05 09:28:11 crc kubenswrapper[4997]: I1205 09:28:11.368999 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-5q88v" podStartSLOduration=2.598667479 podStartE2EDuration="5.368977709s" podCreationTimestamp="2025-12-05 09:28:06 +0000 UTC" firstStartedPulling="2025-12-05 09:28:07.281590285 +0000 UTC m=+9187.810497546" lastFinishedPulling="2025-12-05 09:28:10.051900505 +0000 UTC m=+9190.580807776" observedRunningTime="2025-12-05 09:28:11.359359209 +0000 UTC m=+9191.888266470" watchObservedRunningTime="2025-12-05 09:28:11.368977709 +0000 UTC m=+9191.897884970" Dec 05 09:30:00 crc kubenswrapper[4997]: I1205 09:30:00.156949 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc"] Dec 05 09:30:00 crc kubenswrapper[4997]: I1205 09:30:00.160243 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc" Dec 05 09:30:00 crc kubenswrapper[4997]: I1205 09:30:00.162958 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 09:30:00 crc kubenswrapper[4997]: I1205 09:30:00.162958 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 09:30:00 crc kubenswrapper[4997]: I1205 09:30:00.187093 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc"] Dec 05 09:30:00 crc kubenswrapper[4997]: I1205 09:30:00.279830 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pc2p\" (UniqueName: \"kubernetes.io/projected/115f37ed-c40f-4655-986f-a5313b62cf1e-kube-api-access-8pc2p\") pod \"collect-profiles-29415450-9tsfc\" (UID: \"115f37ed-c40f-4655-986f-a5313b62cf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc" Dec 05 09:30:00 crc kubenswrapper[4997]: I1205 09:30:00.279903 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/115f37ed-c40f-4655-986f-a5313b62cf1e-config-volume\") pod \"collect-profiles-29415450-9tsfc\" (UID: \"115f37ed-c40f-4655-986f-a5313b62cf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc" Dec 05 09:30:00 crc kubenswrapper[4997]: I1205 09:30:00.280045 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/115f37ed-c40f-4655-986f-a5313b62cf1e-secret-volume\") pod \"collect-profiles-29415450-9tsfc\" (UID: \"115f37ed-c40f-4655-986f-a5313b62cf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc" Dec 05 09:30:00 crc kubenswrapper[4997]: I1205 09:30:00.382181 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pc2p\" (UniqueName: \"kubernetes.io/projected/115f37ed-c40f-4655-986f-a5313b62cf1e-kube-api-access-8pc2p\") pod \"collect-profiles-29415450-9tsfc\" (UID: \"115f37ed-c40f-4655-986f-a5313b62cf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc" Dec 05 09:30:00 crc kubenswrapper[4997]: I1205 09:30:00.382281 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/115f37ed-c40f-4655-986f-a5313b62cf1e-config-volume\") pod \"collect-profiles-29415450-9tsfc\" (UID: \"115f37ed-c40f-4655-986f-a5313b62cf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc" Dec 05 09:30:00 crc kubenswrapper[4997]: I1205 09:30:00.382300 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/115f37ed-c40f-4655-986f-a5313b62cf1e-secret-volume\") pod \"collect-profiles-29415450-9tsfc\" (UID: \"115f37ed-c40f-4655-986f-a5313b62cf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc" Dec 05 09:30:00 crc kubenswrapper[4997]: I1205 09:30:00.383432 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/115f37ed-c40f-4655-986f-a5313b62cf1e-config-volume\") pod \"collect-profiles-29415450-9tsfc\" (UID: \"115f37ed-c40f-4655-986f-a5313b62cf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc" Dec 05 09:30:00 crc kubenswrapper[4997]: I1205 09:30:00.406276 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/115f37ed-c40f-4655-986f-a5313b62cf1e-secret-volume\") pod \"collect-profiles-29415450-9tsfc\" (UID: \"115f37ed-c40f-4655-986f-a5313b62cf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc" Dec 05 09:30:00 crc kubenswrapper[4997]: I1205 09:30:00.414492 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pc2p\" (UniqueName: \"kubernetes.io/projected/115f37ed-c40f-4655-986f-a5313b62cf1e-kube-api-access-8pc2p\") pod \"collect-profiles-29415450-9tsfc\" (UID: \"115f37ed-c40f-4655-986f-a5313b62cf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc" Dec 05 09:30:00 crc kubenswrapper[4997]: I1205 09:30:00.488800 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc" Dec 05 09:30:00 crc kubenswrapper[4997]: I1205 09:30:00.990223 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc"] Dec 05 09:30:01 crc kubenswrapper[4997]: I1205 09:30:01.541093 4997 generic.go:334] "Generic (PLEG): container finished" podID="115f37ed-c40f-4655-986f-a5313b62cf1e" containerID="37a5df49d0f5a50fd25a0baa67fc12f249c5b5565b01ea4439e4afb29b3c1b88" exitCode=0 Dec 05 09:30:01 crc kubenswrapper[4997]: I1205 09:30:01.541154 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc" event={"ID":"115f37ed-c40f-4655-986f-a5313b62cf1e","Type":"ContainerDied","Data":"37a5df49d0f5a50fd25a0baa67fc12f249c5b5565b01ea4439e4afb29b3c1b88"} Dec 05 09:30:01 crc kubenswrapper[4997]: I1205 09:30:01.541381 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc" event={"ID":"115f37ed-c40f-4655-986f-a5313b62cf1e","Type":"ContainerStarted","Data":"9c4b0170c88d05a9f7d829b3042d7cbb678a007b949a2a0fe20275f620800e83"} Dec 05 09:30:02 crc kubenswrapper[4997]: I1205 09:30:02.967710 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc" Dec 05 09:30:03 crc kubenswrapper[4997]: I1205 09:30:03.035641 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/115f37ed-c40f-4655-986f-a5313b62cf1e-config-volume\") pod \"115f37ed-c40f-4655-986f-a5313b62cf1e\" (UID: \"115f37ed-c40f-4655-986f-a5313b62cf1e\") " Dec 05 09:30:03 crc kubenswrapper[4997]: I1205 09:30:03.035896 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pc2p\" (UniqueName: \"kubernetes.io/projected/115f37ed-c40f-4655-986f-a5313b62cf1e-kube-api-access-8pc2p\") pod \"115f37ed-c40f-4655-986f-a5313b62cf1e\" (UID: \"115f37ed-c40f-4655-986f-a5313b62cf1e\") " Dec 05 09:30:03 crc kubenswrapper[4997]: I1205 09:30:03.035946 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/115f37ed-c40f-4655-986f-a5313b62cf1e-secret-volume\") pod \"115f37ed-c40f-4655-986f-a5313b62cf1e\" (UID: \"115f37ed-c40f-4655-986f-a5313b62cf1e\") " Dec 05 09:30:03 crc kubenswrapper[4997]: I1205 09:30:03.037929 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/115f37ed-c40f-4655-986f-a5313b62cf1e-config-volume" (OuterVolumeSpecName: "config-volume") pod "115f37ed-c40f-4655-986f-a5313b62cf1e" (UID: "115f37ed-c40f-4655-986f-a5313b62cf1e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:30:03 crc kubenswrapper[4997]: I1205 09:30:03.043886 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/115f37ed-c40f-4655-986f-a5313b62cf1e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "115f37ed-c40f-4655-986f-a5313b62cf1e" (UID: "115f37ed-c40f-4655-986f-a5313b62cf1e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:03 crc kubenswrapper[4997]: I1205 09:30:03.048094 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/115f37ed-c40f-4655-986f-a5313b62cf1e-kube-api-access-8pc2p" (OuterVolumeSpecName: "kube-api-access-8pc2p") pod "115f37ed-c40f-4655-986f-a5313b62cf1e" (UID: "115f37ed-c40f-4655-986f-a5313b62cf1e"). InnerVolumeSpecName "kube-api-access-8pc2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:03 crc kubenswrapper[4997]: I1205 09:30:03.138798 4997 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/115f37ed-c40f-4655-986f-a5313b62cf1e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:03 crc kubenswrapper[4997]: I1205 09:30:03.138833 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pc2p\" (UniqueName: \"kubernetes.io/projected/115f37ed-c40f-4655-986f-a5313b62cf1e-kube-api-access-8pc2p\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:03 crc kubenswrapper[4997]: I1205 09:30:03.138843 4997 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/115f37ed-c40f-4655-986f-a5313b62cf1e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:03 crc kubenswrapper[4997]: I1205 09:30:03.566503 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc" event={"ID":"115f37ed-c40f-4655-986f-a5313b62cf1e","Type":"ContainerDied","Data":"9c4b0170c88d05a9f7d829b3042d7cbb678a007b949a2a0fe20275f620800e83"} Dec 05 09:30:03 crc kubenswrapper[4997]: I1205 09:30:03.566550 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c4b0170c88d05a9f7d829b3042d7cbb678a007b949a2a0fe20275f620800e83" Dec 05 09:30:03 crc kubenswrapper[4997]: I1205 09:30:03.566663 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc" Dec 05 09:30:04 crc kubenswrapper[4997]: I1205 09:30:04.036564 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw"] Dec 05 09:30:04 crc kubenswrapper[4997]: I1205 09:30:04.046525 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415405-pvsdw"] Dec 05 09:30:05 crc kubenswrapper[4997]: I1205 09:30:05.763163 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bb99647-f591-4b3a-88ab-85ef7d051bbc" path="/var/lib/kubelet/pods/8bb99647-f591-4b3a-88ab-85ef7d051bbc/volumes" Dec 05 09:30:19 crc kubenswrapper[4997]: I1205 09:30:19.776409 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:30:19 crc kubenswrapper[4997]: I1205 09:30:19.777089 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:30:20 crc kubenswrapper[4997]: I1205 09:30:20.022810 4997 scope.go:117] "RemoveContainer" containerID="b0567b7982a7fdafe0cf620caadc10fd9e07419fe4b71fa066dad17a6481181c" Dec 05 09:30:37 crc kubenswrapper[4997]: I1205 09:30:37.624468 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lb92z"] Dec 05 09:30:37 crc kubenswrapper[4997]: E1205 09:30:37.625562 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="115f37ed-c40f-4655-986f-a5313b62cf1e" containerName="collect-profiles" Dec 05 09:30:37 crc kubenswrapper[4997]: I1205 09:30:37.625576 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="115f37ed-c40f-4655-986f-a5313b62cf1e" containerName="collect-profiles" Dec 05 09:30:37 crc kubenswrapper[4997]: I1205 09:30:37.625835 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="115f37ed-c40f-4655-986f-a5313b62cf1e" containerName="collect-profiles" Dec 05 09:30:37 crc kubenswrapper[4997]: I1205 09:30:37.627586 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb92z" Dec 05 09:30:37 crc kubenswrapper[4997]: I1205 09:30:37.635226 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb92z"] Dec 05 09:30:37 crc kubenswrapper[4997]: I1205 09:30:37.790824 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/910dc6a9-6016-4b2f-8174-532df92b9777-utilities\") pod \"redhat-marketplace-lb92z\" (UID: \"910dc6a9-6016-4b2f-8174-532df92b9777\") " pod="openshift-marketplace/redhat-marketplace-lb92z" Dec 05 09:30:37 crc kubenswrapper[4997]: I1205 09:30:37.790909 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9p4b\" (UniqueName: \"kubernetes.io/projected/910dc6a9-6016-4b2f-8174-532df92b9777-kube-api-access-c9p4b\") pod \"redhat-marketplace-lb92z\" (UID: \"910dc6a9-6016-4b2f-8174-532df92b9777\") " pod="openshift-marketplace/redhat-marketplace-lb92z" Dec 05 09:30:37 crc kubenswrapper[4997]: I1205 09:30:37.790988 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/910dc6a9-6016-4b2f-8174-532df92b9777-catalog-content\") pod \"redhat-marketplace-lb92z\" (UID: \"910dc6a9-6016-4b2f-8174-532df92b9777\") " pod="openshift-marketplace/redhat-marketplace-lb92z" Dec 05 09:30:37 crc kubenswrapper[4997]: I1205 09:30:37.892473 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9p4b\" (UniqueName: \"kubernetes.io/projected/910dc6a9-6016-4b2f-8174-532df92b9777-kube-api-access-c9p4b\") pod \"redhat-marketplace-lb92z\" (UID: \"910dc6a9-6016-4b2f-8174-532df92b9777\") " pod="openshift-marketplace/redhat-marketplace-lb92z" Dec 05 09:30:37 crc kubenswrapper[4997]: I1205 09:30:37.893179 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/910dc6a9-6016-4b2f-8174-532df92b9777-catalog-content\") pod \"redhat-marketplace-lb92z\" (UID: \"910dc6a9-6016-4b2f-8174-532df92b9777\") " pod="openshift-marketplace/redhat-marketplace-lb92z" Dec 05 09:30:37 crc kubenswrapper[4997]: I1205 09:30:37.893531 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/910dc6a9-6016-4b2f-8174-532df92b9777-utilities\") pod \"redhat-marketplace-lb92z\" (UID: \"910dc6a9-6016-4b2f-8174-532df92b9777\") " pod="openshift-marketplace/redhat-marketplace-lb92z" Dec 05 09:30:37 crc kubenswrapper[4997]: I1205 09:30:37.893658 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/910dc6a9-6016-4b2f-8174-532df92b9777-catalog-content\") pod \"redhat-marketplace-lb92z\" (UID: \"910dc6a9-6016-4b2f-8174-532df92b9777\") " pod="openshift-marketplace/redhat-marketplace-lb92z" Dec 05 09:30:37 crc kubenswrapper[4997]: I1205 09:30:37.894786 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/910dc6a9-6016-4b2f-8174-532df92b9777-utilities\") pod \"redhat-marketplace-lb92z\" (UID: \"910dc6a9-6016-4b2f-8174-532df92b9777\") " pod="openshift-marketplace/redhat-marketplace-lb92z" Dec 05 09:30:37 crc kubenswrapper[4997]: I1205 09:30:37.915137 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9p4b\" (UniqueName: \"kubernetes.io/projected/910dc6a9-6016-4b2f-8174-532df92b9777-kube-api-access-c9p4b\") pod \"redhat-marketplace-lb92z\" (UID: \"910dc6a9-6016-4b2f-8174-532df92b9777\") " pod="openshift-marketplace/redhat-marketplace-lb92z" Dec 05 09:30:37 crc kubenswrapper[4997]: I1205 09:30:37.966124 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb92z" Dec 05 09:30:38 crc kubenswrapper[4997]: I1205 09:30:38.438132 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb92z"] Dec 05 09:30:38 crc kubenswrapper[4997]: I1205 09:30:38.989342 4997 generic.go:334] "Generic (PLEG): container finished" podID="910dc6a9-6016-4b2f-8174-532df92b9777" containerID="dd78093603be890862ecf79e8036f0e123e8862c7ffe472b2b3d5e401e206859" exitCode=0 Dec 05 09:30:38 crc kubenswrapper[4997]: I1205 09:30:38.989494 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb92z" event={"ID":"910dc6a9-6016-4b2f-8174-532df92b9777","Type":"ContainerDied","Data":"dd78093603be890862ecf79e8036f0e123e8862c7ffe472b2b3d5e401e206859"} Dec 05 09:30:38 crc kubenswrapper[4997]: I1205 09:30:38.989698 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb92z" event={"ID":"910dc6a9-6016-4b2f-8174-532df92b9777","Type":"ContainerStarted","Data":"eb46b1219bbc17ce4693f9410a2cb97a23529a1b25b1f87be1593ba58f01c830"} Dec 05 09:30:40 crc kubenswrapper[4997]: I1205 09:30:40.000824 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb92z" event={"ID":"910dc6a9-6016-4b2f-8174-532df92b9777","Type":"ContainerStarted","Data":"37b393bb59f82e4cd783b874b180b959e31d45210f037b3413cd0e5b66b893b9"} Dec 05 09:30:41 crc kubenswrapper[4997]: I1205 09:30:41.014209 4997 generic.go:334] "Generic (PLEG): container finished" podID="910dc6a9-6016-4b2f-8174-532df92b9777" containerID="37b393bb59f82e4cd783b874b180b959e31d45210f037b3413cd0e5b66b893b9" exitCode=0 Dec 05 09:30:41 crc kubenswrapper[4997]: I1205 09:30:41.014252 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb92z" event={"ID":"910dc6a9-6016-4b2f-8174-532df92b9777","Type":"ContainerDied","Data":"37b393bb59f82e4cd783b874b180b959e31d45210f037b3413cd0e5b66b893b9"} Dec 05 09:30:42 crc kubenswrapper[4997]: I1205 09:30:42.026836 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb92z" event={"ID":"910dc6a9-6016-4b2f-8174-532df92b9777","Type":"ContainerStarted","Data":"a9656cff2109182539ea84cae180672da20ef073d7e6bc19881f1b2b6198dd57"} Dec 05 09:30:42 crc kubenswrapper[4997]: I1205 09:30:42.047950 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lb92z" podStartSLOduration=2.616694202 podStartE2EDuration="5.047929558s" podCreationTimestamp="2025-12-05 09:30:37 +0000 UTC" firstStartedPulling="2025-12-05 09:30:38.99186357 +0000 UTC m=+9339.520770851" lastFinishedPulling="2025-12-05 09:30:41.423098946 +0000 UTC m=+9341.952006207" observedRunningTime="2025-12-05 09:30:42.043451107 +0000 UTC m=+9342.572358378" watchObservedRunningTime="2025-12-05 09:30:42.047929558 +0000 UTC m=+9342.576836819" Dec 05 09:30:47 crc kubenswrapper[4997]: I1205 09:30:47.966486 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lb92z" Dec 05 09:30:47 crc kubenswrapper[4997]: I1205 09:30:47.967066 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lb92z" Dec 05 09:30:48 crc kubenswrapper[4997]: I1205 09:30:48.016201 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lb92z" Dec 05 09:30:48 crc kubenswrapper[4997]: I1205 09:30:48.140428 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lb92z" Dec 05 09:30:48 crc kubenswrapper[4997]: I1205 09:30:48.254282 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb92z"] Dec 05 09:30:49 crc kubenswrapper[4997]: I1205 09:30:49.770220 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:30:49 crc kubenswrapper[4997]: I1205 09:30:49.770478 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:30:50 crc kubenswrapper[4997]: I1205 09:30:50.115876 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lb92z" podUID="910dc6a9-6016-4b2f-8174-532df92b9777" containerName="registry-server" containerID="cri-o://a9656cff2109182539ea84cae180672da20ef073d7e6bc19881f1b2b6198dd57" gracePeriod=2 Dec 05 09:30:50 crc kubenswrapper[4997]: I1205 09:30:50.613308 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb92z" Dec 05 09:30:50 crc kubenswrapper[4997]: I1205 09:30:50.710077 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/910dc6a9-6016-4b2f-8174-532df92b9777-catalog-content\") pod \"910dc6a9-6016-4b2f-8174-532df92b9777\" (UID: \"910dc6a9-6016-4b2f-8174-532df92b9777\") " Dec 05 09:30:50 crc kubenswrapper[4997]: I1205 09:30:50.710133 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9p4b\" (UniqueName: \"kubernetes.io/projected/910dc6a9-6016-4b2f-8174-532df92b9777-kube-api-access-c9p4b\") pod \"910dc6a9-6016-4b2f-8174-532df92b9777\" (UID: \"910dc6a9-6016-4b2f-8174-532df92b9777\") " Dec 05 09:30:50 crc kubenswrapper[4997]: I1205 09:30:50.710269 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/910dc6a9-6016-4b2f-8174-532df92b9777-utilities\") pod \"910dc6a9-6016-4b2f-8174-532df92b9777\" (UID: \"910dc6a9-6016-4b2f-8174-532df92b9777\") " Dec 05 09:30:50 crc kubenswrapper[4997]: I1205 09:30:50.711223 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/910dc6a9-6016-4b2f-8174-532df92b9777-utilities" (OuterVolumeSpecName: "utilities") pod "910dc6a9-6016-4b2f-8174-532df92b9777" (UID: "910dc6a9-6016-4b2f-8174-532df92b9777"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:30:50 crc kubenswrapper[4997]: I1205 09:30:50.715331 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/910dc6a9-6016-4b2f-8174-532df92b9777-kube-api-access-c9p4b" (OuterVolumeSpecName: "kube-api-access-c9p4b") pod "910dc6a9-6016-4b2f-8174-532df92b9777" (UID: "910dc6a9-6016-4b2f-8174-532df92b9777"). InnerVolumeSpecName "kube-api-access-c9p4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:50 crc kubenswrapper[4997]: I1205 09:30:50.727164 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/910dc6a9-6016-4b2f-8174-532df92b9777-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "910dc6a9-6016-4b2f-8174-532df92b9777" (UID: "910dc6a9-6016-4b2f-8174-532df92b9777"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:30:50 crc kubenswrapper[4997]: I1205 09:30:50.812550 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/910dc6a9-6016-4b2f-8174-532df92b9777-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:50 crc kubenswrapper[4997]: I1205 09:30:50.812907 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/910dc6a9-6016-4b2f-8174-532df92b9777-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:50 crc kubenswrapper[4997]: I1205 09:30:50.812923 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9p4b\" (UniqueName: \"kubernetes.io/projected/910dc6a9-6016-4b2f-8174-532df92b9777-kube-api-access-c9p4b\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:51 crc kubenswrapper[4997]: I1205 09:30:51.130079 4997 generic.go:334] "Generic (PLEG): container finished" podID="910dc6a9-6016-4b2f-8174-532df92b9777" containerID="a9656cff2109182539ea84cae180672da20ef073d7e6bc19881f1b2b6198dd57" exitCode=0 Dec 05 09:30:51 crc kubenswrapper[4997]: I1205 09:30:51.130255 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb92z" Dec 05 09:30:51 crc kubenswrapper[4997]: I1205 09:30:51.130468 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb92z" event={"ID":"910dc6a9-6016-4b2f-8174-532df92b9777","Type":"ContainerDied","Data":"a9656cff2109182539ea84cae180672da20ef073d7e6bc19881f1b2b6198dd57"} Dec 05 09:30:51 crc kubenswrapper[4997]: I1205 09:30:51.130596 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb92z" event={"ID":"910dc6a9-6016-4b2f-8174-532df92b9777","Type":"ContainerDied","Data":"eb46b1219bbc17ce4693f9410a2cb97a23529a1b25b1f87be1593ba58f01c830"} Dec 05 09:30:51 crc kubenswrapper[4997]: I1205 09:30:51.130649 4997 scope.go:117] "RemoveContainer" containerID="a9656cff2109182539ea84cae180672da20ef073d7e6bc19881f1b2b6198dd57" Dec 05 09:30:51 crc kubenswrapper[4997]: I1205 09:30:51.157435 4997 scope.go:117] "RemoveContainer" containerID="37b393bb59f82e4cd783b874b180b959e31d45210f037b3413cd0e5b66b893b9" Dec 05 09:30:51 crc kubenswrapper[4997]: I1205 09:30:51.178100 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb92z"] Dec 05 09:30:51 crc kubenswrapper[4997]: I1205 09:30:51.186645 4997 scope.go:117] "RemoveContainer" containerID="dd78093603be890862ecf79e8036f0e123e8862c7ffe472b2b3d5e401e206859" Dec 05 09:30:51 crc kubenswrapper[4997]: I1205 09:30:51.188540 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb92z"] Dec 05 09:30:51 crc kubenswrapper[4997]: I1205 09:30:51.236460 4997 scope.go:117] "RemoveContainer" containerID="a9656cff2109182539ea84cae180672da20ef073d7e6bc19881f1b2b6198dd57" Dec 05 09:30:51 crc kubenswrapper[4997]: E1205 09:30:51.236937 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9656cff2109182539ea84cae180672da20ef073d7e6bc19881f1b2b6198dd57\": container with ID starting with a9656cff2109182539ea84cae180672da20ef073d7e6bc19881f1b2b6198dd57 not found: ID does not exist" containerID="a9656cff2109182539ea84cae180672da20ef073d7e6bc19881f1b2b6198dd57" Dec 05 09:30:51 crc kubenswrapper[4997]: I1205 09:30:51.237054 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9656cff2109182539ea84cae180672da20ef073d7e6bc19881f1b2b6198dd57"} err="failed to get container status \"a9656cff2109182539ea84cae180672da20ef073d7e6bc19881f1b2b6198dd57\": rpc error: code = NotFound desc = could not find container \"a9656cff2109182539ea84cae180672da20ef073d7e6bc19881f1b2b6198dd57\": container with ID starting with a9656cff2109182539ea84cae180672da20ef073d7e6bc19881f1b2b6198dd57 not found: ID does not exist" Dec 05 09:30:51 crc kubenswrapper[4997]: I1205 09:30:51.237143 4997 scope.go:117] "RemoveContainer" containerID="37b393bb59f82e4cd783b874b180b959e31d45210f037b3413cd0e5b66b893b9" Dec 05 09:30:51 crc kubenswrapper[4997]: E1205 09:30:51.238247 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37b393bb59f82e4cd783b874b180b959e31d45210f037b3413cd0e5b66b893b9\": container with ID starting with 37b393bb59f82e4cd783b874b180b959e31d45210f037b3413cd0e5b66b893b9 not found: ID does not exist" containerID="37b393bb59f82e4cd783b874b180b959e31d45210f037b3413cd0e5b66b893b9" Dec 05 09:30:51 crc kubenswrapper[4997]: I1205 09:30:51.238406 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37b393bb59f82e4cd783b874b180b959e31d45210f037b3413cd0e5b66b893b9"} err="failed to get container status \"37b393bb59f82e4cd783b874b180b959e31d45210f037b3413cd0e5b66b893b9\": rpc error: code = NotFound desc = could not find container \"37b393bb59f82e4cd783b874b180b959e31d45210f037b3413cd0e5b66b893b9\": container with ID starting with 37b393bb59f82e4cd783b874b180b959e31d45210f037b3413cd0e5b66b893b9 not found: ID does not exist" Dec 05 09:30:51 crc kubenswrapper[4997]: I1205 09:30:51.238437 4997 scope.go:117] "RemoveContainer" containerID="dd78093603be890862ecf79e8036f0e123e8862c7ffe472b2b3d5e401e206859" Dec 05 09:30:51 crc kubenswrapper[4997]: E1205 09:30:51.239709 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd78093603be890862ecf79e8036f0e123e8862c7ffe472b2b3d5e401e206859\": container with ID starting with dd78093603be890862ecf79e8036f0e123e8862c7ffe472b2b3d5e401e206859 not found: ID does not exist" containerID="dd78093603be890862ecf79e8036f0e123e8862c7ffe472b2b3d5e401e206859" Dec 05 09:30:51 crc kubenswrapper[4997]: I1205 09:30:51.239743 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd78093603be890862ecf79e8036f0e123e8862c7ffe472b2b3d5e401e206859"} err="failed to get container status \"dd78093603be890862ecf79e8036f0e123e8862c7ffe472b2b3d5e401e206859\": rpc error: code = NotFound desc = could not find container \"dd78093603be890862ecf79e8036f0e123e8862c7ffe472b2b3d5e401e206859\": container with ID starting with dd78093603be890862ecf79e8036f0e123e8862c7ffe472b2b3d5e401e206859 not found: ID does not exist" Dec 05 09:30:51 crc kubenswrapper[4997]: I1205 09:30:51.766173 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="910dc6a9-6016-4b2f-8174-532df92b9777" path="/var/lib/kubelet/pods/910dc6a9-6016-4b2f-8174-532df92b9777/volumes" Dec 05 09:31:19 crc kubenswrapper[4997]: I1205 09:31:19.770755 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:31:19 crc kubenswrapper[4997]: I1205 09:31:19.771359 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:31:19 crc kubenswrapper[4997]: I1205 09:31:19.771420 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 09:31:19 crc kubenswrapper[4997]: I1205 09:31:19.772356 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6ce1c08e479b9a4b91cc9213507822cb7d19ab1ed1dfebf53f033a419412ffa0"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:31:19 crc kubenswrapper[4997]: I1205 09:31:19.772416 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://6ce1c08e479b9a4b91cc9213507822cb7d19ab1ed1dfebf53f033a419412ffa0" gracePeriod=600 Dec 05 09:31:19 crc kubenswrapper[4997]: E1205 09:31:19.886733 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28cde6da_1de5_455a_93c8_dd250f9ea4d0.slice/crio-conmon-6ce1c08e479b9a4b91cc9213507822cb7d19ab1ed1dfebf53f033a419412ffa0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28cde6da_1de5_455a_93c8_dd250f9ea4d0.slice/crio-6ce1c08e479b9a4b91cc9213507822cb7d19ab1ed1dfebf53f033a419412ffa0.scope\": RecentStats: unable to find data in memory cache]" Dec 05 09:31:20 crc kubenswrapper[4997]: I1205 09:31:20.505669 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="6ce1c08e479b9a4b91cc9213507822cb7d19ab1ed1dfebf53f033a419412ffa0" exitCode=0 Dec 05 09:31:20 crc kubenswrapper[4997]: I1205 09:31:20.505753 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"6ce1c08e479b9a4b91cc9213507822cb7d19ab1ed1dfebf53f033a419412ffa0"} Dec 05 09:31:20 crc kubenswrapper[4997]: I1205 09:31:20.506341 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06"} Dec 05 09:31:20 crc kubenswrapper[4997]: I1205 09:31:20.506381 4997 scope.go:117] "RemoveContainer" containerID="7893644b833e0404ab0bb201634e72108d4cc2a4e47b44e5b5d5c9c6062f499a" Dec 05 09:31:22 crc kubenswrapper[4997]: I1205 09:31:22.652221 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gghdj"] Dec 05 09:31:22 crc kubenswrapper[4997]: E1205 09:31:22.653294 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="910dc6a9-6016-4b2f-8174-532df92b9777" containerName="registry-server" Dec 05 09:31:22 crc kubenswrapper[4997]: I1205 09:31:22.653317 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="910dc6a9-6016-4b2f-8174-532df92b9777" containerName="registry-server" Dec 05 09:31:22 crc kubenswrapper[4997]: E1205 09:31:22.653334 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="910dc6a9-6016-4b2f-8174-532df92b9777" containerName="extract-content" Dec 05 09:31:22 crc kubenswrapper[4997]: I1205 09:31:22.653342 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="910dc6a9-6016-4b2f-8174-532df92b9777" containerName="extract-content" Dec 05 09:31:22 crc kubenswrapper[4997]: E1205 09:31:22.653376 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="910dc6a9-6016-4b2f-8174-532df92b9777" containerName="extract-utilities" Dec 05 09:31:22 crc kubenswrapper[4997]: I1205 09:31:22.653385 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="910dc6a9-6016-4b2f-8174-532df92b9777" containerName="extract-utilities" Dec 05 09:31:22 crc kubenswrapper[4997]: I1205 09:31:22.653688 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="910dc6a9-6016-4b2f-8174-532df92b9777" containerName="registry-server" Dec 05 09:31:22 crc kubenswrapper[4997]: I1205 09:31:22.655526 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gghdj" Dec 05 09:31:22 crc kubenswrapper[4997]: I1205 09:31:22.702685 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gghdj"] Dec 05 09:31:22 crc kubenswrapper[4997]: I1205 09:31:22.790151 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/649fa519-39dd-4bb3-abef-12c1de34a7f0-catalog-content\") pod \"community-operators-gghdj\" (UID: \"649fa519-39dd-4bb3-abef-12c1de34a7f0\") " pod="openshift-marketplace/community-operators-gghdj" Dec 05 09:31:22 crc kubenswrapper[4997]: I1205 09:31:22.790246 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/649fa519-39dd-4bb3-abef-12c1de34a7f0-utilities\") pod \"community-operators-gghdj\" (UID: \"649fa519-39dd-4bb3-abef-12c1de34a7f0\") " pod="openshift-marketplace/community-operators-gghdj" Dec 05 09:31:22 crc kubenswrapper[4997]: I1205 09:31:22.790425 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8tmh\" (UniqueName: \"kubernetes.io/projected/649fa519-39dd-4bb3-abef-12c1de34a7f0-kube-api-access-z8tmh\") pod \"community-operators-gghdj\" (UID: \"649fa519-39dd-4bb3-abef-12c1de34a7f0\") " pod="openshift-marketplace/community-operators-gghdj" Dec 05 09:31:22 crc kubenswrapper[4997]: I1205 09:31:22.892124 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/649fa519-39dd-4bb3-abef-12c1de34a7f0-catalog-content\") pod \"community-operators-gghdj\" (UID: \"649fa519-39dd-4bb3-abef-12c1de34a7f0\") " pod="openshift-marketplace/community-operators-gghdj" Dec 05 09:31:22 crc kubenswrapper[4997]: I1205 09:31:22.892283 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/649fa519-39dd-4bb3-abef-12c1de34a7f0-utilities\") pod \"community-operators-gghdj\" (UID: \"649fa519-39dd-4bb3-abef-12c1de34a7f0\") " pod="openshift-marketplace/community-operators-gghdj" Dec 05 09:31:22 crc kubenswrapper[4997]: I1205 09:31:22.892451 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8tmh\" (UniqueName: \"kubernetes.io/projected/649fa519-39dd-4bb3-abef-12c1de34a7f0-kube-api-access-z8tmh\") pod \"community-operators-gghdj\" (UID: \"649fa519-39dd-4bb3-abef-12c1de34a7f0\") " pod="openshift-marketplace/community-operators-gghdj" Dec 05 09:31:22 crc kubenswrapper[4997]: I1205 09:31:22.892827 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/649fa519-39dd-4bb3-abef-12c1de34a7f0-catalog-content\") pod \"community-operators-gghdj\" (UID: \"649fa519-39dd-4bb3-abef-12c1de34a7f0\") " pod="openshift-marketplace/community-operators-gghdj" Dec 05 09:31:22 crc kubenswrapper[4997]: I1205 09:31:22.893413 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/649fa519-39dd-4bb3-abef-12c1de34a7f0-utilities\") pod \"community-operators-gghdj\" (UID: \"649fa519-39dd-4bb3-abef-12c1de34a7f0\") " pod="openshift-marketplace/community-operators-gghdj" Dec 05 09:31:22 crc kubenswrapper[4997]: I1205 09:31:22.913505 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8tmh\" (UniqueName: \"kubernetes.io/projected/649fa519-39dd-4bb3-abef-12c1de34a7f0-kube-api-access-z8tmh\") pod \"community-operators-gghdj\" (UID: \"649fa519-39dd-4bb3-abef-12c1de34a7f0\") " pod="openshift-marketplace/community-operators-gghdj" Dec 05 09:31:23 crc kubenswrapper[4997]: I1205 09:31:23.001238 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gghdj" Dec 05 09:31:23 crc kubenswrapper[4997]: I1205 09:31:23.567427 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gghdj"] Dec 05 09:31:24 crc kubenswrapper[4997]: I1205 09:31:24.546840 4997 generic.go:334] "Generic (PLEG): container finished" podID="649fa519-39dd-4bb3-abef-12c1de34a7f0" containerID="c1b971d46ead675069d18e0a746efb4fd365d564232961db1658f37ed7c35627" exitCode=0 Dec 05 09:31:24 crc kubenswrapper[4997]: I1205 09:31:24.546923 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gghdj" event={"ID":"649fa519-39dd-4bb3-abef-12c1de34a7f0","Type":"ContainerDied","Data":"c1b971d46ead675069d18e0a746efb4fd365d564232961db1658f37ed7c35627"} Dec 05 09:31:24 crc kubenswrapper[4997]: I1205 09:31:24.547089 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gghdj" event={"ID":"649fa519-39dd-4bb3-abef-12c1de34a7f0","Type":"ContainerStarted","Data":"aef26f46f1d40c02554345cc563787f15b0981f2b359b35f35f06d342f77ca4b"} Dec 05 09:31:25 crc kubenswrapper[4997]: I1205 09:31:25.560309 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gghdj" event={"ID":"649fa519-39dd-4bb3-abef-12c1de34a7f0","Type":"ContainerStarted","Data":"f71af64e556798ec34f354e3e4c9b2802f1da8dba586eeaecac327caf794b9d0"} Dec 05 09:31:26 crc kubenswrapper[4997]: I1205 09:31:26.570539 4997 generic.go:334] "Generic (PLEG): container finished" podID="649fa519-39dd-4bb3-abef-12c1de34a7f0" containerID="f71af64e556798ec34f354e3e4c9b2802f1da8dba586eeaecac327caf794b9d0" exitCode=0 Dec 05 09:31:26 crc kubenswrapper[4997]: I1205 09:31:26.570635 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gghdj" event={"ID":"649fa519-39dd-4bb3-abef-12c1de34a7f0","Type":"ContainerDied","Data":"f71af64e556798ec34f354e3e4c9b2802f1da8dba586eeaecac327caf794b9d0"} Dec 05 09:31:27 crc kubenswrapper[4997]: I1205 09:31:27.583832 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gghdj" event={"ID":"649fa519-39dd-4bb3-abef-12c1de34a7f0","Type":"ContainerStarted","Data":"4351ec5432cae19b896e868ba83b1f540857f792adb13bbdb6b46e1e887a4031"} Dec 05 09:31:27 crc kubenswrapper[4997]: I1205 09:31:27.610353 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gghdj" podStartSLOduration=3.2078267289999998 podStartE2EDuration="5.610328209s" podCreationTimestamp="2025-12-05 09:31:22 +0000 UTC" firstStartedPulling="2025-12-05 09:31:24.548833045 +0000 UTC m=+9385.077740306" lastFinishedPulling="2025-12-05 09:31:26.951334525 +0000 UTC m=+9387.480241786" observedRunningTime="2025-12-05 09:31:27.604261995 +0000 UTC m=+9388.133169286" watchObservedRunningTime="2025-12-05 09:31:27.610328209 +0000 UTC m=+9388.139235470" Dec 05 09:31:33 crc kubenswrapper[4997]: I1205 09:31:33.001559 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gghdj" Dec 05 09:31:33 crc kubenswrapper[4997]: I1205 09:31:33.002137 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gghdj" Dec 05 09:31:33 crc kubenswrapper[4997]: I1205 09:31:33.059097 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gghdj" Dec 05 09:31:33 crc kubenswrapper[4997]: I1205 09:31:33.703784 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gghdj" Dec 05 09:31:33 crc kubenswrapper[4997]: I1205 09:31:33.765888 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gghdj"] Dec 05 09:31:35 crc kubenswrapper[4997]: I1205 09:31:35.667245 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gghdj" podUID="649fa519-39dd-4bb3-abef-12c1de34a7f0" containerName="registry-server" containerID="cri-o://4351ec5432cae19b896e868ba83b1f540857f792adb13bbdb6b46e1e887a4031" gracePeriod=2 Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.205076 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gghdj" Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.268562 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/649fa519-39dd-4bb3-abef-12c1de34a7f0-utilities\") pod \"649fa519-39dd-4bb3-abef-12c1de34a7f0\" (UID: \"649fa519-39dd-4bb3-abef-12c1de34a7f0\") " Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.268730 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/649fa519-39dd-4bb3-abef-12c1de34a7f0-catalog-content\") pod \"649fa519-39dd-4bb3-abef-12c1de34a7f0\" (UID: \"649fa519-39dd-4bb3-abef-12c1de34a7f0\") " Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.268794 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8tmh\" (UniqueName: \"kubernetes.io/projected/649fa519-39dd-4bb3-abef-12c1de34a7f0-kube-api-access-z8tmh\") pod \"649fa519-39dd-4bb3-abef-12c1de34a7f0\" (UID: \"649fa519-39dd-4bb3-abef-12c1de34a7f0\") " Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.269389 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/649fa519-39dd-4bb3-abef-12c1de34a7f0-utilities" (OuterVolumeSpecName: "utilities") pod "649fa519-39dd-4bb3-abef-12c1de34a7f0" (UID: "649fa519-39dd-4bb3-abef-12c1de34a7f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.277481 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/649fa519-39dd-4bb3-abef-12c1de34a7f0-kube-api-access-z8tmh" (OuterVolumeSpecName: "kube-api-access-z8tmh") pod "649fa519-39dd-4bb3-abef-12c1de34a7f0" (UID: "649fa519-39dd-4bb3-abef-12c1de34a7f0"). InnerVolumeSpecName "kube-api-access-z8tmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.337050 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/649fa519-39dd-4bb3-abef-12c1de34a7f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "649fa519-39dd-4bb3-abef-12c1de34a7f0" (UID: "649fa519-39dd-4bb3-abef-12c1de34a7f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.372713 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/649fa519-39dd-4bb3-abef-12c1de34a7f0-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.372755 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/649fa519-39dd-4bb3-abef-12c1de34a7f0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.372773 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8tmh\" (UniqueName: \"kubernetes.io/projected/649fa519-39dd-4bb3-abef-12c1de34a7f0-kube-api-access-z8tmh\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.679913 4997 generic.go:334] "Generic (PLEG): container finished" podID="649fa519-39dd-4bb3-abef-12c1de34a7f0" containerID="4351ec5432cae19b896e868ba83b1f540857f792adb13bbdb6b46e1e887a4031" exitCode=0 Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.679955 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gghdj" event={"ID":"649fa519-39dd-4bb3-abef-12c1de34a7f0","Type":"ContainerDied","Data":"4351ec5432cae19b896e868ba83b1f540857f792adb13bbdb6b46e1e887a4031"} Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.679974 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gghdj" Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.680052 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gghdj" event={"ID":"649fa519-39dd-4bb3-abef-12c1de34a7f0","Type":"ContainerDied","Data":"aef26f46f1d40c02554345cc563787f15b0981f2b359b35f35f06d342f77ca4b"} Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.680095 4997 scope.go:117] "RemoveContainer" containerID="4351ec5432cae19b896e868ba83b1f540857f792adb13bbdb6b46e1e887a4031" Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.717970 4997 scope.go:117] "RemoveContainer" containerID="f71af64e556798ec34f354e3e4c9b2802f1da8dba586eeaecac327caf794b9d0" Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.730345 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gghdj"] Dec 05 09:31:36 crc kubenswrapper[4997]: I1205 09:31:36.742167 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gghdj"] Dec 05 09:31:37 crc kubenswrapper[4997]: I1205 09:31:37.119717 4997 scope.go:117] "RemoveContainer" containerID="c1b971d46ead675069d18e0a746efb4fd365d564232961db1658f37ed7c35627" Dec 05 09:31:37 crc kubenswrapper[4997]: I1205 09:31:37.192662 4997 scope.go:117] "RemoveContainer" containerID="4351ec5432cae19b896e868ba83b1f540857f792adb13bbdb6b46e1e887a4031" Dec 05 09:31:37 crc kubenswrapper[4997]: E1205 09:31:37.196916 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4351ec5432cae19b896e868ba83b1f540857f792adb13bbdb6b46e1e887a4031\": container with ID starting with 4351ec5432cae19b896e868ba83b1f540857f792adb13bbdb6b46e1e887a4031 not found: ID does not exist" containerID="4351ec5432cae19b896e868ba83b1f540857f792adb13bbdb6b46e1e887a4031" Dec 05 09:31:37 crc kubenswrapper[4997]: I1205 09:31:37.196978 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4351ec5432cae19b896e868ba83b1f540857f792adb13bbdb6b46e1e887a4031"} err="failed to get container status \"4351ec5432cae19b896e868ba83b1f540857f792adb13bbdb6b46e1e887a4031\": rpc error: code = NotFound desc = could not find container \"4351ec5432cae19b896e868ba83b1f540857f792adb13bbdb6b46e1e887a4031\": container with ID starting with 4351ec5432cae19b896e868ba83b1f540857f792adb13bbdb6b46e1e887a4031 not found: ID does not exist" Dec 05 09:31:37 crc kubenswrapper[4997]: I1205 09:31:37.197009 4997 scope.go:117] "RemoveContainer" containerID="f71af64e556798ec34f354e3e4c9b2802f1da8dba586eeaecac327caf794b9d0" Dec 05 09:31:37 crc kubenswrapper[4997]: E1205 09:31:37.197450 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f71af64e556798ec34f354e3e4c9b2802f1da8dba586eeaecac327caf794b9d0\": container with ID starting with f71af64e556798ec34f354e3e4c9b2802f1da8dba586eeaecac327caf794b9d0 not found: ID does not exist" containerID="f71af64e556798ec34f354e3e4c9b2802f1da8dba586eeaecac327caf794b9d0" Dec 05 09:31:37 crc kubenswrapper[4997]: I1205 09:31:37.197501 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f71af64e556798ec34f354e3e4c9b2802f1da8dba586eeaecac327caf794b9d0"} err="failed to get container status \"f71af64e556798ec34f354e3e4c9b2802f1da8dba586eeaecac327caf794b9d0\": rpc error: code = NotFound desc = could not find container \"f71af64e556798ec34f354e3e4c9b2802f1da8dba586eeaecac327caf794b9d0\": container with ID starting with f71af64e556798ec34f354e3e4c9b2802f1da8dba586eeaecac327caf794b9d0 not found: ID does not exist" Dec 05 09:31:37 crc kubenswrapper[4997]: I1205 09:31:37.197517 4997 scope.go:117] "RemoveContainer" containerID="c1b971d46ead675069d18e0a746efb4fd365d564232961db1658f37ed7c35627" Dec 05 09:31:37 crc kubenswrapper[4997]: E1205 09:31:37.197865 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1b971d46ead675069d18e0a746efb4fd365d564232961db1658f37ed7c35627\": container with ID starting with c1b971d46ead675069d18e0a746efb4fd365d564232961db1658f37ed7c35627 not found: ID does not exist" containerID="c1b971d46ead675069d18e0a746efb4fd365d564232961db1658f37ed7c35627" Dec 05 09:31:37 crc kubenswrapper[4997]: I1205 09:31:37.197891 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1b971d46ead675069d18e0a746efb4fd365d564232961db1658f37ed7c35627"} err="failed to get container status \"c1b971d46ead675069d18e0a746efb4fd365d564232961db1658f37ed7c35627\": rpc error: code = NotFound desc = could not find container \"c1b971d46ead675069d18e0a746efb4fd365d564232961db1658f37ed7c35627\": container with ID starting with c1b971d46ead675069d18e0a746efb4fd365d564232961db1658f37ed7c35627 not found: ID does not exist" Dec 05 09:31:37 crc kubenswrapper[4997]: I1205 09:31:37.760518 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="649fa519-39dd-4bb3-abef-12c1de34a7f0" path="/var/lib/kubelet/pods/649fa519-39dd-4bb3-abef-12c1de34a7f0/volumes" Dec 05 09:33:23 crc kubenswrapper[4997]: I1205 09:33:23.871334 4997 generic.go:334] "Generic (PLEG): container finished" podID="93e1015b-7d40-4ad7-a1b8-78f9a243223e" containerID="0a2635e13f6a5da24bbdf19fd436c8b280601d819d12ccb08a23c355c15bdadd" exitCode=0 Dec 05 09:33:23 crc kubenswrapper[4997]: I1205 09:33:23.871411 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-5q88v" event={"ID":"93e1015b-7d40-4ad7-a1b8-78f9a243223e","Type":"ContainerDied","Data":"0a2635e13f6a5da24bbdf19fd436c8b280601d819d12ccb08a23c355c15bdadd"} Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.471325 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.598555 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-ssh-key\") pod \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.598729 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-libvirt-combined-ca-bundle\") pod \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.598848 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-libvirt-secret-0\") pod \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.598999 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-inventory\") pod \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.599043 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-ceph\") pod \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.599153 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm2gs\" (UniqueName: \"kubernetes.io/projected/93e1015b-7d40-4ad7-a1b8-78f9a243223e-kube-api-access-sm2gs\") pod \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\" (UID: \"93e1015b-7d40-4ad7-a1b8-78f9a243223e\") " Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.604676 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "93e1015b-7d40-4ad7-a1b8-78f9a243223e" (UID: "93e1015b-7d40-4ad7-a1b8-78f9a243223e"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.604966 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93e1015b-7d40-4ad7-a1b8-78f9a243223e-kube-api-access-sm2gs" (OuterVolumeSpecName: "kube-api-access-sm2gs") pod "93e1015b-7d40-4ad7-a1b8-78f9a243223e" (UID: "93e1015b-7d40-4ad7-a1b8-78f9a243223e"). InnerVolumeSpecName "kube-api-access-sm2gs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.610837 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-ceph" (OuterVolumeSpecName: "ceph") pod "93e1015b-7d40-4ad7-a1b8-78f9a243223e" (UID: "93e1015b-7d40-4ad7-a1b8-78f9a243223e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.634040 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "93e1015b-7d40-4ad7-a1b8-78f9a243223e" (UID: "93e1015b-7d40-4ad7-a1b8-78f9a243223e"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.643740 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "93e1015b-7d40-4ad7-a1b8-78f9a243223e" (UID: "93e1015b-7d40-4ad7-a1b8-78f9a243223e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.644031 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-inventory" (OuterVolumeSpecName: "inventory") pod "93e1015b-7d40-4ad7-a1b8-78f9a243223e" (UID: "93e1015b-7d40-4ad7-a1b8-78f9a243223e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.701736 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm2gs\" (UniqueName: \"kubernetes.io/projected/93e1015b-7d40-4ad7-a1b8-78f9a243223e-kube-api-access-sm2gs\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.701771 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.701781 4997 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.701792 4997 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.701802 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.701835 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93e1015b-7d40-4ad7-a1b8-78f9a243223e-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.892874 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-5q88v" event={"ID":"93e1015b-7d40-4ad7-a1b8-78f9a243223e","Type":"ContainerDied","Data":"0597bea7eeab354f61cc6ac9061c6c68c533c821490450bf057d013d972348ff"} Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.892932 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0597bea7eeab354f61cc6ac9061c6c68c533c821490450bf057d013d972348ff" Dec 05 09:33:25 crc kubenswrapper[4997]: I1205 09:33:25.892937 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-5q88v" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.010669 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-lltrv"] Dec 05 09:33:26 crc kubenswrapper[4997]: E1205 09:33:26.011185 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="649fa519-39dd-4bb3-abef-12c1de34a7f0" containerName="extract-utilities" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.011198 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="649fa519-39dd-4bb3-abef-12c1de34a7f0" containerName="extract-utilities" Dec 05 09:33:26 crc kubenswrapper[4997]: E1205 09:33:26.011226 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="649fa519-39dd-4bb3-abef-12c1de34a7f0" containerName="registry-server" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.011232 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="649fa519-39dd-4bb3-abef-12c1de34a7f0" containerName="registry-server" Dec 05 09:33:26 crc kubenswrapper[4997]: E1205 09:33:26.011243 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e1015b-7d40-4ad7-a1b8-78f9a243223e" containerName="libvirt-openstack-openstack-cell1" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.011250 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e1015b-7d40-4ad7-a1b8-78f9a243223e" containerName="libvirt-openstack-openstack-cell1" Dec 05 09:33:26 crc kubenswrapper[4997]: E1205 09:33:26.011282 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="649fa519-39dd-4bb3-abef-12c1de34a7f0" containerName="extract-content" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.011288 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="649fa519-39dd-4bb3-abef-12c1de34a7f0" containerName="extract-content" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.011516 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="649fa519-39dd-4bb3-abef-12c1de34a7f0" containerName="registry-server" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.011526 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e1015b-7d40-4ad7-a1b8-78f9a243223e" containerName="libvirt-openstack-openstack-cell1" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.012326 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.016087 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.016153 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.017728 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.017901 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.024002 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-lltrv"] Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.027987 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.028442 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.028722 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.134882 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.134941 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-ceph\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.135036 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.135251 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt7qf\" (UniqueName: \"kubernetes.io/projected/c684f910-aebf-42e8-940b-5abd6b46c2a2-kube-api-access-wt7qf\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.135308 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.135351 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.135377 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.135396 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.135440 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-inventory\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.135538 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.135580 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.239478 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt7qf\" (UniqueName: \"kubernetes.io/projected/c684f910-aebf-42e8-940b-5abd6b46c2a2-kube-api-access-wt7qf\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.239527 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.239548 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.239573 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.239593 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.239653 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-inventory\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.239762 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.239788 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.239869 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.239885 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-ceph\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.239923 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.241874 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.245886 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.247873 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.251216 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.251493 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.251455 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.254401 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.256178 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-ceph\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.258235 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.259097 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-inventory\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.274692 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt7qf\" (UniqueName: \"kubernetes.io/projected/c684f910-aebf-42e8-940b-5abd6b46c2a2-kube-api-access-wt7qf\") pod \"nova-cell1-openstack-openstack-cell1-lltrv\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.341214 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.879514 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.885361 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-lltrv"] Dec 05 09:33:26 crc kubenswrapper[4997]: I1205 09:33:26.902813 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" event={"ID":"c684f910-aebf-42e8-940b-5abd6b46c2a2","Type":"ContainerStarted","Data":"3982c06a682dfd083a675a70258179013ecce67bdc8ebff8a87bc953d9e1dac9"} Dec 05 09:33:27 crc kubenswrapper[4997]: I1205 09:33:27.921373 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" event={"ID":"c684f910-aebf-42e8-940b-5abd6b46c2a2","Type":"ContainerStarted","Data":"92c8452b459264d0b9b96726b98e248a9973a0521060b60c55ecfad4280c99fd"} Dec 05 09:33:27 crc kubenswrapper[4997]: I1205 09:33:27.942113 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" podStartSLOduration=2.497436374 podStartE2EDuration="2.942089911s" podCreationTimestamp="2025-12-05 09:33:25 +0000 UTC" firstStartedPulling="2025-12-05 09:33:26.879173581 +0000 UTC m=+9507.408080842" lastFinishedPulling="2025-12-05 09:33:27.323827118 +0000 UTC m=+9507.852734379" observedRunningTime="2025-12-05 09:33:27.936313945 +0000 UTC m=+9508.465221206" watchObservedRunningTime="2025-12-05 09:33:27.942089911 +0000 UTC m=+9508.470997172" Dec 05 09:33:49 crc kubenswrapper[4997]: I1205 09:33:49.780942 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:33:49 crc kubenswrapper[4997]: I1205 09:33:49.781871 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:34:19 crc kubenswrapper[4997]: I1205 09:34:19.770021 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:34:19 crc kubenswrapper[4997]: I1205 09:34:19.770515 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:34:49 crc kubenswrapper[4997]: I1205 09:34:49.775329 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:34:49 crc kubenswrapper[4997]: I1205 09:34:49.778899 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:34:49 crc kubenswrapper[4997]: I1205 09:34:49.782664 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 09:34:49 crc kubenswrapper[4997]: I1205 09:34:49.785177 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:34:49 crc kubenswrapper[4997]: I1205 09:34:49.785258 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" gracePeriod=600 Dec 05 09:34:49 crc kubenswrapper[4997]: E1205 09:34:49.906983 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:34:50 crc kubenswrapper[4997]: I1205 09:34:50.775538 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" exitCode=0 Dec 05 09:34:50 crc kubenswrapper[4997]: I1205 09:34:50.775576 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06"} Dec 05 09:34:50 crc kubenswrapper[4997]: I1205 09:34:50.775886 4997 scope.go:117] "RemoveContainer" containerID="6ce1c08e479b9a4b91cc9213507822cb7d19ab1ed1dfebf53f033a419412ffa0" Dec 05 09:34:50 crc kubenswrapper[4997]: I1205 09:34:50.776766 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:34:50 crc kubenswrapper[4997]: E1205 09:34:50.777277 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:35:02 crc kubenswrapper[4997]: I1205 09:35:02.749804 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:35:02 crc kubenswrapper[4997]: E1205 09:35:02.750652 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:35:17 crc kubenswrapper[4997]: I1205 09:35:17.749080 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:35:17 crc kubenswrapper[4997]: E1205 09:35:17.749840 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:35:31 crc kubenswrapper[4997]: I1205 09:35:31.749245 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:35:31 crc kubenswrapper[4997]: E1205 09:35:31.750001 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:35:46 crc kubenswrapper[4997]: I1205 09:35:46.749630 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:35:46 crc kubenswrapper[4997]: E1205 09:35:46.750402 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:35:57 crc kubenswrapper[4997]: I1205 09:35:57.750071 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:35:57 crc kubenswrapper[4997]: E1205 09:35:57.750984 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:36:11 crc kubenswrapper[4997]: I1205 09:36:11.749325 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:36:11 crc kubenswrapper[4997]: E1205 09:36:11.750031 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:36:22 crc kubenswrapper[4997]: I1205 09:36:22.749837 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:36:22 crc kubenswrapper[4997]: E1205 09:36:22.750691 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:36:36 crc kubenswrapper[4997]: I1205 09:36:36.748662 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:36:36 crc kubenswrapper[4997]: E1205 09:36:36.750302 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:36:49 crc kubenswrapper[4997]: I1205 09:36:49.758575 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:36:49 crc kubenswrapper[4997]: E1205 09:36:49.759514 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:37:02 crc kubenswrapper[4997]: I1205 09:37:02.749411 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:37:02 crc kubenswrapper[4997]: E1205 09:37:02.750149 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:37:07 crc kubenswrapper[4997]: I1205 09:37:07.324447 4997 generic.go:334] "Generic (PLEG): container finished" podID="c684f910-aebf-42e8-940b-5abd6b46c2a2" containerID="92c8452b459264d0b9b96726b98e248a9973a0521060b60c55ecfad4280c99fd" exitCode=0 Dec 05 09:37:07 crc kubenswrapper[4997]: I1205 09:37:07.324977 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" event={"ID":"c684f910-aebf-42e8-940b-5abd6b46c2a2","Type":"ContainerDied","Data":"92c8452b459264d0b9b96726b98e248a9973a0521060b60c55ecfad4280c99fd"} Dec 05 09:37:08 crc kubenswrapper[4997]: I1205 09:37:08.884857 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:37:08 crc kubenswrapper[4997]: I1205 09:37:08.976528 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-combined-ca-bundle\") pod \"c684f910-aebf-42e8-940b-5abd6b46c2a2\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " Dec 05 09:37:08 crc kubenswrapper[4997]: I1205 09:37:08.976573 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-ceph\") pod \"c684f910-aebf-42e8-940b-5abd6b46c2a2\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " Dec 05 09:37:08 crc kubenswrapper[4997]: I1205 09:37:08.976590 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-migration-ssh-key-0\") pod \"c684f910-aebf-42e8-940b-5abd6b46c2a2\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " Dec 05 09:37:08 crc kubenswrapper[4997]: I1205 09:37:08.976706 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-compute-config-0\") pod \"c684f910-aebf-42e8-940b-5abd6b46c2a2\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " Dec 05 09:37:08 crc kubenswrapper[4997]: I1205 09:37:08.976737 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cells-global-config-0\") pod \"c684f910-aebf-42e8-940b-5abd6b46c2a2\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " Dec 05 09:37:08 crc kubenswrapper[4997]: I1205 09:37:08.976778 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wt7qf\" (UniqueName: \"kubernetes.io/projected/c684f910-aebf-42e8-940b-5abd6b46c2a2-kube-api-access-wt7qf\") pod \"c684f910-aebf-42e8-940b-5abd6b46c2a2\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " Dec 05 09:37:08 crc kubenswrapper[4997]: I1205 09:37:08.976807 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-inventory\") pod \"c684f910-aebf-42e8-940b-5abd6b46c2a2\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " Dec 05 09:37:08 crc kubenswrapper[4997]: I1205 09:37:08.976838 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-compute-config-1\") pod \"c684f910-aebf-42e8-940b-5abd6b46c2a2\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " Dec 05 09:37:08 crc kubenswrapper[4997]: I1205 09:37:08.976876 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cells-global-config-1\") pod \"c684f910-aebf-42e8-940b-5abd6b46c2a2\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " Dec 05 09:37:08 crc kubenswrapper[4997]: I1205 09:37:08.976901 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-ssh-key\") pod \"c684f910-aebf-42e8-940b-5abd6b46c2a2\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " Dec 05 09:37:08 crc kubenswrapper[4997]: I1205 09:37:08.976934 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-migration-ssh-key-1\") pod \"c684f910-aebf-42e8-940b-5abd6b46c2a2\" (UID: \"c684f910-aebf-42e8-940b-5abd6b46c2a2\") " Dec 05 09:37:08 crc kubenswrapper[4997]: I1205 09:37:08.982751 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-ceph" (OuterVolumeSpecName: "ceph") pod "c684f910-aebf-42e8-940b-5abd6b46c2a2" (UID: "c684f910-aebf-42e8-940b-5abd6b46c2a2"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:37:08 crc kubenswrapper[4997]: I1205 09:37:08.982900 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "c684f910-aebf-42e8-940b-5abd6b46c2a2" (UID: "c684f910-aebf-42e8-940b-5abd6b46c2a2"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:37:08 crc kubenswrapper[4997]: I1205 09:37:08.989157 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c684f910-aebf-42e8-940b-5abd6b46c2a2-kube-api-access-wt7qf" (OuterVolumeSpecName: "kube-api-access-wt7qf") pod "c684f910-aebf-42e8-940b-5abd6b46c2a2" (UID: "c684f910-aebf-42e8-940b-5abd6b46c2a2"). InnerVolumeSpecName "kube-api-access-wt7qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.005207 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "c684f910-aebf-42e8-940b-5abd6b46c2a2" (UID: "c684f910-aebf-42e8-940b-5abd6b46c2a2"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.005727 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "c684f910-aebf-42e8-940b-5abd6b46c2a2" (UID: "c684f910-aebf-42e8-940b-5abd6b46c2a2"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.014042 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "c684f910-aebf-42e8-940b-5abd6b46c2a2" (UID: "c684f910-aebf-42e8-940b-5abd6b46c2a2"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.018568 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "c684f910-aebf-42e8-940b-5abd6b46c2a2" (UID: "c684f910-aebf-42e8-940b-5abd6b46c2a2"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.022674 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "c684f910-aebf-42e8-940b-5abd6b46c2a2" (UID: "c684f910-aebf-42e8-940b-5abd6b46c2a2"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.028003 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "c684f910-aebf-42e8-940b-5abd6b46c2a2" (UID: "c684f910-aebf-42e8-940b-5abd6b46c2a2"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.030379 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c684f910-aebf-42e8-940b-5abd6b46c2a2" (UID: "c684f910-aebf-42e8-940b-5abd6b46c2a2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.052239 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-inventory" (OuterVolumeSpecName: "inventory") pod "c684f910-aebf-42e8-940b-5abd6b46c2a2" (UID: "c684f910-aebf-42e8-940b-5abd6b46c2a2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.082107 4997 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.082168 4997 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.082184 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wt7qf\" (UniqueName: \"kubernetes.io/projected/c684f910-aebf-42e8-940b-5abd6b46c2a2-kube-api-access-wt7qf\") on node \"crc\" DevicePath \"\"" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.082197 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.082212 4997 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.082225 4997 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.082237 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.082249 4997 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.082260 4997 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.082272 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.082286 4997 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c684f910-aebf-42e8-940b-5abd6b46c2a2-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.347004 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" event={"ID":"c684f910-aebf-42e8-940b-5abd6b46c2a2","Type":"ContainerDied","Data":"3982c06a682dfd083a675a70258179013ecce67bdc8ebff8a87bc953d9e1dac9"} Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.347042 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3982c06a682dfd083a675a70258179013ecce67bdc8ebff8a87bc953d9e1dac9" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.347062 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-lltrv" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.535146 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-q9b5m"] Dec 05 09:37:09 crc kubenswrapper[4997]: E1205 09:37:09.542210 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c684f910-aebf-42e8-940b-5abd6b46c2a2" containerName="nova-cell1-openstack-openstack-cell1" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.542248 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c684f910-aebf-42e8-940b-5abd6b46c2a2" containerName="nova-cell1-openstack-openstack-cell1" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.542576 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="c684f910-aebf-42e8-940b-5abd6b46c2a2" containerName="nova-cell1-openstack-openstack-cell1" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.543524 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.548209 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.548440 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.548853 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.549009 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.549121 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.558812 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-q9b5m"] Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.703684 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-inventory\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.703735 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ssh-key\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.703805 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.703860 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7nc6\" (UniqueName: \"kubernetes.io/projected/746846bb-be60-4333-a253-f1f3057ea2d7-kube-api-access-q7nc6\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.703881 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.703909 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.703985 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceph\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.704002 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.805364 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceph\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.805414 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.805453 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-inventory\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.805475 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ssh-key\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.805544 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.805597 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7nc6\" (UniqueName: \"kubernetes.io/projected/746846bb-be60-4333-a253-f1f3057ea2d7-kube-api-access-q7nc6\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.805637 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.805976 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.809828 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ssh-key\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.810236 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-inventory\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.810440 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.810464 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceph\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.811346 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.812215 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.814795 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.826222 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7nc6\" (UniqueName: \"kubernetes.io/projected/746846bb-be60-4333-a253-f1f3057ea2d7-kube-api-access-q7nc6\") pod \"telemetry-openstack-openstack-cell1-q9b5m\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:09 crc kubenswrapper[4997]: I1205 09:37:09.863711 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:37:10 crc kubenswrapper[4997]: I1205 09:37:10.398274 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-q9b5m"] Dec 05 09:37:11 crc kubenswrapper[4997]: I1205 09:37:11.369251 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" event={"ID":"746846bb-be60-4333-a253-f1f3057ea2d7","Type":"ContainerStarted","Data":"3c63b6c9a5471fc477c08f1de711354a627c0d94cd2ce3c4a62e1c3c145659a9"} Dec 05 09:37:11 crc kubenswrapper[4997]: I1205 09:37:11.369927 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" event={"ID":"746846bb-be60-4333-a253-f1f3057ea2d7","Type":"ContainerStarted","Data":"2681586441f965014631ff6922da8a9f04bf287d82ab44d2f01693e20b948b81"} Dec 05 09:37:11 crc kubenswrapper[4997]: I1205 09:37:11.402079 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" podStartSLOduration=1.9846424150000002 podStartE2EDuration="2.402054281s" podCreationTimestamp="2025-12-05 09:37:09 +0000 UTC" firstStartedPulling="2025-12-05 09:37:10.37983286 +0000 UTC m=+9730.908740121" lastFinishedPulling="2025-12-05 09:37:10.797244726 +0000 UTC m=+9731.326151987" observedRunningTime="2025-12-05 09:37:11.38542612 +0000 UTC m=+9731.914333401" watchObservedRunningTime="2025-12-05 09:37:11.402054281 +0000 UTC m=+9731.930961552" Dec 05 09:37:15 crc kubenswrapper[4997]: I1205 09:37:15.749311 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:37:15 crc kubenswrapper[4997]: E1205 09:37:15.750238 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:37:30 crc kubenswrapper[4997]: I1205 09:37:30.749169 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:37:30 crc kubenswrapper[4997]: E1205 09:37:30.749830 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:37:44 crc kubenswrapper[4997]: I1205 09:37:44.749542 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:37:44 crc kubenswrapper[4997]: E1205 09:37:44.750289 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:37:56 crc kubenswrapper[4997]: I1205 09:37:56.749896 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:37:56 crc kubenswrapper[4997]: E1205 09:37:56.750691 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:38:07 crc kubenswrapper[4997]: I1205 09:38:07.756330 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:38:07 crc kubenswrapper[4997]: E1205 09:38:07.757551 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:38:14 crc kubenswrapper[4997]: I1205 09:38:14.270292 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m76vp"] Dec 05 09:38:14 crc kubenswrapper[4997]: I1205 09:38:14.274684 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m76vp" Dec 05 09:38:14 crc kubenswrapper[4997]: I1205 09:38:14.299668 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m76vp"] Dec 05 09:38:14 crc kubenswrapper[4997]: I1205 09:38:14.410686 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95d516cb-9a56-4447-85fe-ad315bd3ad76-catalog-content\") pod \"certified-operators-m76vp\" (UID: \"95d516cb-9a56-4447-85fe-ad315bd3ad76\") " pod="openshift-marketplace/certified-operators-m76vp" Dec 05 09:38:14 crc kubenswrapper[4997]: I1205 09:38:14.411350 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv762\" (UniqueName: \"kubernetes.io/projected/95d516cb-9a56-4447-85fe-ad315bd3ad76-kube-api-access-hv762\") pod \"certified-operators-m76vp\" (UID: \"95d516cb-9a56-4447-85fe-ad315bd3ad76\") " pod="openshift-marketplace/certified-operators-m76vp" Dec 05 09:38:14 crc kubenswrapper[4997]: I1205 09:38:14.411515 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95d516cb-9a56-4447-85fe-ad315bd3ad76-utilities\") pod \"certified-operators-m76vp\" (UID: \"95d516cb-9a56-4447-85fe-ad315bd3ad76\") " pod="openshift-marketplace/certified-operators-m76vp" Dec 05 09:38:14 crc kubenswrapper[4997]: I1205 09:38:14.513365 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95d516cb-9a56-4447-85fe-ad315bd3ad76-catalog-content\") pod \"certified-operators-m76vp\" (UID: \"95d516cb-9a56-4447-85fe-ad315bd3ad76\") " pod="openshift-marketplace/certified-operators-m76vp" Dec 05 09:38:14 crc kubenswrapper[4997]: I1205 09:38:14.513444 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv762\" (UniqueName: \"kubernetes.io/projected/95d516cb-9a56-4447-85fe-ad315bd3ad76-kube-api-access-hv762\") pod \"certified-operators-m76vp\" (UID: \"95d516cb-9a56-4447-85fe-ad315bd3ad76\") " pod="openshift-marketplace/certified-operators-m76vp" Dec 05 09:38:14 crc kubenswrapper[4997]: I1205 09:38:14.513469 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95d516cb-9a56-4447-85fe-ad315bd3ad76-utilities\") pod \"certified-operators-m76vp\" (UID: \"95d516cb-9a56-4447-85fe-ad315bd3ad76\") " pod="openshift-marketplace/certified-operators-m76vp" Dec 05 09:38:14 crc kubenswrapper[4997]: I1205 09:38:14.514313 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95d516cb-9a56-4447-85fe-ad315bd3ad76-utilities\") pod \"certified-operators-m76vp\" (UID: \"95d516cb-9a56-4447-85fe-ad315bd3ad76\") " pod="openshift-marketplace/certified-operators-m76vp" Dec 05 09:38:14 crc kubenswrapper[4997]: I1205 09:38:14.514314 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95d516cb-9a56-4447-85fe-ad315bd3ad76-catalog-content\") pod \"certified-operators-m76vp\" (UID: \"95d516cb-9a56-4447-85fe-ad315bd3ad76\") " pod="openshift-marketplace/certified-operators-m76vp" Dec 05 09:38:14 crc kubenswrapper[4997]: I1205 09:38:14.558091 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv762\" (UniqueName: \"kubernetes.io/projected/95d516cb-9a56-4447-85fe-ad315bd3ad76-kube-api-access-hv762\") pod \"certified-operators-m76vp\" (UID: \"95d516cb-9a56-4447-85fe-ad315bd3ad76\") " pod="openshift-marketplace/certified-operators-m76vp" Dec 05 09:38:14 crc kubenswrapper[4997]: I1205 09:38:14.601686 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m76vp" Dec 05 09:38:15 crc kubenswrapper[4997]: W1205 09:38:15.112870 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95d516cb_9a56_4447_85fe_ad315bd3ad76.slice/crio-79ce51cb1927c09c38952aa3d60c9259128125e4f1f3403058ee23fe0088a1c7 WatchSource:0}: Error finding container 79ce51cb1927c09c38952aa3d60c9259128125e4f1f3403058ee23fe0088a1c7: Status 404 returned error can't find the container with id 79ce51cb1927c09c38952aa3d60c9259128125e4f1f3403058ee23fe0088a1c7 Dec 05 09:38:15 crc kubenswrapper[4997]: I1205 09:38:15.118172 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m76vp"] Dec 05 09:38:16 crc kubenswrapper[4997]: I1205 09:38:16.046286 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m76vp" event={"ID":"95d516cb-9a56-4447-85fe-ad315bd3ad76","Type":"ContainerStarted","Data":"79ce51cb1927c09c38952aa3d60c9259128125e4f1f3403058ee23fe0088a1c7"} Dec 05 09:38:19 crc kubenswrapper[4997]: I1205 09:38:19.765209 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:38:19 crc kubenswrapper[4997]: E1205 09:38:19.766481 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:38:20 crc kubenswrapper[4997]: I1205 09:38:20.092221 4997 generic.go:334] "Generic (PLEG): container finished" podID="95d516cb-9a56-4447-85fe-ad315bd3ad76" containerID="622c0f6ef22ff826ce3d42231c68bff2dc3390e91c7179b72b53c97b949761d1" exitCode=0 Dec 05 09:38:20 crc kubenswrapper[4997]: I1205 09:38:20.092273 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m76vp" event={"ID":"95d516cb-9a56-4447-85fe-ad315bd3ad76","Type":"ContainerDied","Data":"622c0f6ef22ff826ce3d42231c68bff2dc3390e91c7179b72b53c97b949761d1"} Dec 05 09:38:22 crc kubenswrapper[4997]: I1205 09:38:22.131346 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m76vp" event={"ID":"95d516cb-9a56-4447-85fe-ad315bd3ad76","Type":"ContainerStarted","Data":"0773b2fe68fe3389ddac13ee842e09d282af48e1725f5b2e347b5759877b9a0c"} Dec 05 09:38:23 crc kubenswrapper[4997]: I1205 09:38:23.143347 4997 generic.go:334] "Generic (PLEG): container finished" podID="95d516cb-9a56-4447-85fe-ad315bd3ad76" containerID="0773b2fe68fe3389ddac13ee842e09d282af48e1725f5b2e347b5759877b9a0c" exitCode=0 Dec 05 09:38:23 crc kubenswrapper[4997]: I1205 09:38:23.143399 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m76vp" event={"ID":"95d516cb-9a56-4447-85fe-ad315bd3ad76","Type":"ContainerDied","Data":"0773b2fe68fe3389ddac13ee842e09d282af48e1725f5b2e347b5759877b9a0c"} Dec 05 09:38:24 crc kubenswrapper[4997]: I1205 09:38:24.156661 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m76vp" event={"ID":"95d516cb-9a56-4447-85fe-ad315bd3ad76","Type":"ContainerStarted","Data":"4a103e09df91ce6012d6fbdc2e38b67dbe26eeb72349c52526df48c8f9e061ec"} Dec 05 09:38:24 crc kubenswrapper[4997]: I1205 09:38:24.186680 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m76vp" podStartSLOduration=6.473859326 podStartE2EDuration="10.186645225s" podCreationTimestamp="2025-12-05 09:38:14 +0000 UTC" firstStartedPulling="2025-12-05 09:38:20.094350157 +0000 UTC m=+9800.623257418" lastFinishedPulling="2025-12-05 09:38:23.807136066 +0000 UTC m=+9804.336043317" observedRunningTime="2025-12-05 09:38:24.176507609 +0000 UTC m=+9804.705414880" watchObservedRunningTime="2025-12-05 09:38:24.186645225 +0000 UTC m=+9804.715552496" Dec 05 09:38:24 crc kubenswrapper[4997]: I1205 09:38:24.602249 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m76vp" Dec 05 09:38:24 crc kubenswrapper[4997]: I1205 09:38:24.602296 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m76vp" Dec 05 09:38:25 crc kubenswrapper[4997]: I1205 09:38:25.646549 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-m76vp" podUID="95d516cb-9a56-4447-85fe-ad315bd3ad76" containerName="registry-server" probeResult="failure" output=< Dec 05 09:38:25 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 09:38:25 crc kubenswrapper[4997]: > Dec 05 09:38:29 crc kubenswrapper[4997]: I1205 09:38:29.779913 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-52hks"] Dec 05 09:38:29 crc kubenswrapper[4997]: I1205 09:38:29.782820 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-52hks" Dec 05 09:38:29 crc kubenswrapper[4997]: I1205 09:38:29.799331 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-52hks"] Dec 05 09:38:29 crc kubenswrapper[4997]: I1205 09:38:29.885132 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-utilities\") pod \"redhat-operators-52hks\" (UID: \"698b0a8e-bfbe-4654-974a-a41c0a1a0b08\") " pod="openshift-marketplace/redhat-operators-52hks" Dec 05 09:38:29 crc kubenswrapper[4997]: I1205 09:38:29.886112 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s87m\" (UniqueName: \"kubernetes.io/projected/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-kube-api-access-4s87m\") pod \"redhat-operators-52hks\" (UID: \"698b0a8e-bfbe-4654-974a-a41c0a1a0b08\") " pod="openshift-marketplace/redhat-operators-52hks" Dec 05 09:38:29 crc kubenswrapper[4997]: I1205 09:38:29.886151 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-catalog-content\") pod \"redhat-operators-52hks\" (UID: \"698b0a8e-bfbe-4654-974a-a41c0a1a0b08\") " pod="openshift-marketplace/redhat-operators-52hks" Dec 05 09:38:29 crc kubenswrapper[4997]: I1205 09:38:29.987732 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-utilities\") pod \"redhat-operators-52hks\" (UID: \"698b0a8e-bfbe-4654-974a-a41c0a1a0b08\") " pod="openshift-marketplace/redhat-operators-52hks" Dec 05 09:38:29 crc kubenswrapper[4997]: I1205 09:38:29.988166 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-utilities\") pod \"redhat-operators-52hks\" (UID: \"698b0a8e-bfbe-4654-974a-a41c0a1a0b08\") " pod="openshift-marketplace/redhat-operators-52hks" Dec 05 09:38:29 crc kubenswrapper[4997]: I1205 09:38:29.988394 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s87m\" (UniqueName: \"kubernetes.io/projected/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-kube-api-access-4s87m\") pod \"redhat-operators-52hks\" (UID: \"698b0a8e-bfbe-4654-974a-a41c0a1a0b08\") " pod="openshift-marketplace/redhat-operators-52hks" Dec 05 09:38:29 crc kubenswrapper[4997]: I1205 09:38:29.988520 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-catalog-content\") pod \"redhat-operators-52hks\" (UID: \"698b0a8e-bfbe-4654-974a-a41c0a1a0b08\") " pod="openshift-marketplace/redhat-operators-52hks" Dec 05 09:38:29 crc kubenswrapper[4997]: I1205 09:38:29.988814 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-catalog-content\") pod \"redhat-operators-52hks\" (UID: \"698b0a8e-bfbe-4654-974a-a41c0a1a0b08\") " pod="openshift-marketplace/redhat-operators-52hks" Dec 05 09:38:30 crc kubenswrapper[4997]: I1205 09:38:30.007800 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s87m\" (UniqueName: \"kubernetes.io/projected/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-kube-api-access-4s87m\") pod \"redhat-operators-52hks\" (UID: \"698b0a8e-bfbe-4654-974a-a41c0a1a0b08\") " pod="openshift-marketplace/redhat-operators-52hks" Dec 05 09:38:30 crc kubenswrapper[4997]: I1205 09:38:30.115210 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-52hks" Dec 05 09:38:30 crc kubenswrapper[4997]: I1205 09:38:30.634287 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-52hks"] Dec 05 09:38:30 crc kubenswrapper[4997]: I1205 09:38:30.748734 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:38:30 crc kubenswrapper[4997]: E1205 09:38:30.749024 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:38:31 crc kubenswrapper[4997]: I1205 09:38:31.229126 4997 generic.go:334] "Generic (PLEG): container finished" podID="698b0a8e-bfbe-4654-974a-a41c0a1a0b08" containerID="661ddb6ea1fab0d635dc730da5741a5e24ee77eec25c93b2f892891e68edb0a7" exitCode=0 Dec 05 09:38:31 crc kubenswrapper[4997]: I1205 09:38:31.229218 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52hks" event={"ID":"698b0a8e-bfbe-4654-974a-a41c0a1a0b08","Type":"ContainerDied","Data":"661ddb6ea1fab0d635dc730da5741a5e24ee77eec25c93b2f892891e68edb0a7"} Dec 05 09:38:31 crc kubenswrapper[4997]: I1205 09:38:31.229470 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52hks" event={"ID":"698b0a8e-bfbe-4654-974a-a41c0a1a0b08","Type":"ContainerStarted","Data":"c6418fa4f6ca09962bdbf6306793c06e3871c78a40a44b23bc78baba7073c8c4"} Dec 05 09:38:31 crc kubenswrapper[4997]: I1205 09:38:31.231205 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:38:32 crc kubenswrapper[4997]: I1205 09:38:32.243436 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52hks" event={"ID":"698b0a8e-bfbe-4654-974a-a41c0a1a0b08","Type":"ContainerStarted","Data":"8230c6b59d7969f60cc0d7d26d27b54bc088c018123fbb2d1829af93c44c449f"} Dec 05 09:38:34 crc kubenswrapper[4997]: I1205 09:38:34.649882 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m76vp" Dec 05 09:38:34 crc kubenswrapper[4997]: I1205 09:38:34.713259 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m76vp" Dec 05 09:38:35 crc kubenswrapper[4997]: I1205 09:38:35.279290 4997 generic.go:334] "Generic (PLEG): container finished" podID="698b0a8e-bfbe-4654-974a-a41c0a1a0b08" containerID="8230c6b59d7969f60cc0d7d26d27b54bc088c018123fbb2d1829af93c44c449f" exitCode=0 Dec 05 09:38:35 crc kubenswrapper[4997]: I1205 09:38:35.279412 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52hks" event={"ID":"698b0a8e-bfbe-4654-974a-a41c0a1a0b08","Type":"ContainerDied","Data":"8230c6b59d7969f60cc0d7d26d27b54bc088c018123fbb2d1829af93c44c449f"} Dec 05 09:38:35 crc kubenswrapper[4997]: I1205 09:38:35.765409 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m76vp"] Dec 05 09:38:36 crc kubenswrapper[4997]: I1205 09:38:36.294794 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52hks" event={"ID":"698b0a8e-bfbe-4654-974a-a41c0a1a0b08","Type":"ContainerStarted","Data":"f48b9222434c67682bde6bacfe7f14e410717b294c3a1405954b62ef40775d2e"} Dec 05 09:38:36 crc kubenswrapper[4997]: I1205 09:38:36.295114 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m76vp" podUID="95d516cb-9a56-4447-85fe-ad315bd3ad76" containerName="registry-server" containerID="cri-o://4a103e09df91ce6012d6fbdc2e38b67dbe26eeb72349c52526df48c8f9e061ec" gracePeriod=2 Dec 05 09:38:36 crc kubenswrapper[4997]: I1205 09:38:36.322327 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-52hks" podStartSLOduration=2.83492693 podStartE2EDuration="7.322291967s" podCreationTimestamp="2025-12-05 09:38:29 +0000 UTC" firstStartedPulling="2025-12-05 09:38:31.230991398 +0000 UTC m=+9811.759898659" lastFinishedPulling="2025-12-05 09:38:35.718356435 +0000 UTC m=+9816.247263696" observedRunningTime="2025-12-05 09:38:36.321043523 +0000 UTC m=+9816.849950784" watchObservedRunningTime="2025-12-05 09:38:36.322291967 +0000 UTC m=+9816.851199228" Dec 05 09:38:36 crc kubenswrapper[4997]: I1205 09:38:36.797140 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m76vp" Dec 05 09:38:36 crc kubenswrapper[4997]: I1205 09:38:36.956005 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95d516cb-9a56-4447-85fe-ad315bd3ad76-catalog-content\") pod \"95d516cb-9a56-4447-85fe-ad315bd3ad76\" (UID: \"95d516cb-9a56-4447-85fe-ad315bd3ad76\") " Dec 05 09:38:36 crc kubenswrapper[4997]: I1205 09:38:36.956389 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv762\" (UniqueName: \"kubernetes.io/projected/95d516cb-9a56-4447-85fe-ad315bd3ad76-kube-api-access-hv762\") pod \"95d516cb-9a56-4447-85fe-ad315bd3ad76\" (UID: \"95d516cb-9a56-4447-85fe-ad315bd3ad76\") " Dec 05 09:38:36 crc kubenswrapper[4997]: I1205 09:38:36.957240 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95d516cb-9a56-4447-85fe-ad315bd3ad76-utilities\") pod \"95d516cb-9a56-4447-85fe-ad315bd3ad76\" (UID: \"95d516cb-9a56-4447-85fe-ad315bd3ad76\") " Dec 05 09:38:36 crc kubenswrapper[4997]: I1205 09:38:36.958428 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95d516cb-9a56-4447-85fe-ad315bd3ad76-utilities" (OuterVolumeSpecName: "utilities") pod "95d516cb-9a56-4447-85fe-ad315bd3ad76" (UID: "95d516cb-9a56-4447-85fe-ad315bd3ad76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:38:36 crc kubenswrapper[4997]: I1205 09:38:36.962044 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95d516cb-9a56-4447-85fe-ad315bd3ad76-kube-api-access-hv762" (OuterVolumeSpecName: "kube-api-access-hv762") pod "95d516cb-9a56-4447-85fe-ad315bd3ad76" (UID: "95d516cb-9a56-4447-85fe-ad315bd3ad76"). InnerVolumeSpecName "kube-api-access-hv762". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:36.999804 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95d516cb-9a56-4447-85fe-ad315bd3ad76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95d516cb-9a56-4447-85fe-ad315bd3ad76" (UID: "95d516cb-9a56-4447-85fe-ad315bd3ad76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.059768 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv762\" (UniqueName: \"kubernetes.io/projected/95d516cb-9a56-4447-85fe-ad315bd3ad76-kube-api-access-hv762\") on node \"crc\" DevicePath \"\"" Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.059820 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95d516cb-9a56-4447-85fe-ad315bd3ad76-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.059832 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95d516cb-9a56-4447-85fe-ad315bd3ad76-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.305907 4997 generic.go:334] "Generic (PLEG): container finished" podID="95d516cb-9a56-4447-85fe-ad315bd3ad76" containerID="4a103e09df91ce6012d6fbdc2e38b67dbe26eeb72349c52526df48c8f9e061ec" exitCode=0 Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.305970 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m76vp" event={"ID":"95d516cb-9a56-4447-85fe-ad315bd3ad76","Type":"ContainerDied","Data":"4a103e09df91ce6012d6fbdc2e38b67dbe26eeb72349c52526df48c8f9e061ec"} Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.306012 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m76vp" event={"ID":"95d516cb-9a56-4447-85fe-ad315bd3ad76","Type":"ContainerDied","Data":"79ce51cb1927c09c38952aa3d60c9259128125e4f1f3403058ee23fe0088a1c7"} Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.306034 4997 scope.go:117] "RemoveContainer" containerID="4a103e09df91ce6012d6fbdc2e38b67dbe26eeb72349c52526df48c8f9e061ec" Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.306277 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m76vp" Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.328434 4997 scope.go:117] "RemoveContainer" containerID="0773b2fe68fe3389ddac13ee842e09d282af48e1725f5b2e347b5759877b9a0c" Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.350921 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m76vp"] Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.360448 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m76vp"] Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.368150 4997 scope.go:117] "RemoveContainer" containerID="622c0f6ef22ff826ce3d42231c68bff2dc3390e91c7179b72b53c97b949761d1" Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.404128 4997 scope.go:117] "RemoveContainer" containerID="4a103e09df91ce6012d6fbdc2e38b67dbe26eeb72349c52526df48c8f9e061ec" Dec 05 09:38:37 crc kubenswrapper[4997]: E1205 09:38:37.404834 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a103e09df91ce6012d6fbdc2e38b67dbe26eeb72349c52526df48c8f9e061ec\": container with ID starting with 4a103e09df91ce6012d6fbdc2e38b67dbe26eeb72349c52526df48c8f9e061ec not found: ID does not exist" containerID="4a103e09df91ce6012d6fbdc2e38b67dbe26eeb72349c52526df48c8f9e061ec" Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.404893 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a103e09df91ce6012d6fbdc2e38b67dbe26eeb72349c52526df48c8f9e061ec"} err="failed to get container status \"4a103e09df91ce6012d6fbdc2e38b67dbe26eeb72349c52526df48c8f9e061ec\": rpc error: code = NotFound desc = could not find container \"4a103e09df91ce6012d6fbdc2e38b67dbe26eeb72349c52526df48c8f9e061ec\": container with ID starting with 4a103e09df91ce6012d6fbdc2e38b67dbe26eeb72349c52526df48c8f9e061ec not found: ID does not exist" Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.404928 4997 scope.go:117] "RemoveContainer" containerID="0773b2fe68fe3389ddac13ee842e09d282af48e1725f5b2e347b5759877b9a0c" Dec 05 09:38:37 crc kubenswrapper[4997]: E1205 09:38:37.406367 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0773b2fe68fe3389ddac13ee842e09d282af48e1725f5b2e347b5759877b9a0c\": container with ID starting with 0773b2fe68fe3389ddac13ee842e09d282af48e1725f5b2e347b5759877b9a0c not found: ID does not exist" containerID="0773b2fe68fe3389ddac13ee842e09d282af48e1725f5b2e347b5759877b9a0c" Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.406438 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0773b2fe68fe3389ddac13ee842e09d282af48e1725f5b2e347b5759877b9a0c"} err="failed to get container status \"0773b2fe68fe3389ddac13ee842e09d282af48e1725f5b2e347b5759877b9a0c\": rpc error: code = NotFound desc = could not find container \"0773b2fe68fe3389ddac13ee842e09d282af48e1725f5b2e347b5759877b9a0c\": container with ID starting with 0773b2fe68fe3389ddac13ee842e09d282af48e1725f5b2e347b5759877b9a0c not found: ID does not exist" Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.406475 4997 scope.go:117] "RemoveContainer" containerID="622c0f6ef22ff826ce3d42231c68bff2dc3390e91c7179b72b53c97b949761d1" Dec 05 09:38:37 crc kubenswrapper[4997]: E1205 09:38:37.406950 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"622c0f6ef22ff826ce3d42231c68bff2dc3390e91c7179b72b53c97b949761d1\": container with ID starting with 622c0f6ef22ff826ce3d42231c68bff2dc3390e91c7179b72b53c97b949761d1 not found: ID does not exist" containerID="622c0f6ef22ff826ce3d42231c68bff2dc3390e91c7179b72b53c97b949761d1" Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.406975 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"622c0f6ef22ff826ce3d42231c68bff2dc3390e91c7179b72b53c97b949761d1"} err="failed to get container status \"622c0f6ef22ff826ce3d42231c68bff2dc3390e91c7179b72b53c97b949761d1\": rpc error: code = NotFound desc = could not find container \"622c0f6ef22ff826ce3d42231c68bff2dc3390e91c7179b72b53c97b949761d1\": container with ID starting with 622c0f6ef22ff826ce3d42231c68bff2dc3390e91c7179b72b53c97b949761d1 not found: ID does not exist" Dec 05 09:38:37 crc kubenswrapper[4997]: I1205 09:38:37.768552 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95d516cb-9a56-4447-85fe-ad315bd3ad76" path="/var/lib/kubelet/pods/95d516cb-9a56-4447-85fe-ad315bd3ad76/volumes" Dec 05 09:38:40 crc kubenswrapper[4997]: I1205 09:38:40.116225 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-52hks" Dec 05 09:38:40 crc kubenswrapper[4997]: I1205 09:38:40.116630 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-52hks" Dec 05 09:38:41 crc kubenswrapper[4997]: I1205 09:38:41.161327 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-52hks" podUID="698b0a8e-bfbe-4654-974a-a41c0a1a0b08" containerName="registry-server" probeResult="failure" output=< Dec 05 09:38:41 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 09:38:41 crc kubenswrapper[4997]: > Dec 05 09:38:43 crc kubenswrapper[4997]: I1205 09:38:43.749546 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:38:43 crc kubenswrapper[4997]: E1205 09:38:43.750382 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:38:50 crc kubenswrapper[4997]: I1205 09:38:50.163637 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-52hks" Dec 05 09:38:50 crc kubenswrapper[4997]: I1205 09:38:50.214427 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-52hks" Dec 05 09:38:53 crc kubenswrapper[4997]: I1205 09:38:53.608168 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-52hks"] Dec 05 09:38:53 crc kubenswrapper[4997]: I1205 09:38:53.608955 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-52hks" podUID="698b0a8e-bfbe-4654-974a-a41c0a1a0b08" containerName="registry-server" containerID="cri-o://f48b9222434c67682bde6bacfe7f14e410717b294c3a1405954b62ef40775d2e" gracePeriod=2 Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.144236 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-52hks" Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.250191 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4s87m\" (UniqueName: \"kubernetes.io/projected/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-kube-api-access-4s87m\") pod \"698b0a8e-bfbe-4654-974a-a41c0a1a0b08\" (UID: \"698b0a8e-bfbe-4654-974a-a41c0a1a0b08\") " Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.250568 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-catalog-content\") pod \"698b0a8e-bfbe-4654-974a-a41c0a1a0b08\" (UID: \"698b0a8e-bfbe-4654-974a-a41c0a1a0b08\") " Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.250731 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-utilities\") pod \"698b0a8e-bfbe-4654-974a-a41c0a1a0b08\" (UID: \"698b0a8e-bfbe-4654-974a-a41c0a1a0b08\") " Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.251669 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-utilities" (OuterVolumeSpecName: "utilities") pod "698b0a8e-bfbe-4654-974a-a41c0a1a0b08" (UID: "698b0a8e-bfbe-4654-974a-a41c0a1a0b08"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.255753 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-kube-api-access-4s87m" (OuterVolumeSpecName: "kube-api-access-4s87m") pod "698b0a8e-bfbe-4654-974a-a41c0a1a0b08" (UID: "698b0a8e-bfbe-4654-974a-a41c0a1a0b08"). InnerVolumeSpecName "kube-api-access-4s87m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.353169 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4s87m\" (UniqueName: \"kubernetes.io/projected/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-kube-api-access-4s87m\") on node \"crc\" DevicePath \"\"" Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.353199 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.358242 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "698b0a8e-bfbe-4654-974a-a41c0a1a0b08" (UID: "698b0a8e-bfbe-4654-974a-a41c0a1a0b08"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.454955 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/698b0a8e-bfbe-4654-974a-a41c0a1a0b08-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.496094 4997 generic.go:334] "Generic (PLEG): container finished" podID="698b0a8e-bfbe-4654-974a-a41c0a1a0b08" containerID="f48b9222434c67682bde6bacfe7f14e410717b294c3a1405954b62ef40775d2e" exitCode=0 Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.496138 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52hks" event={"ID":"698b0a8e-bfbe-4654-974a-a41c0a1a0b08","Type":"ContainerDied","Data":"f48b9222434c67682bde6bacfe7f14e410717b294c3a1405954b62ef40775d2e"} Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.496168 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52hks" event={"ID":"698b0a8e-bfbe-4654-974a-a41c0a1a0b08","Type":"ContainerDied","Data":"c6418fa4f6ca09962bdbf6306793c06e3871c78a40a44b23bc78baba7073c8c4"} Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.496174 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-52hks" Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.496186 4997 scope.go:117] "RemoveContainer" containerID="f48b9222434c67682bde6bacfe7f14e410717b294c3a1405954b62ef40775d2e" Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.522240 4997 scope.go:117] "RemoveContainer" containerID="8230c6b59d7969f60cc0d7d26d27b54bc088c018123fbb2d1829af93c44c449f" Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.535222 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-52hks"] Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.552137 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-52hks"] Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.565388 4997 scope.go:117] "RemoveContainer" containerID="661ddb6ea1fab0d635dc730da5741a5e24ee77eec25c93b2f892891e68edb0a7" Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.608875 4997 scope.go:117] "RemoveContainer" containerID="f48b9222434c67682bde6bacfe7f14e410717b294c3a1405954b62ef40775d2e" Dec 05 09:38:54 crc kubenswrapper[4997]: E1205 09:38:54.609342 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f48b9222434c67682bde6bacfe7f14e410717b294c3a1405954b62ef40775d2e\": container with ID starting with f48b9222434c67682bde6bacfe7f14e410717b294c3a1405954b62ef40775d2e not found: ID does not exist" containerID="f48b9222434c67682bde6bacfe7f14e410717b294c3a1405954b62ef40775d2e" Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.609429 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f48b9222434c67682bde6bacfe7f14e410717b294c3a1405954b62ef40775d2e"} err="failed to get container status \"f48b9222434c67682bde6bacfe7f14e410717b294c3a1405954b62ef40775d2e\": rpc error: code = NotFound desc = could not find container \"f48b9222434c67682bde6bacfe7f14e410717b294c3a1405954b62ef40775d2e\": container with ID starting with f48b9222434c67682bde6bacfe7f14e410717b294c3a1405954b62ef40775d2e not found: ID does not exist" Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.609462 4997 scope.go:117] "RemoveContainer" containerID="8230c6b59d7969f60cc0d7d26d27b54bc088c018123fbb2d1829af93c44c449f" Dec 05 09:38:54 crc kubenswrapper[4997]: E1205 09:38:54.609919 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8230c6b59d7969f60cc0d7d26d27b54bc088c018123fbb2d1829af93c44c449f\": container with ID starting with 8230c6b59d7969f60cc0d7d26d27b54bc088c018123fbb2d1829af93c44c449f not found: ID does not exist" containerID="8230c6b59d7969f60cc0d7d26d27b54bc088c018123fbb2d1829af93c44c449f" Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.609944 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8230c6b59d7969f60cc0d7d26d27b54bc088c018123fbb2d1829af93c44c449f"} err="failed to get container status \"8230c6b59d7969f60cc0d7d26d27b54bc088c018123fbb2d1829af93c44c449f\": rpc error: code = NotFound desc = could not find container \"8230c6b59d7969f60cc0d7d26d27b54bc088c018123fbb2d1829af93c44c449f\": container with ID starting with 8230c6b59d7969f60cc0d7d26d27b54bc088c018123fbb2d1829af93c44c449f not found: ID does not exist" Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.609960 4997 scope.go:117] "RemoveContainer" containerID="661ddb6ea1fab0d635dc730da5741a5e24ee77eec25c93b2f892891e68edb0a7" Dec 05 09:38:54 crc kubenswrapper[4997]: E1205 09:38:54.610234 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"661ddb6ea1fab0d635dc730da5741a5e24ee77eec25c93b2f892891e68edb0a7\": container with ID starting with 661ddb6ea1fab0d635dc730da5741a5e24ee77eec25c93b2f892891e68edb0a7 not found: ID does not exist" containerID="661ddb6ea1fab0d635dc730da5741a5e24ee77eec25c93b2f892891e68edb0a7" Dec 05 09:38:54 crc kubenswrapper[4997]: I1205 09:38:54.610270 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"661ddb6ea1fab0d635dc730da5741a5e24ee77eec25c93b2f892891e68edb0a7"} err="failed to get container status \"661ddb6ea1fab0d635dc730da5741a5e24ee77eec25c93b2f892891e68edb0a7\": rpc error: code = NotFound desc = could not find container \"661ddb6ea1fab0d635dc730da5741a5e24ee77eec25c93b2f892891e68edb0a7\": container with ID starting with 661ddb6ea1fab0d635dc730da5741a5e24ee77eec25c93b2f892891e68edb0a7 not found: ID does not exist" Dec 05 09:38:55 crc kubenswrapper[4997]: I1205 09:38:55.765030 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="698b0a8e-bfbe-4654-974a-a41c0a1a0b08" path="/var/lib/kubelet/pods/698b0a8e-bfbe-4654-974a-a41c0a1a0b08/volumes" Dec 05 09:38:57 crc kubenswrapper[4997]: I1205 09:38:57.749086 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:38:57 crc kubenswrapper[4997]: E1205 09:38:57.749919 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:39:11 crc kubenswrapper[4997]: I1205 09:39:11.750103 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:39:11 crc kubenswrapper[4997]: E1205 09:39:11.751079 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:39:24 crc kubenswrapper[4997]: I1205 09:39:24.748947 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:39:24 crc kubenswrapper[4997]: E1205 09:39:24.749762 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:39:39 crc kubenswrapper[4997]: I1205 09:39:39.762979 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:39:39 crc kubenswrapper[4997]: E1205 09:39:39.764186 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:39:54 crc kubenswrapper[4997]: I1205 09:39:54.748821 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:39:55 crc kubenswrapper[4997]: I1205 09:39:55.119294 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"553255d50247709601e534c39b4889a350b0020aa3393fb72def52e3ad4252bf"} Dec 05 09:41:03 crc kubenswrapper[4997]: I1205 09:41:03.821870 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-477xj"] Dec 05 09:41:03 crc kubenswrapper[4997]: E1205 09:41:03.822945 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="698b0a8e-bfbe-4654-974a-a41c0a1a0b08" containerName="extract-content" Dec 05 09:41:03 crc kubenswrapper[4997]: I1205 09:41:03.822969 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="698b0a8e-bfbe-4654-974a-a41c0a1a0b08" containerName="extract-content" Dec 05 09:41:03 crc kubenswrapper[4997]: E1205 09:41:03.823001 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95d516cb-9a56-4447-85fe-ad315bd3ad76" containerName="registry-server" Dec 05 09:41:03 crc kubenswrapper[4997]: I1205 09:41:03.823007 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="95d516cb-9a56-4447-85fe-ad315bd3ad76" containerName="registry-server" Dec 05 09:41:03 crc kubenswrapper[4997]: E1205 09:41:03.823022 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="698b0a8e-bfbe-4654-974a-a41c0a1a0b08" containerName="registry-server" Dec 05 09:41:03 crc kubenswrapper[4997]: I1205 09:41:03.823028 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="698b0a8e-bfbe-4654-974a-a41c0a1a0b08" containerName="registry-server" Dec 05 09:41:03 crc kubenswrapper[4997]: E1205 09:41:03.823036 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="698b0a8e-bfbe-4654-974a-a41c0a1a0b08" containerName="extract-utilities" Dec 05 09:41:03 crc kubenswrapper[4997]: I1205 09:41:03.823050 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="698b0a8e-bfbe-4654-974a-a41c0a1a0b08" containerName="extract-utilities" Dec 05 09:41:03 crc kubenswrapper[4997]: E1205 09:41:03.823067 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95d516cb-9a56-4447-85fe-ad315bd3ad76" containerName="extract-utilities" Dec 05 09:41:03 crc kubenswrapper[4997]: I1205 09:41:03.823074 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="95d516cb-9a56-4447-85fe-ad315bd3ad76" containerName="extract-utilities" Dec 05 09:41:03 crc kubenswrapper[4997]: E1205 09:41:03.823091 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95d516cb-9a56-4447-85fe-ad315bd3ad76" containerName="extract-content" Dec 05 09:41:03 crc kubenswrapper[4997]: I1205 09:41:03.823099 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="95d516cb-9a56-4447-85fe-ad315bd3ad76" containerName="extract-content" Dec 05 09:41:03 crc kubenswrapper[4997]: I1205 09:41:03.823342 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="95d516cb-9a56-4447-85fe-ad315bd3ad76" containerName="registry-server" Dec 05 09:41:03 crc kubenswrapper[4997]: I1205 09:41:03.823379 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="698b0a8e-bfbe-4654-974a-a41c0a1a0b08" containerName="registry-server" Dec 05 09:41:03 crc kubenswrapper[4997]: I1205 09:41:03.825238 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-477xj" Dec 05 09:41:03 crc kubenswrapper[4997]: I1205 09:41:03.854345 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-477xj"] Dec 05 09:41:03 crc kubenswrapper[4997]: I1205 09:41:03.948776 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6k9v\" (UniqueName: \"kubernetes.io/projected/6d2a2784-93e2-411c-b7da-707a3cd53d50-kube-api-access-d6k9v\") pod \"redhat-marketplace-477xj\" (UID: \"6d2a2784-93e2-411c-b7da-707a3cd53d50\") " pod="openshift-marketplace/redhat-marketplace-477xj" Dec 05 09:41:03 crc kubenswrapper[4997]: I1205 09:41:03.949011 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d2a2784-93e2-411c-b7da-707a3cd53d50-utilities\") pod \"redhat-marketplace-477xj\" (UID: \"6d2a2784-93e2-411c-b7da-707a3cd53d50\") " pod="openshift-marketplace/redhat-marketplace-477xj" Dec 05 09:41:03 crc kubenswrapper[4997]: I1205 09:41:03.949334 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d2a2784-93e2-411c-b7da-707a3cd53d50-catalog-content\") pod \"redhat-marketplace-477xj\" (UID: \"6d2a2784-93e2-411c-b7da-707a3cd53d50\") " pod="openshift-marketplace/redhat-marketplace-477xj" Dec 05 09:41:04 crc kubenswrapper[4997]: I1205 09:41:04.051461 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6k9v\" (UniqueName: \"kubernetes.io/projected/6d2a2784-93e2-411c-b7da-707a3cd53d50-kube-api-access-d6k9v\") pod \"redhat-marketplace-477xj\" (UID: \"6d2a2784-93e2-411c-b7da-707a3cd53d50\") " pod="openshift-marketplace/redhat-marketplace-477xj" Dec 05 09:41:04 crc kubenswrapper[4997]: I1205 09:41:04.051547 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d2a2784-93e2-411c-b7da-707a3cd53d50-utilities\") pod \"redhat-marketplace-477xj\" (UID: \"6d2a2784-93e2-411c-b7da-707a3cd53d50\") " pod="openshift-marketplace/redhat-marketplace-477xj" Dec 05 09:41:04 crc kubenswrapper[4997]: I1205 09:41:04.051626 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d2a2784-93e2-411c-b7da-707a3cd53d50-catalog-content\") pod \"redhat-marketplace-477xj\" (UID: \"6d2a2784-93e2-411c-b7da-707a3cd53d50\") " pod="openshift-marketplace/redhat-marketplace-477xj" Dec 05 09:41:04 crc kubenswrapper[4997]: I1205 09:41:04.052132 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d2a2784-93e2-411c-b7da-707a3cd53d50-utilities\") pod \"redhat-marketplace-477xj\" (UID: \"6d2a2784-93e2-411c-b7da-707a3cd53d50\") " pod="openshift-marketplace/redhat-marketplace-477xj" Dec 05 09:41:04 crc kubenswrapper[4997]: I1205 09:41:04.052182 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d2a2784-93e2-411c-b7da-707a3cd53d50-catalog-content\") pod \"redhat-marketplace-477xj\" (UID: \"6d2a2784-93e2-411c-b7da-707a3cd53d50\") " pod="openshift-marketplace/redhat-marketplace-477xj" Dec 05 09:41:04 crc kubenswrapper[4997]: I1205 09:41:04.071270 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6k9v\" (UniqueName: \"kubernetes.io/projected/6d2a2784-93e2-411c-b7da-707a3cd53d50-kube-api-access-d6k9v\") pod \"redhat-marketplace-477xj\" (UID: \"6d2a2784-93e2-411c-b7da-707a3cd53d50\") " pod="openshift-marketplace/redhat-marketplace-477xj" Dec 05 09:41:04 crc kubenswrapper[4997]: I1205 09:41:04.147104 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-477xj" Dec 05 09:41:04 crc kubenswrapper[4997]: I1205 09:41:04.669722 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-477xj"] Dec 05 09:41:04 crc kubenswrapper[4997]: I1205 09:41:04.851801 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-477xj" event={"ID":"6d2a2784-93e2-411c-b7da-707a3cd53d50","Type":"ContainerStarted","Data":"2d3cb4d9d533780e6122dc8b5049a00dd85471c4508c088e87fb839b9a9a8340"} Dec 05 09:41:05 crc kubenswrapper[4997]: I1205 09:41:05.865939 4997 generic.go:334] "Generic (PLEG): container finished" podID="6d2a2784-93e2-411c-b7da-707a3cd53d50" containerID="bdc199c0c28580ffc5da3ec5c120dd9adc9cde5d8345fa38f34941cf2a5a1ee7" exitCode=0 Dec 05 09:41:05 crc kubenswrapper[4997]: I1205 09:41:05.866284 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-477xj" event={"ID":"6d2a2784-93e2-411c-b7da-707a3cd53d50","Type":"ContainerDied","Data":"bdc199c0c28580ffc5da3ec5c120dd9adc9cde5d8345fa38f34941cf2a5a1ee7"} Dec 05 09:41:06 crc kubenswrapper[4997]: I1205 09:41:06.879818 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-477xj" event={"ID":"6d2a2784-93e2-411c-b7da-707a3cd53d50","Type":"ContainerStarted","Data":"fe3f520b8d1b67333b254fb63b4da289b0a0a4edc7a965d9ae1baaec32472bba"} Dec 05 09:41:07 crc kubenswrapper[4997]: I1205 09:41:07.891869 4997 generic.go:334] "Generic (PLEG): container finished" podID="6d2a2784-93e2-411c-b7da-707a3cd53d50" containerID="fe3f520b8d1b67333b254fb63b4da289b0a0a4edc7a965d9ae1baaec32472bba" exitCode=0 Dec 05 09:41:07 crc kubenswrapper[4997]: I1205 09:41:07.892189 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-477xj" event={"ID":"6d2a2784-93e2-411c-b7da-707a3cd53d50","Type":"ContainerDied","Data":"fe3f520b8d1b67333b254fb63b4da289b0a0a4edc7a965d9ae1baaec32472bba"} Dec 05 09:41:08 crc kubenswrapper[4997]: I1205 09:41:08.903788 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-477xj" event={"ID":"6d2a2784-93e2-411c-b7da-707a3cd53d50","Type":"ContainerStarted","Data":"93d135bad0b701906eeaf645b58ba4cbe780321d11c41fa8713f046c41ca7e50"} Dec 05 09:41:08 crc kubenswrapper[4997]: I1205 09:41:08.948210 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-477xj" podStartSLOduration=3.517293575 podStartE2EDuration="5.948188544s" podCreationTimestamp="2025-12-05 09:41:03 +0000 UTC" firstStartedPulling="2025-12-05 09:41:05.868506178 +0000 UTC m=+9966.397413439" lastFinishedPulling="2025-12-05 09:41:08.299401147 +0000 UTC m=+9968.828308408" observedRunningTime="2025-12-05 09:41:08.923215627 +0000 UTC m=+9969.452122908" watchObservedRunningTime="2025-12-05 09:41:08.948188544 +0000 UTC m=+9969.477095815" Dec 05 09:41:14 crc kubenswrapper[4997]: I1205 09:41:14.147546 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-477xj" Dec 05 09:41:14 crc kubenswrapper[4997]: I1205 09:41:14.148109 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-477xj" Dec 05 09:41:14 crc kubenswrapper[4997]: I1205 09:41:14.198950 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-477xj" Dec 05 09:41:15 crc kubenswrapper[4997]: I1205 09:41:15.015983 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-477xj" Dec 05 09:41:15 crc kubenswrapper[4997]: I1205 09:41:15.079816 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-477xj"] Dec 05 09:41:16 crc kubenswrapper[4997]: I1205 09:41:16.982111 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-477xj" podUID="6d2a2784-93e2-411c-b7da-707a3cd53d50" containerName="registry-server" containerID="cri-o://93d135bad0b701906eeaf645b58ba4cbe780321d11c41fa8713f046c41ca7e50" gracePeriod=2 Dec 05 09:41:17 crc kubenswrapper[4997]: I1205 09:41:17.517718 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-477xj" Dec 05 09:41:17 crc kubenswrapper[4997]: I1205 09:41:17.624397 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6k9v\" (UniqueName: \"kubernetes.io/projected/6d2a2784-93e2-411c-b7da-707a3cd53d50-kube-api-access-d6k9v\") pod \"6d2a2784-93e2-411c-b7da-707a3cd53d50\" (UID: \"6d2a2784-93e2-411c-b7da-707a3cd53d50\") " Dec 05 09:41:17 crc kubenswrapper[4997]: I1205 09:41:17.624544 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d2a2784-93e2-411c-b7da-707a3cd53d50-utilities\") pod \"6d2a2784-93e2-411c-b7da-707a3cd53d50\" (UID: \"6d2a2784-93e2-411c-b7da-707a3cd53d50\") " Dec 05 09:41:17 crc kubenswrapper[4997]: I1205 09:41:17.624575 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d2a2784-93e2-411c-b7da-707a3cd53d50-catalog-content\") pod \"6d2a2784-93e2-411c-b7da-707a3cd53d50\" (UID: \"6d2a2784-93e2-411c-b7da-707a3cd53d50\") " Dec 05 09:41:17 crc kubenswrapper[4997]: I1205 09:41:17.625632 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d2a2784-93e2-411c-b7da-707a3cd53d50-utilities" (OuterVolumeSpecName: "utilities") pod "6d2a2784-93e2-411c-b7da-707a3cd53d50" (UID: "6d2a2784-93e2-411c-b7da-707a3cd53d50"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:41:17 crc kubenswrapper[4997]: I1205 09:41:17.631814 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d2a2784-93e2-411c-b7da-707a3cd53d50-kube-api-access-d6k9v" (OuterVolumeSpecName: "kube-api-access-d6k9v") pod "6d2a2784-93e2-411c-b7da-707a3cd53d50" (UID: "6d2a2784-93e2-411c-b7da-707a3cd53d50"). InnerVolumeSpecName "kube-api-access-d6k9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:41:17 crc kubenswrapper[4997]: I1205 09:41:17.644142 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d2a2784-93e2-411c-b7da-707a3cd53d50-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d2a2784-93e2-411c-b7da-707a3cd53d50" (UID: "6d2a2784-93e2-411c-b7da-707a3cd53d50"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:41:17 crc kubenswrapper[4997]: I1205 09:41:17.727824 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6k9v\" (UniqueName: \"kubernetes.io/projected/6d2a2784-93e2-411c-b7da-707a3cd53d50-kube-api-access-d6k9v\") on node \"crc\" DevicePath \"\"" Dec 05 09:41:17 crc kubenswrapper[4997]: I1205 09:41:17.727865 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d2a2784-93e2-411c-b7da-707a3cd53d50-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:41:17 crc kubenswrapper[4997]: I1205 09:41:17.727875 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d2a2784-93e2-411c-b7da-707a3cd53d50-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:41:17 crc kubenswrapper[4997]: I1205 09:41:17.993907 4997 generic.go:334] "Generic (PLEG): container finished" podID="6d2a2784-93e2-411c-b7da-707a3cd53d50" containerID="93d135bad0b701906eeaf645b58ba4cbe780321d11c41fa8713f046c41ca7e50" exitCode=0 Dec 05 09:41:17 crc kubenswrapper[4997]: I1205 09:41:17.993963 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-477xj" event={"ID":"6d2a2784-93e2-411c-b7da-707a3cd53d50","Type":"ContainerDied","Data":"93d135bad0b701906eeaf645b58ba4cbe780321d11c41fa8713f046c41ca7e50"} Dec 05 09:41:17 crc kubenswrapper[4997]: I1205 09:41:17.993965 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-477xj" Dec 05 09:41:17 crc kubenswrapper[4997]: I1205 09:41:17.994004 4997 scope.go:117] "RemoveContainer" containerID="93d135bad0b701906eeaf645b58ba4cbe780321d11c41fa8713f046c41ca7e50" Dec 05 09:41:17 crc kubenswrapper[4997]: I1205 09:41:17.993993 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-477xj" event={"ID":"6d2a2784-93e2-411c-b7da-707a3cd53d50","Type":"ContainerDied","Data":"2d3cb4d9d533780e6122dc8b5049a00dd85471c4508c088e87fb839b9a9a8340"} Dec 05 09:41:18 crc kubenswrapper[4997]: I1205 09:41:18.025379 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-477xj"] Dec 05 09:41:18 crc kubenswrapper[4997]: I1205 09:41:18.025428 4997 scope.go:117] "RemoveContainer" containerID="fe3f520b8d1b67333b254fb63b4da289b0a0a4edc7a965d9ae1baaec32472bba" Dec 05 09:41:18 crc kubenswrapper[4997]: I1205 09:41:18.060110 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-477xj"] Dec 05 09:41:18 crc kubenswrapper[4997]: I1205 09:41:18.067041 4997 scope.go:117] "RemoveContainer" containerID="bdc199c0c28580ffc5da3ec5c120dd9adc9cde5d8345fa38f34941cf2a5a1ee7" Dec 05 09:41:18 crc kubenswrapper[4997]: I1205 09:41:18.098949 4997 scope.go:117] "RemoveContainer" containerID="93d135bad0b701906eeaf645b58ba4cbe780321d11c41fa8713f046c41ca7e50" Dec 05 09:41:18 crc kubenswrapper[4997]: E1205 09:41:18.099311 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93d135bad0b701906eeaf645b58ba4cbe780321d11c41fa8713f046c41ca7e50\": container with ID starting with 93d135bad0b701906eeaf645b58ba4cbe780321d11c41fa8713f046c41ca7e50 not found: ID does not exist" containerID="93d135bad0b701906eeaf645b58ba4cbe780321d11c41fa8713f046c41ca7e50" Dec 05 09:41:18 crc kubenswrapper[4997]: I1205 09:41:18.099360 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93d135bad0b701906eeaf645b58ba4cbe780321d11c41fa8713f046c41ca7e50"} err="failed to get container status \"93d135bad0b701906eeaf645b58ba4cbe780321d11c41fa8713f046c41ca7e50\": rpc error: code = NotFound desc = could not find container \"93d135bad0b701906eeaf645b58ba4cbe780321d11c41fa8713f046c41ca7e50\": container with ID starting with 93d135bad0b701906eeaf645b58ba4cbe780321d11c41fa8713f046c41ca7e50 not found: ID does not exist" Dec 05 09:41:18 crc kubenswrapper[4997]: I1205 09:41:18.099474 4997 scope.go:117] "RemoveContainer" containerID="fe3f520b8d1b67333b254fb63b4da289b0a0a4edc7a965d9ae1baaec32472bba" Dec 05 09:41:18 crc kubenswrapper[4997]: E1205 09:41:18.099824 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe3f520b8d1b67333b254fb63b4da289b0a0a4edc7a965d9ae1baaec32472bba\": container with ID starting with fe3f520b8d1b67333b254fb63b4da289b0a0a4edc7a965d9ae1baaec32472bba not found: ID does not exist" containerID="fe3f520b8d1b67333b254fb63b4da289b0a0a4edc7a965d9ae1baaec32472bba" Dec 05 09:41:18 crc kubenswrapper[4997]: I1205 09:41:18.099848 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe3f520b8d1b67333b254fb63b4da289b0a0a4edc7a965d9ae1baaec32472bba"} err="failed to get container status \"fe3f520b8d1b67333b254fb63b4da289b0a0a4edc7a965d9ae1baaec32472bba\": rpc error: code = NotFound desc = could not find container \"fe3f520b8d1b67333b254fb63b4da289b0a0a4edc7a965d9ae1baaec32472bba\": container with ID starting with fe3f520b8d1b67333b254fb63b4da289b0a0a4edc7a965d9ae1baaec32472bba not found: ID does not exist" Dec 05 09:41:18 crc kubenswrapper[4997]: I1205 09:41:18.099884 4997 scope.go:117] "RemoveContainer" containerID="bdc199c0c28580ffc5da3ec5c120dd9adc9cde5d8345fa38f34941cf2a5a1ee7" Dec 05 09:41:18 crc kubenswrapper[4997]: E1205 09:41:18.100089 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdc199c0c28580ffc5da3ec5c120dd9adc9cde5d8345fa38f34941cf2a5a1ee7\": container with ID starting with bdc199c0c28580ffc5da3ec5c120dd9adc9cde5d8345fa38f34941cf2a5a1ee7 not found: ID does not exist" containerID="bdc199c0c28580ffc5da3ec5c120dd9adc9cde5d8345fa38f34941cf2a5a1ee7" Dec 05 09:41:18 crc kubenswrapper[4997]: I1205 09:41:18.100124 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdc199c0c28580ffc5da3ec5c120dd9adc9cde5d8345fa38f34941cf2a5a1ee7"} err="failed to get container status \"bdc199c0c28580ffc5da3ec5c120dd9adc9cde5d8345fa38f34941cf2a5a1ee7\": rpc error: code = NotFound desc = could not find container \"bdc199c0c28580ffc5da3ec5c120dd9adc9cde5d8345fa38f34941cf2a5a1ee7\": container with ID starting with bdc199c0c28580ffc5da3ec5c120dd9adc9cde5d8345fa38f34941cf2a5a1ee7 not found: ID does not exist" Dec 05 09:41:19 crc kubenswrapper[4997]: I1205 09:41:19.762013 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d2a2784-93e2-411c-b7da-707a3cd53d50" path="/var/lib/kubelet/pods/6d2a2784-93e2-411c-b7da-707a3cd53d50/volumes" Dec 05 09:41:34 crc kubenswrapper[4997]: I1205 09:41:34.373714 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9p4fh"] Dec 05 09:41:34 crc kubenswrapper[4997]: E1205 09:41:34.374593 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d2a2784-93e2-411c-b7da-707a3cd53d50" containerName="extract-utilities" Dec 05 09:41:34 crc kubenswrapper[4997]: I1205 09:41:34.374607 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d2a2784-93e2-411c-b7da-707a3cd53d50" containerName="extract-utilities" Dec 05 09:41:34 crc kubenswrapper[4997]: E1205 09:41:34.374649 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d2a2784-93e2-411c-b7da-707a3cd53d50" containerName="extract-content" Dec 05 09:41:34 crc kubenswrapper[4997]: I1205 09:41:34.374655 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d2a2784-93e2-411c-b7da-707a3cd53d50" containerName="extract-content" Dec 05 09:41:34 crc kubenswrapper[4997]: E1205 09:41:34.374672 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d2a2784-93e2-411c-b7da-707a3cd53d50" containerName="registry-server" Dec 05 09:41:34 crc kubenswrapper[4997]: I1205 09:41:34.374678 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d2a2784-93e2-411c-b7da-707a3cd53d50" containerName="registry-server" Dec 05 09:41:34 crc kubenswrapper[4997]: I1205 09:41:34.374904 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d2a2784-93e2-411c-b7da-707a3cd53d50" containerName="registry-server" Dec 05 09:41:34 crc kubenswrapper[4997]: I1205 09:41:34.376452 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9p4fh" Dec 05 09:41:34 crc kubenswrapper[4997]: I1205 09:41:34.391676 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9p4fh"] Dec 05 09:41:34 crc kubenswrapper[4997]: I1205 09:41:34.483097 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6f152d6-402e-4008-8faa-e4535b1ce96d-catalog-content\") pod \"community-operators-9p4fh\" (UID: \"f6f152d6-402e-4008-8faa-e4535b1ce96d\") " pod="openshift-marketplace/community-operators-9p4fh" Dec 05 09:41:34 crc kubenswrapper[4997]: I1205 09:41:34.483413 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6f152d6-402e-4008-8faa-e4535b1ce96d-utilities\") pod \"community-operators-9p4fh\" (UID: \"f6f152d6-402e-4008-8faa-e4535b1ce96d\") " pod="openshift-marketplace/community-operators-9p4fh" Dec 05 09:41:34 crc kubenswrapper[4997]: I1205 09:41:34.483629 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knttz\" (UniqueName: \"kubernetes.io/projected/f6f152d6-402e-4008-8faa-e4535b1ce96d-kube-api-access-knttz\") pod \"community-operators-9p4fh\" (UID: \"f6f152d6-402e-4008-8faa-e4535b1ce96d\") " pod="openshift-marketplace/community-operators-9p4fh" Dec 05 09:41:34 crc kubenswrapper[4997]: I1205 09:41:34.586015 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6f152d6-402e-4008-8faa-e4535b1ce96d-catalog-content\") pod \"community-operators-9p4fh\" (UID: \"f6f152d6-402e-4008-8faa-e4535b1ce96d\") " pod="openshift-marketplace/community-operators-9p4fh" Dec 05 09:41:34 crc kubenswrapper[4997]: I1205 09:41:34.586159 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6f152d6-402e-4008-8faa-e4535b1ce96d-utilities\") pod \"community-operators-9p4fh\" (UID: \"f6f152d6-402e-4008-8faa-e4535b1ce96d\") " pod="openshift-marketplace/community-operators-9p4fh" Dec 05 09:41:34 crc kubenswrapper[4997]: I1205 09:41:34.586246 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knttz\" (UniqueName: \"kubernetes.io/projected/f6f152d6-402e-4008-8faa-e4535b1ce96d-kube-api-access-knttz\") pod \"community-operators-9p4fh\" (UID: \"f6f152d6-402e-4008-8faa-e4535b1ce96d\") " pod="openshift-marketplace/community-operators-9p4fh" Dec 05 09:41:34 crc kubenswrapper[4997]: I1205 09:41:34.586714 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6f152d6-402e-4008-8faa-e4535b1ce96d-catalog-content\") pod \"community-operators-9p4fh\" (UID: \"f6f152d6-402e-4008-8faa-e4535b1ce96d\") " pod="openshift-marketplace/community-operators-9p4fh" Dec 05 09:41:34 crc kubenswrapper[4997]: I1205 09:41:34.586734 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6f152d6-402e-4008-8faa-e4535b1ce96d-utilities\") pod \"community-operators-9p4fh\" (UID: \"f6f152d6-402e-4008-8faa-e4535b1ce96d\") " pod="openshift-marketplace/community-operators-9p4fh" Dec 05 09:41:34 crc kubenswrapper[4997]: I1205 09:41:34.606071 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knttz\" (UniqueName: \"kubernetes.io/projected/f6f152d6-402e-4008-8faa-e4535b1ce96d-kube-api-access-knttz\") pod \"community-operators-9p4fh\" (UID: \"f6f152d6-402e-4008-8faa-e4535b1ce96d\") " pod="openshift-marketplace/community-operators-9p4fh" Dec 05 09:41:34 crc kubenswrapper[4997]: I1205 09:41:34.706362 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9p4fh" Dec 05 09:41:36 crc kubenswrapper[4997]: I1205 09:41:36.076770 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9p4fh"] Dec 05 09:41:36 crc kubenswrapper[4997]: E1205 09:41:36.506056 4997 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6f152d6_402e_4008_8faa_e4535b1ce96d.slice/crio-4ab675055e264ea009da7e4f9ab3bd538bf5402b34fa3253d88a3d926fa30608.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6f152d6_402e_4008_8faa_e4535b1ce96d.slice/crio-conmon-4ab675055e264ea009da7e4f9ab3bd538bf5402b34fa3253d88a3d926fa30608.scope\": RecentStats: unable to find data in memory cache]" Dec 05 09:41:36 crc kubenswrapper[4997]: I1205 09:41:36.608369 4997 generic.go:334] "Generic (PLEG): container finished" podID="f6f152d6-402e-4008-8faa-e4535b1ce96d" containerID="4ab675055e264ea009da7e4f9ab3bd538bf5402b34fa3253d88a3d926fa30608" exitCode=0 Dec 05 09:41:36 crc kubenswrapper[4997]: I1205 09:41:36.608431 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p4fh" event={"ID":"f6f152d6-402e-4008-8faa-e4535b1ce96d","Type":"ContainerDied","Data":"4ab675055e264ea009da7e4f9ab3bd538bf5402b34fa3253d88a3d926fa30608"} Dec 05 09:41:36 crc kubenswrapper[4997]: I1205 09:41:36.608701 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p4fh" event={"ID":"f6f152d6-402e-4008-8faa-e4535b1ce96d","Type":"ContainerStarted","Data":"443675c01e36b54f1681f35cfbd73fc134e567270a0f1216883df4b20c9af9ed"} Dec 05 09:41:37 crc kubenswrapper[4997]: I1205 09:41:37.621629 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p4fh" event={"ID":"f6f152d6-402e-4008-8faa-e4535b1ce96d","Type":"ContainerStarted","Data":"e170f121b95b0294dab155b281cb806290b9c0e433fe3df0ad180c29627e7d0d"} Dec 05 09:41:38 crc kubenswrapper[4997]: I1205 09:41:38.633798 4997 generic.go:334] "Generic (PLEG): container finished" podID="f6f152d6-402e-4008-8faa-e4535b1ce96d" containerID="e170f121b95b0294dab155b281cb806290b9c0e433fe3df0ad180c29627e7d0d" exitCode=0 Dec 05 09:41:38 crc kubenswrapper[4997]: I1205 09:41:38.633901 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p4fh" event={"ID":"f6f152d6-402e-4008-8faa-e4535b1ce96d","Type":"ContainerDied","Data":"e170f121b95b0294dab155b281cb806290b9c0e433fe3df0ad180c29627e7d0d"} Dec 05 09:41:39 crc kubenswrapper[4997]: I1205 09:41:39.647165 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p4fh" event={"ID":"f6f152d6-402e-4008-8faa-e4535b1ce96d","Type":"ContainerStarted","Data":"808b3c643dd1e5a337b4799df129a7e0c3d3a8aa43ecbdf7c556128a32163cb1"} Dec 05 09:41:39 crc kubenswrapper[4997]: I1205 09:41:39.669241 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9p4fh" podStartSLOduration=3.260470986 podStartE2EDuration="5.669210635s" podCreationTimestamp="2025-12-05 09:41:34 +0000 UTC" firstStartedPulling="2025-12-05 09:41:36.610786834 +0000 UTC m=+9997.139694095" lastFinishedPulling="2025-12-05 09:41:39.019526483 +0000 UTC m=+9999.548433744" observedRunningTime="2025-12-05 09:41:39.666977654 +0000 UTC m=+10000.195884945" watchObservedRunningTime="2025-12-05 09:41:39.669210635 +0000 UTC m=+10000.198117926" Dec 05 09:41:44 crc kubenswrapper[4997]: I1205 09:41:44.707214 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9p4fh" Dec 05 09:41:44 crc kubenswrapper[4997]: I1205 09:41:44.707887 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9p4fh" Dec 05 09:41:44 crc kubenswrapper[4997]: I1205 09:41:44.755477 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9p4fh" Dec 05 09:41:45 crc kubenswrapper[4997]: I1205 09:41:45.760510 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9p4fh" Dec 05 09:41:45 crc kubenswrapper[4997]: I1205 09:41:45.801299 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9p4fh"] Dec 05 09:41:47 crc kubenswrapper[4997]: I1205 09:41:47.735915 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9p4fh" podUID="f6f152d6-402e-4008-8faa-e4535b1ce96d" containerName="registry-server" containerID="cri-o://808b3c643dd1e5a337b4799df129a7e0c3d3a8aa43ecbdf7c556128a32163cb1" gracePeriod=2 Dec 05 09:41:48 crc kubenswrapper[4997]: I1205 09:41:48.750132 4997 generic.go:334] "Generic (PLEG): container finished" podID="f6f152d6-402e-4008-8faa-e4535b1ce96d" containerID="808b3c643dd1e5a337b4799df129a7e0c3d3a8aa43ecbdf7c556128a32163cb1" exitCode=0 Dec 05 09:41:48 crc kubenswrapper[4997]: I1205 09:41:48.750213 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p4fh" event={"ID":"f6f152d6-402e-4008-8faa-e4535b1ce96d","Type":"ContainerDied","Data":"808b3c643dd1e5a337b4799df129a7e0c3d3a8aa43ecbdf7c556128a32163cb1"} Dec 05 09:41:48 crc kubenswrapper[4997]: I1205 09:41:48.750471 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p4fh" event={"ID":"f6f152d6-402e-4008-8faa-e4535b1ce96d","Type":"ContainerDied","Data":"443675c01e36b54f1681f35cfbd73fc134e567270a0f1216883df4b20c9af9ed"} Dec 05 09:41:48 crc kubenswrapper[4997]: I1205 09:41:48.750500 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="443675c01e36b54f1681f35cfbd73fc134e567270a0f1216883df4b20c9af9ed" Dec 05 09:41:49 crc kubenswrapper[4997]: I1205 09:41:49.020585 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9p4fh" Dec 05 09:41:49 crc kubenswrapper[4997]: I1205 09:41:49.188605 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knttz\" (UniqueName: \"kubernetes.io/projected/f6f152d6-402e-4008-8faa-e4535b1ce96d-kube-api-access-knttz\") pod \"f6f152d6-402e-4008-8faa-e4535b1ce96d\" (UID: \"f6f152d6-402e-4008-8faa-e4535b1ce96d\") " Dec 05 09:41:49 crc kubenswrapper[4997]: I1205 09:41:49.189000 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6f152d6-402e-4008-8faa-e4535b1ce96d-catalog-content\") pod \"f6f152d6-402e-4008-8faa-e4535b1ce96d\" (UID: \"f6f152d6-402e-4008-8faa-e4535b1ce96d\") " Dec 05 09:41:49 crc kubenswrapper[4997]: I1205 09:41:49.189414 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6f152d6-402e-4008-8faa-e4535b1ce96d-utilities\") pod \"f6f152d6-402e-4008-8faa-e4535b1ce96d\" (UID: \"f6f152d6-402e-4008-8faa-e4535b1ce96d\") " Dec 05 09:41:49 crc kubenswrapper[4997]: I1205 09:41:49.190718 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6f152d6-402e-4008-8faa-e4535b1ce96d-utilities" (OuterVolumeSpecName: "utilities") pod "f6f152d6-402e-4008-8faa-e4535b1ce96d" (UID: "f6f152d6-402e-4008-8faa-e4535b1ce96d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:41:49 crc kubenswrapper[4997]: I1205 09:41:49.195449 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6f152d6-402e-4008-8faa-e4535b1ce96d-kube-api-access-knttz" (OuterVolumeSpecName: "kube-api-access-knttz") pod "f6f152d6-402e-4008-8faa-e4535b1ce96d" (UID: "f6f152d6-402e-4008-8faa-e4535b1ce96d"). InnerVolumeSpecName "kube-api-access-knttz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:41:49 crc kubenswrapper[4997]: I1205 09:41:49.247231 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6f152d6-402e-4008-8faa-e4535b1ce96d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f6f152d6-402e-4008-8faa-e4535b1ce96d" (UID: "f6f152d6-402e-4008-8faa-e4535b1ce96d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:41:49 crc kubenswrapper[4997]: I1205 09:41:49.293418 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6f152d6-402e-4008-8faa-e4535b1ce96d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:41:49 crc kubenswrapper[4997]: I1205 09:41:49.293505 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knttz\" (UniqueName: \"kubernetes.io/projected/f6f152d6-402e-4008-8faa-e4535b1ce96d-kube-api-access-knttz\") on node \"crc\" DevicePath \"\"" Dec 05 09:41:49 crc kubenswrapper[4997]: I1205 09:41:49.293524 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6f152d6-402e-4008-8faa-e4535b1ce96d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:41:49 crc kubenswrapper[4997]: I1205 09:41:49.762358 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9p4fh" Dec 05 09:41:49 crc kubenswrapper[4997]: I1205 09:41:49.816097 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9p4fh"] Dec 05 09:41:49 crc kubenswrapper[4997]: I1205 09:41:49.830821 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9p4fh"] Dec 05 09:41:51 crc kubenswrapper[4997]: I1205 09:41:51.762879 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6f152d6-402e-4008-8faa-e4535b1ce96d" path="/var/lib/kubelet/pods/f6f152d6-402e-4008-8faa-e4535b1ce96d/volumes" Dec 05 09:42:19 crc kubenswrapper[4997]: I1205 09:42:19.770523 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:42:19 crc kubenswrapper[4997]: I1205 09:42:19.771764 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:42:49 crc kubenswrapper[4997]: I1205 09:42:49.770091 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:42:49 crc kubenswrapper[4997]: I1205 09:42:49.770758 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:43:19 crc kubenswrapper[4997]: I1205 09:43:19.769923 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:43:19 crc kubenswrapper[4997]: I1205 09:43:19.770533 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:43:19 crc kubenswrapper[4997]: I1205 09:43:19.770579 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 09:43:19 crc kubenswrapper[4997]: I1205 09:43:19.771475 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"553255d50247709601e534c39b4889a350b0020aa3393fb72def52e3ad4252bf"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:43:19 crc kubenswrapper[4997]: I1205 09:43:19.771530 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://553255d50247709601e534c39b4889a350b0020aa3393fb72def52e3ad4252bf" gracePeriod=600 Dec 05 09:43:20 crc kubenswrapper[4997]: I1205 09:43:20.710292 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="553255d50247709601e534c39b4889a350b0020aa3393fb72def52e3ad4252bf" exitCode=0 Dec 05 09:43:20 crc kubenswrapper[4997]: I1205 09:43:20.710389 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"553255d50247709601e534c39b4889a350b0020aa3393fb72def52e3ad4252bf"} Dec 05 09:43:20 crc kubenswrapper[4997]: I1205 09:43:20.710918 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86"} Dec 05 09:43:20 crc kubenswrapper[4997]: I1205 09:43:20.710952 4997 scope.go:117] "RemoveContainer" containerID="0822c5878ae80208264d2213716e24ff4a95a198054631a1c58f61f48d43ec06" Dec 05 09:44:17 crc kubenswrapper[4997]: I1205 09:44:17.316732 4997 generic.go:334] "Generic (PLEG): container finished" podID="746846bb-be60-4333-a253-f1f3057ea2d7" containerID="3c63b6c9a5471fc477c08f1de711354a627c0d94cd2ce3c4a62e1c3c145659a9" exitCode=0 Dec 05 09:44:17 crc kubenswrapper[4997]: I1205 09:44:17.316803 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" event={"ID":"746846bb-be60-4333-a253-f1f3057ea2d7","Type":"ContainerDied","Data":"3c63b6c9a5471fc477c08f1de711354a627c0d94cd2ce3c4a62e1c3c145659a9"} Dec 05 09:44:18 crc kubenswrapper[4997]: I1205 09:44:18.956187 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.122320 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-inventory\") pod \"746846bb-be60-4333-a253-f1f3057ea2d7\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.122371 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-0\") pod \"746846bb-be60-4333-a253-f1f3057ea2d7\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.122566 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceph\") pod \"746846bb-be60-4333-a253-f1f3057ea2d7\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.122665 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-telemetry-combined-ca-bundle\") pod \"746846bb-be60-4333-a253-f1f3057ea2d7\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.122705 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7nc6\" (UniqueName: \"kubernetes.io/projected/746846bb-be60-4333-a253-f1f3057ea2d7-kube-api-access-q7nc6\") pod \"746846bb-be60-4333-a253-f1f3057ea2d7\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.122743 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-2\") pod \"746846bb-be60-4333-a253-f1f3057ea2d7\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.122778 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ssh-key\") pod \"746846bb-be60-4333-a253-f1f3057ea2d7\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.122858 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-1\") pod \"746846bb-be60-4333-a253-f1f3057ea2d7\" (UID: \"746846bb-be60-4333-a253-f1f3057ea2d7\") " Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.128165 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceph" (OuterVolumeSpecName: "ceph") pod "746846bb-be60-4333-a253-f1f3057ea2d7" (UID: "746846bb-be60-4333-a253-f1f3057ea2d7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.128220 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "746846bb-be60-4333-a253-f1f3057ea2d7" (UID: "746846bb-be60-4333-a253-f1f3057ea2d7"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.128538 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/746846bb-be60-4333-a253-f1f3057ea2d7-kube-api-access-q7nc6" (OuterVolumeSpecName: "kube-api-access-q7nc6") pod "746846bb-be60-4333-a253-f1f3057ea2d7" (UID: "746846bb-be60-4333-a253-f1f3057ea2d7"). InnerVolumeSpecName "kube-api-access-q7nc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.154630 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "746846bb-be60-4333-a253-f1f3057ea2d7" (UID: "746846bb-be60-4333-a253-f1f3057ea2d7"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.156117 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "746846bb-be60-4333-a253-f1f3057ea2d7" (UID: "746846bb-be60-4333-a253-f1f3057ea2d7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.163272 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-inventory" (OuterVolumeSpecName: "inventory") pod "746846bb-be60-4333-a253-f1f3057ea2d7" (UID: "746846bb-be60-4333-a253-f1f3057ea2d7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.165108 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "746846bb-be60-4333-a253-f1f3057ea2d7" (UID: "746846bb-be60-4333-a253-f1f3057ea2d7"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.183341 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "746846bb-be60-4333-a253-f1f3057ea2d7" (UID: "746846bb-be60-4333-a253-f1f3057ea2d7"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.225512 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.225546 4997 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.225558 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7nc6\" (UniqueName: \"kubernetes.io/projected/746846bb-be60-4333-a253-f1f3057ea2d7-kube-api-access-q7nc6\") on node \"crc\" DevicePath \"\"" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.225566 4997 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.225574 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.225583 4997 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.225591 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.225600 4997 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/746846bb-be60-4333-a253-f1f3057ea2d7-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.365977 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" event={"ID":"746846bb-be60-4333-a253-f1f3057ea2d7","Type":"ContainerDied","Data":"2681586441f965014631ff6922da8a9f04bf287d82ab44d2f01693e20b948b81"} Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.366596 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2681586441f965014631ff6922da8a9f04bf287d82ab44d2f01693e20b948b81" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.366293 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-q9b5m" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.470045 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-qmm94"] Dec 05 09:44:19 crc kubenswrapper[4997]: E1205 09:44:19.470608 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6f152d6-402e-4008-8faa-e4535b1ce96d" containerName="registry-server" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.470688 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6f152d6-402e-4008-8faa-e4535b1ce96d" containerName="registry-server" Dec 05 09:44:19 crc kubenswrapper[4997]: E1205 09:44:19.470720 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="746846bb-be60-4333-a253-f1f3057ea2d7" containerName="telemetry-openstack-openstack-cell1" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.470729 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="746846bb-be60-4333-a253-f1f3057ea2d7" containerName="telemetry-openstack-openstack-cell1" Dec 05 09:44:19 crc kubenswrapper[4997]: E1205 09:44:19.470743 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6f152d6-402e-4008-8faa-e4535b1ce96d" containerName="extract-content" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.470751 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6f152d6-402e-4008-8faa-e4535b1ce96d" containerName="extract-content" Dec 05 09:44:19 crc kubenswrapper[4997]: E1205 09:44:19.470784 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6f152d6-402e-4008-8faa-e4535b1ce96d" containerName="extract-utilities" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.470793 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6f152d6-402e-4008-8faa-e4535b1ce96d" containerName="extract-utilities" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.471075 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="746846bb-be60-4333-a253-f1f3057ea2d7" containerName="telemetry-openstack-openstack-cell1" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.471101 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6f152d6-402e-4008-8faa-e4535b1ce96d" containerName="registry-server" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.472054 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.481957 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-qmm94"] Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.484012 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.484133 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.484313 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.484339 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.484639 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.531591 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnjh4\" (UniqueName: \"kubernetes.io/projected/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-kube-api-access-nnjh4\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.531660 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.531706 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.531843 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.531869 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.531923 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.635601 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.635680 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.635757 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.635813 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnjh4\" (UniqueName: \"kubernetes.io/projected/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-kube-api-access-nnjh4\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.635842 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:19 crc kubenswrapper[4997]: I1205 09:44:19.636835 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:20 crc kubenswrapper[4997]: I1205 09:44:20.103000 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:20 crc kubenswrapper[4997]: I1205 09:44:20.103081 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:20 crc kubenswrapper[4997]: I1205 09:44:20.103220 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:20 crc kubenswrapper[4997]: I1205 09:44:20.103648 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:20 crc kubenswrapper[4997]: I1205 09:44:20.103821 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnjh4\" (UniqueName: \"kubernetes.io/projected/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-kube-api-access-nnjh4\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:20 crc kubenswrapper[4997]: I1205 09:44:20.116736 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-qmm94\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:20 crc kubenswrapper[4997]: I1205 09:44:20.399675 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:44:20 crc kubenswrapper[4997]: I1205 09:44:20.934244 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-qmm94"] Dec 05 09:44:20 crc kubenswrapper[4997]: I1205 09:44:20.938014 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:44:21 crc kubenswrapper[4997]: I1205 09:44:21.388867 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" event={"ID":"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337","Type":"ContainerStarted","Data":"39d74200ef9d78b122a0197a28c8428a0ed4059d0f05b57d1f28f93a287b5609"} Dec 05 09:44:22 crc kubenswrapper[4997]: I1205 09:44:22.399745 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" event={"ID":"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337","Type":"ContainerStarted","Data":"bbdf5b7646453519efa3c66bd043922ae140fe247a0d8f998decac93d8bfd16d"} Dec 05 09:44:22 crc kubenswrapper[4997]: I1205 09:44:22.422747 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" podStartSLOduration=2.961076246 podStartE2EDuration="3.422725671s" podCreationTimestamp="2025-12-05 09:44:19 +0000 UTC" firstStartedPulling="2025-12-05 09:44:20.937775765 +0000 UTC m=+10161.466683026" lastFinishedPulling="2025-12-05 09:44:21.39942519 +0000 UTC m=+10161.928332451" observedRunningTime="2025-12-05 09:44:22.416667807 +0000 UTC m=+10162.945575088" watchObservedRunningTime="2025-12-05 09:44:22.422725671 +0000 UTC m=+10162.951632952" Dec 05 09:45:00 crc kubenswrapper[4997]: I1205 09:45:00.143726 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4"] Dec 05 09:45:00 crc kubenswrapper[4997]: I1205 09:45:00.145884 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" Dec 05 09:45:00 crc kubenswrapper[4997]: I1205 09:45:00.154657 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 09:45:00 crc kubenswrapper[4997]: I1205 09:45:00.154799 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 09:45:00 crc kubenswrapper[4997]: I1205 09:45:00.177342 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4"] Dec 05 09:45:00 crc kubenswrapper[4997]: I1205 09:45:00.310730 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a04c950b-c9b4-4052-bad7-53a68c7568c5-secret-volume\") pod \"collect-profiles-29415465-w57s4\" (UID: \"a04c950b-c9b4-4052-bad7-53a68c7568c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" Dec 05 09:45:00 crc kubenswrapper[4997]: I1205 09:45:00.311133 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a04c950b-c9b4-4052-bad7-53a68c7568c5-config-volume\") pod \"collect-profiles-29415465-w57s4\" (UID: \"a04c950b-c9b4-4052-bad7-53a68c7568c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" Dec 05 09:45:00 crc kubenswrapper[4997]: I1205 09:45:00.311304 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sshp5\" (UniqueName: \"kubernetes.io/projected/a04c950b-c9b4-4052-bad7-53a68c7568c5-kube-api-access-sshp5\") pod \"collect-profiles-29415465-w57s4\" (UID: \"a04c950b-c9b4-4052-bad7-53a68c7568c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" Dec 05 09:45:00 crc kubenswrapper[4997]: I1205 09:45:00.413552 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a04c950b-c9b4-4052-bad7-53a68c7568c5-config-volume\") pod \"collect-profiles-29415465-w57s4\" (UID: \"a04c950b-c9b4-4052-bad7-53a68c7568c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" Dec 05 09:45:00 crc kubenswrapper[4997]: I1205 09:45:00.413692 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sshp5\" (UniqueName: \"kubernetes.io/projected/a04c950b-c9b4-4052-bad7-53a68c7568c5-kube-api-access-sshp5\") pod \"collect-profiles-29415465-w57s4\" (UID: \"a04c950b-c9b4-4052-bad7-53a68c7568c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" Dec 05 09:45:00 crc kubenswrapper[4997]: I1205 09:45:00.413803 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a04c950b-c9b4-4052-bad7-53a68c7568c5-secret-volume\") pod \"collect-profiles-29415465-w57s4\" (UID: \"a04c950b-c9b4-4052-bad7-53a68c7568c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" Dec 05 09:45:00 crc kubenswrapper[4997]: I1205 09:45:00.414987 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a04c950b-c9b4-4052-bad7-53a68c7568c5-config-volume\") pod \"collect-profiles-29415465-w57s4\" (UID: \"a04c950b-c9b4-4052-bad7-53a68c7568c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" Dec 05 09:45:00 crc kubenswrapper[4997]: I1205 09:45:00.420093 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a04c950b-c9b4-4052-bad7-53a68c7568c5-secret-volume\") pod \"collect-profiles-29415465-w57s4\" (UID: \"a04c950b-c9b4-4052-bad7-53a68c7568c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" Dec 05 09:45:00 crc kubenswrapper[4997]: I1205 09:45:00.434381 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sshp5\" (UniqueName: \"kubernetes.io/projected/a04c950b-c9b4-4052-bad7-53a68c7568c5-kube-api-access-sshp5\") pod \"collect-profiles-29415465-w57s4\" (UID: \"a04c950b-c9b4-4052-bad7-53a68c7568c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" Dec 05 09:45:00 crc kubenswrapper[4997]: I1205 09:45:00.477926 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" Dec 05 09:45:00 crc kubenswrapper[4997]: I1205 09:45:00.959927 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4"] Dec 05 09:45:01 crc kubenswrapper[4997]: I1205 09:45:01.813704 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" event={"ID":"a04c950b-c9b4-4052-bad7-53a68c7568c5","Type":"ContainerStarted","Data":"a9cfe68a24f23da0b5b08d562f31949505a689769f05eec7b642770c3df12856"} Dec 05 09:45:01 crc kubenswrapper[4997]: I1205 09:45:01.814096 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" event={"ID":"a04c950b-c9b4-4052-bad7-53a68c7568c5","Type":"ContainerStarted","Data":"36f24d29b10533089475e75492a25f379f4b3305d2743ae4544ccac51abfb2e6"} Dec 05 09:45:01 crc kubenswrapper[4997]: I1205 09:45:01.836641 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" podStartSLOduration=1.836596642 podStartE2EDuration="1.836596642s" podCreationTimestamp="2025-12-05 09:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:45:01.832242304 +0000 UTC m=+10202.361149585" watchObservedRunningTime="2025-12-05 09:45:01.836596642 +0000 UTC m=+10202.365503903" Dec 05 09:45:02 crc kubenswrapper[4997]: I1205 09:45:02.822997 4997 generic.go:334] "Generic (PLEG): container finished" podID="a04c950b-c9b4-4052-bad7-53a68c7568c5" containerID="a9cfe68a24f23da0b5b08d562f31949505a689769f05eec7b642770c3df12856" exitCode=0 Dec 05 09:45:02 crc kubenswrapper[4997]: I1205 09:45:02.823031 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" event={"ID":"a04c950b-c9b4-4052-bad7-53a68c7568c5","Type":"ContainerDied","Data":"a9cfe68a24f23da0b5b08d562f31949505a689769f05eec7b642770c3df12856"} Dec 05 09:45:04 crc kubenswrapper[4997]: I1205 09:45:04.229343 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" Dec 05 09:45:04 crc kubenswrapper[4997]: I1205 09:45:04.289169 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sshp5\" (UniqueName: \"kubernetes.io/projected/a04c950b-c9b4-4052-bad7-53a68c7568c5-kube-api-access-sshp5\") pod \"a04c950b-c9b4-4052-bad7-53a68c7568c5\" (UID: \"a04c950b-c9b4-4052-bad7-53a68c7568c5\") " Dec 05 09:45:04 crc kubenswrapper[4997]: I1205 09:45:04.289393 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a04c950b-c9b4-4052-bad7-53a68c7568c5-secret-volume\") pod \"a04c950b-c9b4-4052-bad7-53a68c7568c5\" (UID: \"a04c950b-c9b4-4052-bad7-53a68c7568c5\") " Dec 05 09:45:04 crc kubenswrapper[4997]: I1205 09:45:04.289419 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a04c950b-c9b4-4052-bad7-53a68c7568c5-config-volume\") pod \"a04c950b-c9b4-4052-bad7-53a68c7568c5\" (UID: \"a04c950b-c9b4-4052-bad7-53a68c7568c5\") " Dec 05 09:45:04 crc kubenswrapper[4997]: I1205 09:45:04.290402 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a04c950b-c9b4-4052-bad7-53a68c7568c5-config-volume" (OuterVolumeSpecName: "config-volume") pod "a04c950b-c9b4-4052-bad7-53a68c7568c5" (UID: "a04c950b-c9b4-4052-bad7-53a68c7568c5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:45:04 crc kubenswrapper[4997]: I1205 09:45:04.291136 4997 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a04c950b-c9b4-4052-bad7-53a68c7568c5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:45:04 crc kubenswrapper[4997]: I1205 09:45:04.295022 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a04c950b-c9b4-4052-bad7-53a68c7568c5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a04c950b-c9b4-4052-bad7-53a68c7568c5" (UID: "a04c950b-c9b4-4052-bad7-53a68c7568c5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:45:04 crc kubenswrapper[4997]: I1205 09:45:04.301389 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a04c950b-c9b4-4052-bad7-53a68c7568c5-kube-api-access-sshp5" (OuterVolumeSpecName: "kube-api-access-sshp5") pod "a04c950b-c9b4-4052-bad7-53a68c7568c5" (UID: "a04c950b-c9b4-4052-bad7-53a68c7568c5"). InnerVolumeSpecName "kube-api-access-sshp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:45:04 crc kubenswrapper[4997]: I1205 09:45:04.393399 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sshp5\" (UniqueName: \"kubernetes.io/projected/a04c950b-c9b4-4052-bad7-53a68c7568c5-kube-api-access-sshp5\") on node \"crc\" DevicePath \"\"" Dec 05 09:45:04 crc kubenswrapper[4997]: I1205 09:45:04.393428 4997 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a04c950b-c9b4-4052-bad7-53a68c7568c5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:45:04 crc kubenswrapper[4997]: I1205 09:45:04.867316 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" event={"ID":"a04c950b-c9b4-4052-bad7-53a68c7568c5","Type":"ContainerDied","Data":"36f24d29b10533089475e75492a25f379f4b3305d2743ae4544ccac51abfb2e6"} Dec 05 09:45:04 crc kubenswrapper[4997]: I1205 09:45:04.867369 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36f24d29b10533089475e75492a25f379f4b3305d2743ae4544ccac51abfb2e6" Dec 05 09:45:04 crc kubenswrapper[4997]: I1205 09:45:04.867413 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4" Dec 05 09:45:04 crc kubenswrapper[4997]: I1205 09:45:04.912174 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t"] Dec 05 09:45:04 crc kubenswrapper[4997]: I1205 09:45:04.926220 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415420-cqs2t"] Dec 05 09:45:05 crc kubenswrapper[4997]: I1205 09:45:05.762034 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dc60482-e3d1-455f-9d81-58ef71b08668" path="/var/lib/kubelet/pods/4dc60482-e3d1-455f-9d81-58ef71b08668/volumes" Dec 05 09:45:06 crc kubenswrapper[4997]: I1205 09:45:06.888740 4997 generic.go:334] "Generic (PLEG): container finished" podID="8f306f28-44f2-4a8c-95e2-9b6d1c7a5337" containerID="bbdf5b7646453519efa3c66bd043922ae140fe247a0d8f998decac93d8bfd16d" exitCode=0 Dec 05 09:45:06 crc kubenswrapper[4997]: I1205 09:45:06.888793 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" event={"ID":"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337","Type":"ContainerDied","Data":"bbdf5b7646453519efa3c66bd043922ae140fe247a0d8f998decac93d8bfd16d"} Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.394225 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.483530 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-ceph\") pod \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.483602 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-ssh-key\") pod \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.483654 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-neutron-sriov-agent-neutron-config-0\") pod \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.483713 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-neutron-sriov-combined-ca-bundle\") pod \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.483737 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnjh4\" (UniqueName: \"kubernetes.io/projected/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-kube-api-access-nnjh4\") pod \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.483759 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-inventory\") pod \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\" (UID: \"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337\") " Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.489790 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-kube-api-access-nnjh4" (OuterVolumeSpecName: "kube-api-access-nnjh4") pod "8f306f28-44f2-4a8c-95e2-9b6d1c7a5337" (UID: "8f306f28-44f2-4a8c-95e2-9b6d1c7a5337"). InnerVolumeSpecName "kube-api-access-nnjh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.489916 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-ceph" (OuterVolumeSpecName: "ceph") pod "8f306f28-44f2-4a8c-95e2-9b6d1c7a5337" (UID: "8f306f28-44f2-4a8c-95e2-9b6d1c7a5337"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.500799 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "8f306f28-44f2-4a8c-95e2-9b6d1c7a5337" (UID: "8f306f28-44f2-4a8c-95e2-9b6d1c7a5337"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.517908 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "8f306f28-44f2-4a8c-95e2-9b6d1c7a5337" (UID: "8f306f28-44f2-4a8c-95e2-9b6d1c7a5337"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.518757 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8f306f28-44f2-4a8c-95e2-9b6d1c7a5337" (UID: "8f306f28-44f2-4a8c-95e2-9b6d1c7a5337"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.528159 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-inventory" (OuterVolumeSpecName: "inventory") pod "8f306f28-44f2-4a8c-95e2-9b6d1c7a5337" (UID: "8f306f28-44f2-4a8c-95e2-9b6d1c7a5337"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.586339 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.586655 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.586668 4997 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.586680 4997 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.586692 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnjh4\" (UniqueName: \"kubernetes.io/projected/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-kube-api-access-nnjh4\") on node \"crc\" DevicePath \"\"" Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.586702 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f306f28-44f2-4a8c-95e2-9b6d1c7a5337-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.908976 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" event={"ID":"8f306f28-44f2-4a8c-95e2-9b6d1c7a5337","Type":"ContainerDied","Data":"39d74200ef9d78b122a0197a28c8428a0ed4059d0f05b57d1f28f93a287b5609"} Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.909023 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39d74200ef9d78b122a0197a28c8428a0ed4059d0f05b57d1f28f93a287b5609" Dec 05 09:45:08 crc kubenswrapper[4997]: I1205 09:45:08.909112 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-qmm94" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.014534 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc"] Dec 05 09:45:09 crc kubenswrapper[4997]: E1205 09:45:09.015767 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a04c950b-c9b4-4052-bad7-53a68c7568c5" containerName="collect-profiles" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.015795 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a04c950b-c9b4-4052-bad7-53a68c7568c5" containerName="collect-profiles" Dec 05 09:45:09 crc kubenswrapper[4997]: E1205 09:45:09.015808 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f306f28-44f2-4a8c-95e2-9b6d1c7a5337" containerName="neutron-sriov-openstack-openstack-cell1" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.015816 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f306f28-44f2-4a8c-95e2-9b6d1c7a5337" containerName="neutron-sriov-openstack-openstack-cell1" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.016100 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a04c950b-c9b4-4052-bad7-53a68c7568c5" containerName="collect-profiles" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.016141 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f306f28-44f2-4a8c-95e2-9b6d1c7a5337" containerName="neutron-sriov-openstack-openstack-cell1" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.017283 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.020495 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.020781 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.020809 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.020905 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.020978 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.029830 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc"] Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.098253 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrk6c\" (UniqueName: \"kubernetes.io/projected/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-kube-api-access-nrk6c\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.098320 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.098354 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.098434 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.098503 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.098580 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.200861 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrk6c\" (UniqueName: \"kubernetes.io/projected/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-kube-api-access-nrk6c\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.200992 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.201031 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.201108 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.201183 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.201249 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.204944 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.205567 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.205729 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.206119 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.212971 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.231180 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrk6c\" (UniqueName: \"kubernetes.io/projected/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-kube-api-access-nrk6c\") pod \"neutron-dhcp-openstack-openstack-cell1-qbcxc\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.355195 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.885163 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc"] Dec 05 09:45:09 crc kubenswrapper[4997]: I1205 09:45:09.924087 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" event={"ID":"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8","Type":"ContainerStarted","Data":"033259516caac56fc74bbb1b97c304efa0de8bfe37b6446292e29f3279bbf9d3"} Dec 05 09:45:10 crc kubenswrapper[4997]: I1205 09:45:10.935496 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" event={"ID":"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8","Type":"ContainerStarted","Data":"cbde4db04b087b76a13a27780d782e6bac51144e64ff3e50cf4e0b7e00166e1a"} Dec 05 09:45:10 crc kubenswrapper[4997]: I1205 09:45:10.964868 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" podStartSLOduration=2.5793855470000002 podStartE2EDuration="2.964840266s" podCreationTimestamp="2025-12-05 09:45:08 +0000 UTC" firstStartedPulling="2025-12-05 09:45:09.886254678 +0000 UTC m=+10210.415161929" lastFinishedPulling="2025-12-05 09:45:10.271709387 +0000 UTC m=+10210.800616648" observedRunningTime="2025-12-05 09:45:10.956411368 +0000 UTC m=+10211.485318639" watchObservedRunningTime="2025-12-05 09:45:10.964840266 +0000 UTC m=+10211.493747547" Dec 05 09:45:20 crc kubenswrapper[4997]: I1205 09:45:20.447528 4997 scope.go:117] "RemoveContainer" containerID="e92ba1096f1801b80d0ee209fb07cd20aaed5df90ee61f5e60c6db5269749d9e" Dec 05 09:45:49 crc kubenswrapper[4997]: I1205 09:45:49.770464 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:45:49 crc kubenswrapper[4997]: I1205 09:45:49.771289 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:46:08 crc kubenswrapper[4997]: I1205 09:46:08.492868 4997 generic.go:334] "Generic (PLEG): container finished" podID="f6398c5c-8fdf-49dc-8de8-c0bc70637ac8" containerID="cbde4db04b087b76a13a27780d782e6bac51144e64ff3e50cf4e0b7e00166e1a" exitCode=0 Dec 05 09:46:08 crc kubenswrapper[4997]: I1205 09:46:08.492972 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" event={"ID":"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8","Type":"ContainerDied","Data":"cbde4db04b087b76a13a27780d782e6bac51144e64ff3e50cf4e0b7e00166e1a"} Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.029976 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.051652 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-ceph\") pod \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.052017 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-neutron-dhcp-agent-neutron-config-0\") pod \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.052170 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-neutron-dhcp-combined-ca-bundle\") pod \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.052371 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-ssh-key\") pod \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.052469 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrk6c\" (UniqueName: \"kubernetes.io/projected/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-kube-api-access-nrk6c\") pod \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.052563 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-inventory\") pod \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\" (UID: \"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8\") " Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.057560 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-ceph" (OuterVolumeSpecName: "ceph") pod "f6398c5c-8fdf-49dc-8de8-c0bc70637ac8" (UID: "f6398c5c-8fdf-49dc-8de8-c0bc70637ac8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.062557 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-kube-api-access-nrk6c" (OuterVolumeSpecName: "kube-api-access-nrk6c") pod "f6398c5c-8fdf-49dc-8de8-c0bc70637ac8" (UID: "f6398c5c-8fdf-49dc-8de8-c0bc70637ac8"). InnerVolumeSpecName "kube-api-access-nrk6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.062682 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "f6398c5c-8fdf-49dc-8de8-c0bc70637ac8" (UID: "f6398c5c-8fdf-49dc-8de8-c0bc70637ac8"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.086328 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-inventory" (OuterVolumeSpecName: "inventory") pod "f6398c5c-8fdf-49dc-8de8-c0bc70637ac8" (UID: "f6398c5c-8fdf-49dc-8de8-c0bc70637ac8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.087991 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f6398c5c-8fdf-49dc-8de8-c0bc70637ac8" (UID: "f6398c5c-8fdf-49dc-8de8-c0bc70637ac8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.090744 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "f6398c5c-8fdf-49dc-8de8-c0bc70637ac8" (UID: "f6398c5c-8fdf-49dc-8de8-c0bc70637ac8"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.155914 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.155955 4997 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.155967 4997 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.155976 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.155987 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrk6c\" (UniqueName: \"kubernetes.io/projected/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-kube-api-access-nrk6c\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.155997 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6398c5c-8fdf-49dc-8de8-c0bc70637ac8-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.553061 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" event={"ID":"f6398c5c-8fdf-49dc-8de8-c0bc70637ac8","Type":"ContainerDied","Data":"033259516caac56fc74bbb1b97c304efa0de8bfe37b6446292e29f3279bbf9d3"} Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.553125 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="033259516caac56fc74bbb1b97c304efa0de8bfe37b6446292e29f3279bbf9d3" Dec 05 09:46:10 crc kubenswrapper[4997]: I1205 09:46:10.553131 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qbcxc" Dec 05 09:46:13 crc kubenswrapper[4997]: I1205 09:46:13.781569 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 09:46:13 crc kubenswrapper[4997]: I1205 09:46:13.782284 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="04d5f9e7-d90b-4887-af25-cbf4e8a16db0" containerName="nova-cell0-conductor-conductor" containerID="cri-o://28b07c5c22ab9328db79dbaaf5998a6025761c2994ace11bb7e357dd95249f1a" gracePeriod=30 Dec 05 09:46:13 crc kubenswrapper[4997]: I1205 09:46:13.828218 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 09:46:13 crc kubenswrapper[4997]: I1205 09:46:13.828441 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="7a6e7c78-6f77-4d5e-96f8-95f67debcd97" containerName="nova-cell1-conductor-conductor" containerID="cri-o://a765187e12ab594edbd3c92b89d19f7b86dabdc712477b007ef3882121061020" gracePeriod=30 Dec 05 09:46:14 crc kubenswrapper[4997]: I1205 09:46:14.595178 4997 generic.go:334] "Generic (PLEG): container finished" podID="7a6e7c78-6f77-4d5e-96f8-95f67debcd97" containerID="a765187e12ab594edbd3c92b89d19f7b86dabdc712477b007ef3882121061020" exitCode=0 Dec 05 09:46:14 crc kubenswrapper[4997]: I1205 09:46:14.595248 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7a6e7c78-6f77-4d5e-96f8-95f67debcd97","Type":"ContainerDied","Data":"a765187e12ab594edbd3c92b89d19f7b86dabdc712477b007ef3882121061020"} Dec 05 09:46:14 crc kubenswrapper[4997]: I1205 09:46:14.797094 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 09:46:14 crc kubenswrapper[4997]: I1205 09:46:14.855926 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgzmq\" (UniqueName: \"kubernetes.io/projected/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-kube-api-access-lgzmq\") pod \"7a6e7c78-6f77-4d5e-96f8-95f67debcd97\" (UID: \"7a6e7c78-6f77-4d5e-96f8-95f67debcd97\") " Dec 05 09:46:14 crc kubenswrapper[4997]: I1205 09:46:14.856016 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-combined-ca-bundle\") pod \"7a6e7c78-6f77-4d5e-96f8-95f67debcd97\" (UID: \"7a6e7c78-6f77-4d5e-96f8-95f67debcd97\") " Dec 05 09:46:14 crc kubenswrapper[4997]: I1205 09:46:14.856059 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-config-data\") pod \"7a6e7c78-6f77-4d5e-96f8-95f67debcd97\" (UID: \"7a6e7c78-6f77-4d5e-96f8-95f67debcd97\") " Dec 05 09:46:14 crc kubenswrapper[4997]: I1205 09:46:14.875068 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-kube-api-access-lgzmq" (OuterVolumeSpecName: "kube-api-access-lgzmq") pod "7a6e7c78-6f77-4d5e-96f8-95f67debcd97" (UID: "7a6e7c78-6f77-4d5e-96f8-95f67debcd97"). InnerVolumeSpecName "kube-api-access-lgzmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:46:14 crc kubenswrapper[4997]: I1205 09:46:14.897172 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a6e7c78-6f77-4d5e-96f8-95f67debcd97" (UID: "7a6e7c78-6f77-4d5e-96f8-95f67debcd97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:46:14 crc kubenswrapper[4997]: I1205 09:46:14.901701 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-config-data" (OuterVolumeSpecName: "config-data") pod "7a6e7c78-6f77-4d5e-96f8-95f67debcd97" (UID: "7a6e7c78-6f77-4d5e-96f8-95f67debcd97"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:46:14 crc kubenswrapper[4997]: I1205 09:46:14.960043 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgzmq\" (UniqueName: \"kubernetes.io/projected/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-kube-api-access-lgzmq\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:14 crc kubenswrapper[4997]: I1205 09:46:14.960080 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:14 crc kubenswrapper[4997]: I1205 09:46:14.960092 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6e7c78-6f77-4d5e-96f8-95f67debcd97-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.042276 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.043600 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f3cbe3c8-d087-461e-a663-c0e8229640b4" containerName="nova-api-api" containerID="cri-o://6deab0d56789b3905d671ebbc986574fd95edd3cc69a8e63b485e79625592e1d" gracePeriod=30 Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.043536 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f3cbe3c8-d087-461e-a663-c0e8229640b4" containerName="nova-api-log" containerID="cri-o://944cbd4353d191849652e26d5cd7806761de0e71e59a72861dcce5e9f3b7a7ff" gracePeriod=30 Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.069872 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.070216 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="51c4332f-d8d3-479f-af91-3314d7111adc" containerName="nova-scheduler-scheduler" containerID="cri-o://35db5a225a2271f9e2cace9eeead88d9507db1e4658527d66fe51579558496fb" gracePeriod=30 Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.086134 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.086394 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="79d909f6-46e3-401d-aafa-3d649b632284" containerName="nova-metadata-log" containerID="cri-o://a78ece514e46e4b92e3bb36eb7f2ecd2ff6c5bd515c06adbe710b954d25f52f6" gracePeriod=30 Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.086520 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="79d909f6-46e3-401d-aafa-3d649b632284" containerName="nova-metadata-metadata" containerID="cri-o://747c13e8d23d170bcf696dca900cf69ae22620556ef8a97a6ad5111684d57f17" gracePeriod=30 Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.607548 4997 generic.go:334] "Generic (PLEG): container finished" podID="f3cbe3c8-d087-461e-a663-c0e8229640b4" containerID="944cbd4353d191849652e26d5cd7806761de0e71e59a72861dcce5e9f3b7a7ff" exitCode=143 Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.607632 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f3cbe3c8-d087-461e-a663-c0e8229640b4","Type":"ContainerDied","Data":"944cbd4353d191849652e26d5cd7806761de0e71e59a72861dcce5e9f3b7a7ff"} Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.609750 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7a6e7c78-6f77-4d5e-96f8-95f67debcd97","Type":"ContainerDied","Data":"aa55d63292154af2420a002e5b662be1664497a35a635f46e63107192049e0f8"} Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.609761 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.609786 4997 scope.go:117] "RemoveContainer" containerID="a765187e12ab594edbd3c92b89d19f7b86dabdc712477b007ef3882121061020" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.612394 4997 generic.go:334] "Generic (PLEG): container finished" podID="79d909f6-46e3-401d-aafa-3d649b632284" containerID="a78ece514e46e4b92e3bb36eb7f2ecd2ff6c5bd515c06adbe710b954d25f52f6" exitCode=143 Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.612424 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79d909f6-46e3-401d-aafa-3d649b632284","Type":"ContainerDied","Data":"a78ece514e46e4b92e3bb36eb7f2ecd2ff6c5bd515c06adbe710b954d25f52f6"} Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.647469 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.656895 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.682424 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 09:46:15 crc kubenswrapper[4997]: E1205 09:46:15.682944 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6e7c78-6f77-4d5e-96f8-95f67debcd97" containerName="nova-cell1-conductor-conductor" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.682961 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6e7c78-6f77-4d5e-96f8-95f67debcd97" containerName="nova-cell1-conductor-conductor" Dec 05 09:46:15 crc kubenswrapper[4997]: E1205 09:46:15.682977 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6398c5c-8fdf-49dc-8de8-c0bc70637ac8" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.682984 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6398c5c-8fdf-49dc-8de8-c0bc70637ac8" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.683198 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6e7c78-6f77-4d5e-96f8-95f67debcd97" containerName="nova-cell1-conductor-conductor" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.683218 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6398c5c-8fdf-49dc-8de8-c0bc70637ac8" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.683959 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.691394 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.699982 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.774656 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbc803d-3d9a-425a-9ac3-65e7010d30a9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0fbc803d-3d9a-425a-9ac3-65e7010d30a9\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.774834 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds2wq\" (UniqueName: \"kubernetes.io/projected/0fbc803d-3d9a-425a-9ac3-65e7010d30a9-kube-api-access-ds2wq\") pod \"nova-cell1-conductor-0\" (UID: \"0fbc803d-3d9a-425a-9ac3-65e7010d30a9\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.774916 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbc803d-3d9a-425a-9ac3-65e7010d30a9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0fbc803d-3d9a-425a-9ac3-65e7010d30a9\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.786031 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a6e7c78-6f77-4d5e-96f8-95f67debcd97" path="/var/lib/kubelet/pods/7a6e7c78-6f77-4d5e-96f8-95f67debcd97/volumes" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.877170 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds2wq\" (UniqueName: \"kubernetes.io/projected/0fbc803d-3d9a-425a-9ac3-65e7010d30a9-kube-api-access-ds2wq\") pod \"nova-cell1-conductor-0\" (UID: \"0fbc803d-3d9a-425a-9ac3-65e7010d30a9\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.877284 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbc803d-3d9a-425a-9ac3-65e7010d30a9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0fbc803d-3d9a-425a-9ac3-65e7010d30a9\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.878419 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbc803d-3d9a-425a-9ac3-65e7010d30a9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0fbc803d-3d9a-425a-9ac3-65e7010d30a9\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.882288 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbc803d-3d9a-425a-9ac3-65e7010d30a9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0fbc803d-3d9a-425a-9ac3-65e7010d30a9\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.883121 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbc803d-3d9a-425a-9ac3-65e7010d30a9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0fbc803d-3d9a-425a-9ac3-65e7010d30a9\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:46:15 crc kubenswrapper[4997]: I1205 09:46:15.899035 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds2wq\" (UniqueName: \"kubernetes.io/projected/0fbc803d-3d9a-425a-9ac3-65e7010d30a9-kube-api-access-ds2wq\") pod \"nova-cell1-conductor-0\" (UID: \"0fbc803d-3d9a-425a-9ac3-65e7010d30a9\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:46:16 crc kubenswrapper[4997]: I1205 09:46:16.046988 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 09:46:16 crc kubenswrapper[4997]: I1205 09:46:16.506482 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 09:46:16 crc kubenswrapper[4997]: I1205 09:46:16.625848 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0fbc803d-3d9a-425a-9ac3-65e7010d30a9","Type":"ContainerStarted","Data":"e5028457e19a33101241dd90201b433bc8c6f43f6c5f036c5e6c29e455801fb1"} Dec 05 09:46:17 crc kubenswrapper[4997]: I1205 09:46:17.636170 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0fbc803d-3d9a-425a-9ac3-65e7010d30a9","Type":"ContainerStarted","Data":"ae05ec97a67570c36c65c344b3b0b2099d71fe5fb1593515d54e898ea147cc22"} Dec 05 09:46:17 crc kubenswrapper[4997]: I1205 09:46:17.637673 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 05 09:46:17 crc kubenswrapper[4997]: I1205 09:46:17.662112 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.662094837 podStartE2EDuration="2.662094837s" podCreationTimestamp="2025-12-05 09:46:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:46:17.652086355 +0000 UTC m=+10278.180993616" watchObservedRunningTime="2025-12-05 09:46:17.662094837 +0000 UTC m=+10278.191002098" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.220428 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.332682 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfwz\" (UniqueName: \"kubernetes.io/projected/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-kube-api-access-htfwz\") pod \"04d5f9e7-d90b-4887-af25-cbf4e8a16db0\" (UID: \"04d5f9e7-d90b-4887-af25-cbf4e8a16db0\") " Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.332826 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-combined-ca-bundle\") pod \"04d5f9e7-d90b-4887-af25-cbf4e8a16db0\" (UID: \"04d5f9e7-d90b-4887-af25-cbf4e8a16db0\") " Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.332979 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-config-data\") pod \"04d5f9e7-d90b-4887-af25-cbf4e8a16db0\" (UID: \"04d5f9e7-d90b-4887-af25-cbf4e8a16db0\") " Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.334096 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="79d909f6-46e3-401d-aafa-3d649b632284" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.98:8775/\": dial tcp 10.217.1.98:8775: connect: connection refused" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.334115 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="79d909f6-46e3-401d-aafa-3d649b632284" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.98:8775/\": dial tcp 10.217.1.98:8775: connect: connection refused" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.341358 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-kube-api-access-htfwz" (OuterVolumeSpecName: "kube-api-access-htfwz") pod "04d5f9e7-d90b-4887-af25-cbf4e8a16db0" (UID: "04d5f9e7-d90b-4887-af25-cbf4e8a16db0"). InnerVolumeSpecName "kube-api-access-htfwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.371824 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04d5f9e7-d90b-4887-af25-cbf4e8a16db0" (UID: "04d5f9e7-d90b-4887-af25-cbf4e8a16db0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.377285 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-config-data" (OuterVolumeSpecName: "config-data") pod "04d5f9e7-d90b-4887-af25-cbf4e8a16db0" (UID: "04d5f9e7-d90b-4887-af25-cbf4e8a16db0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.435321 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfwz\" (UniqueName: \"kubernetes.io/projected/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-kube-api-access-htfwz\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.435354 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.435366 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d5f9e7-d90b-4887-af25-cbf4e8a16db0-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.658180 4997 generic.go:334] "Generic (PLEG): container finished" podID="04d5f9e7-d90b-4887-af25-cbf4e8a16db0" containerID="28b07c5c22ab9328db79dbaaf5998a6025761c2994ace11bb7e357dd95249f1a" exitCode=0 Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.658241 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"04d5f9e7-d90b-4887-af25-cbf4e8a16db0","Type":"ContainerDied","Data":"28b07c5c22ab9328db79dbaaf5998a6025761c2994ace11bb7e357dd95249f1a"} Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.658267 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"04d5f9e7-d90b-4887-af25-cbf4e8a16db0","Type":"ContainerDied","Data":"8ce17ff382b14b6033ec249aa5b67540f64a3d66ea1c8a38bd6349d39baeb0e6"} Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.658285 4997 scope.go:117] "RemoveContainer" containerID="28b07c5c22ab9328db79dbaaf5998a6025761c2994ace11bb7e357dd95249f1a" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.658460 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.659697 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.689274 4997 generic.go:334] "Generic (PLEG): container finished" podID="f3cbe3c8-d087-461e-a663-c0e8229640b4" containerID="6deab0d56789b3905d671ebbc986574fd95edd3cc69a8e63b485e79625592e1d" exitCode=0 Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.689379 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f3cbe3c8-d087-461e-a663-c0e8229640b4","Type":"ContainerDied","Data":"6deab0d56789b3905d671ebbc986574fd95edd3cc69a8e63b485e79625592e1d"} Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.689405 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f3cbe3c8-d087-461e-a663-c0e8229640b4","Type":"ContainerDied","Data":"be32345d9465c8f0a5383ad4eda6f3ab78331e94ba462968c628d5004f5db8a7"} Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.700579 4997 generic.go:334] "Generic (PLEG): container finished" podID="51c4332f-d8d3-479f-af91-3314d7111adc" containerID="35db5a225a2271f9e2cace9eeead88d9507db1e4658527d66fe51579558496fb" exitCode=0 Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.700736 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"51c4332f-d8d3-479f-af91-3314d7111adc","Type":"ContainerDied","Data":"35db5a225a2271f9e2cace9eeead88d9507db1e4658527d66fe51579558496fb"} Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.714632 4997 generic.go:334] "Generic (PLEG): container finished" podID="79d909f6-46e3-401d-aafa-3d649b632284" containerID="747c13e8d23d170bcf696dca900cf69ae22620556ef8a97a6ad5111684d57f17" exitCode=0 Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.715985 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79d909f6-46e3-401d-aafa-3d649b632284","Type":"ContainerDied","Data":"747c13e8d23d170bcf696dca900cf69ae22620556ef8a97a6ad5111684d57f17"} Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.735151 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.740052 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3cbe3c8-d087-461e-a663-c0e8229640b4-logs\") pod \"f3cbe3c8-d087-461e-a663-c0e8229640b4\" (UID: \"f3cbe3c8-d087-461e-a663-c0e8229640b4\") " Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.740096 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3cbe3c8-d087-461e-a663-c0e8229640b4-config-data\") pod \"f3cbe3c8-d087-461e-a663-c0e8229640b4\" (UID: \"f3cbe3c8-d087-461e-a663-c0e8229640b4\") " Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.740200 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpgsr\" (UniqueName: \"kubernetes.io/projected/f3cbe3c8-d087-461e-a663-c0e8229640b4-kube-api-access-fpgsr\") pod \"f3cbe3c8-d087-461e-a663-c0e8229640b4\" (UID: \"f3cbe3c8-d087-461e-a663-c0e8229640b4\") " Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.740280 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3cbe3c8-d087-461e-a663-c0e8229640b4-combined-ca-bundle\") pod \"f3cbe3c8-d087-461e-a663-c0e8229640b4\" (UID: \"f3cbe3c8-d087-461e-a663-c0e8229640b4\") " Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.741581 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3cbe3c8-d087-461e-a663-c0e8229640b4-logs" (OuterVolumeSpecName: "logs") pod "f3cbe3c8-d087-461e-a663-c0e8229640b4" (UID: "f3cbe3c8-d087-461e-a663-c0e8229640b4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.744068 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3cbe3c8-d087-461e-a663-c0e8229640b4-logs\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.744354 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.756901 4997 scope.go:117] "RemoveContainer" containerID="28b07c5c22ab9328db79dbaaf5998a6025761c2994ace11bb7e357dd95249f1a" Dec 05 09:46:18 crc kubenswrapper[4997]: E1205 09:46:18.759383 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28b07c5c22ab9328db79dbaaf5998a6025761c2994ace11bb7e357dd95249f1a\": container with ID starting with 28b07c5c22ab9328db79dbaaf5998a6025761c2994ace11bb7e357dd95249f1a not found: ID does not exist" containerID="28b07c5c22ab9328db79dbaaf5998a6025761c2994ace11bb7e357dd95249f1a" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.759429 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28b07c5c22ab9328db79dbaaf5998a6025761c2994ace11bb7e357dd95249f1a"} err="failed to get container status \"28b07c5c22ab9328db79dbaaf5998a6025761c2994ace11bb7e357dd95249f1a\": rpc error: code = NotFound desc = could not find container \"28b07c5c22ab9328db79dbaaf5998a6025761c2994ace11bb7e357dd95249f1a\": container with ID starting with 28b07c5c22ab9328db79dbaaf5998a6025761c2994ace11bb7e357dd95249f1a not found: ID does not exist" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.759457 4997 scope.go:117] "RemoveContainer" containerID="6deab0d56789b3905d671ebbc986574fd95edd3cc69a8e63b485e79625592e1d" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.766117 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3cbe3c8-d087-461e-a663-c0e8229640b4-kube-api-access-fpgsr" (OuterVolumeSpecName: "kube-api-access-fpgsr") pod "f3cbe3c8-d087-461e-a663-c0e8229640b4" (UID: "f3cbe3c8-d087-461e-a663-c0e8229640b4"). InnerVolumeSpecName "kube-api-access-fpgsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.771795 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.776423 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3cbe3c8-d087-461e-a663-c0e8229640b4-config-data" (OuterVolumeSpecName: "config-data") pod "f3cbe3c8-d087-461e-a663-c0e8229640b4" (UID: "f3cbe3c8-d087-461e-a663-c0e8229640b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:46:18 crc kubenswrapper[4997]: E1205 09:46:18.776929 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3cbe3c8-d087-461e-a663-c0e8229640b4" containerName="nova-api-api" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.776980 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3cbe3c8-d087-461e-a663-c0e8229640b4" containerName="nova-api-api" Dec 05 09:46:18 crc kubenswrapper[4997]: E1205 09:46:18.777028 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04d5f9e7-d90b-4887-af25-cbf4e8a16db0" containerName="nova-cell0-conductor-conductor" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.777061 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="04d5f9e7-d90b-4887-af25-cbf4e8a16db0" containerName="nova-cell0-conductor-conductor" Dec 05 09:46:18 crc kubenswrapper[4997]: E1205 09:46:18.777084 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3cbe3c8-d087-461e-a663-c0e8229640b4" containerName="nova-api-log" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.777094 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3cbe3c8-d087-461e-a663-c0e8229640b4" containerName="nova-api-log" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.778902 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3cbe3c8-d087-461e-a663-c0e8229640b4" containerName="nova-api-log" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.778927 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3cbe3c8-d087-461e-a663-c0e8229640b4" containerName="nova-api-api" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.778953 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="04d5f9e7-d90b-4887-af25-cbf4e8a16db0" containerName="nova-cell0-conductor-conductor" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.779821 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.785201 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.785195 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.792286 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.806141 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3cbe3c8-d087-461e-a663-c0e8229640b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3cbe3c8-d087-461e-a663-c0e8229640b4" (UID: "f3cbe3c8-d087-461e-a663-c0e8229640b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.806483 4997 scope.go:117] "RemoveContainer" containerID="944cbd4353d191849652e26d5cd7806761de0e71e59a72861dcce5e9f3b7a7ff" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.847324 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d909f6-46e3-401d-aafa-3d649b632284-combined-ca-bundle\") pod \"79d909f6-46e3-401d-aafa-3d649b632284\" (UID: \"79d909f6-46e3-401d-aafa-3d649b632284\") " Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.847466 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scq8m\" (UniqueName: \"kubernetes.io/projected/79d909f6-46e3-401d-aafa-3d649b632284-kube-api-access-scq8m\") pod \"79d909f6-46e3-401d-aafa-3d649b632284\" (UID: \"79d909f6-46e3-401d-aafa-3d649b632284\") " Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.847535 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79d909f6-46e3-401d-aafa-3d649b632284-logs\") pod \"79d909f6-46e3-401d-aafa-3d649b632284\" (UID: \"79d909f6-46e3-401d-aafa-3d649b632284\") " Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.847718 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79d909f6-46e3-401d-aafa-3d649b632284-config-data\") pod \"79d909f6-46e3-401d-aafa-3d649b632284\" (UID: \"79d909f6-46e3-401d-aafa-3d649b632284\") " Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.848035 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3180318-e875-4261-bc09-b10c13e79ca5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b3180318-e875-4261-bc09-b10c13e79ca5\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.848098 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zptx\" (UniqueName: \"kubernetes.io/projected/b3180318-e875-4261-bc09-b10c13e79ca5-kube-api-access-2zptx\") pod \"nova-cell0-conductor-0\" (UID: \"b3180318-e875-4261-bc09-b10c13e79ca5\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.848240 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3180318-e875-4261-bc09-b10c13e79ca5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b3180318-e875-4261-bc09-b10c13e79ca5\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.848684 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79d909f6-46e3-401d-aafa-3d649b632284-logs" (OuterVolumeSpecName: "logs") pod "79d909f6-46e3-401d-aafa-3d649b632284" (UID: "79d909f6-46e3-401d-aafa-3d649b632284"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.849929 4997 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79d909f6-46e3-401d-aafa-3d649b632284-logs\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.849961 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3cbe3c8-d087-461e-a663-c0e8229640b4-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.849972 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpgsr\" (UniqueName: \"kubernetes.io/projected/f3cbe3c8-d087-461e-a663-c0e8229640b4-kube-api-access-fpgsr\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.849981 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3cbe3c8-d087-461e-a663-c0e8229640b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.852095 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79d909f6-46e3-401d-aafa-3d649b632284-kube-api-access-scq8m" (OuterVolumeSpecName: "kube-api-access-scq8m") pod "79d909f6-46e3-401d-aafa-3d649b632284" (UID: "79d909f6-46e3-401d-aafa-3d649b632284"). InnerVolumeSpecName "kube-api-access-scq8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.898715 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79d909f6-46e3-401d-aafa-3d649b632284-config-data" (OuterVolumeSpecName: "config-data") pod "79d909f6-46e3-401d-aafa-3d649b632284" (UID: "79d909f6-46e3-401d-aafa-3d649b632284"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.903025 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79d909f6-46e3-401d-aafa-3d649b632284-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79d909f6-46e3-401d-aafa-3d649b632284" (UID: "79d909f6-46e3-401d-aafa-3d649b632284"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.951698 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zptx\" (UniqueName: \"kubernetes.io/projected/b3180318-e875-4261-bc09-b10c13e79ca5-kube-api-access-2zptx\") pod \"nova-cell0-conductor-0\" (UID: \"b3180318-e875-4261-bc09-b10c13e79ca5\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.951871 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3180318-e875-4261-bc09-b10c13e79ca5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b3180318-e875-4261-bc09-b10c13e79ca5\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.951912 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3180318-e875-4261-bc09-b10c13e79ca5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b3180318-e875-4261-bc09-b10c13e79ca5\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.951968 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scq8m\" (UniqueName: \"kubernetes.io/projected/79d909f6-46e3-401d-aafa-3d649b632284-kube-api-access-scq8m\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.951984 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79d909f6-46e3-401d-aafa-3d649b632284-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.951993 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d909f6-46e3-401d-aafa-3d649b632284-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.956509 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3180318-e875-4261-bc09-b10c13e79ca5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b3180318-e875-4261-bc09-b10c13e79ca5\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.958088 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3180318-e875-4261-bc09-b10c13e79ca5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b3180318-e875-4261-bc09-b10c13e79ca5\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:46:18 crc kubenswrapper[4997]: I1205 09:46:18.968005 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zptx\" (UniqueName: \"kubernetes.io/projected/b3180318-e875-4261-bc09-b10c13e79ca5-kube-api-access-2zptx\") pod \"nova-cell0-conductor-0\" (UID: \"b3180318-e875-4261-bc09-b10c13e79ca5\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.058155 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.061845 4997 scope.go:117] "RemoveContainer" containerID="6deab0d56789b3905d671ebbc986574fd95edd3cc69a8e63b485e79625592e1d" Dec 05 09:46:19 crc kubenswrapper[4997]: E1205 09:46:19.062190 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6deab0d56789b3905d671ebbc986574fd95edd3cc69a8e63b485e79625592e1d\": container with ID starting with 6deab0d56789b3905d671ebbc986574fd95edd3cc69a8e63b485e79625592e1d not found: ID does not exist" containerID="6deab0d56789b3905d671ebbc986574fd95edd3cc69a8e63b485e79625592e1d" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.062222 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6deab0d56789b3905d671ebbc986574fd95edd3cc69a8e63b485e79625592e1d"} err="failed to get container status \"6deab0d56789b3905d671ebbc986574fd95edd3cc69a8e63b485e79625592e1d\": rpc error: code = NotFound desc = could not find container \"6deab0d56789b3905d671ebbc986574fd95edd3cc69a8e63b485e79625592e1d\": container with ID starting with 6deab0d56789b3905d671ebbc986574fd95edd3cc69a8e63b485e79625592e1d not found: ID does not exist" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.062240 4997 scope.go:117] "RemoveContainer" containerID="944cbd4353d191849652e26d5cd7806761de0e71e59a72861dcce5e9f3b7a7ff" Dec 05 09:46:19 crc kubenswrapper[4997]: E1205 09:46:19.062584 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"944cbd4353d191849652e26d5cd7806761de0e71e59a72861dcce5e9f3b7a7ff\": container with ID starting with 944cbd4353d191849652e26d5cd7806761de0e71e59a72861dcce5e9f3b7a7ff not found: ID does not exist" containerID="944cbd4353d191849652e26d5cd7806761de0e71e59a72861dcce5e9f3b7a7ff" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.062621 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"944cbd4353d191849652e26d5cd7806761de0e71e59a72861dcce5e9f3b7a7ff"} err="failed to get container status \"944cbd4353d191849652e26d5cd7806761de0e71e59a72861dcce5e9f3b7a7ff\": rpc error: code = NotFound desc = could not find container \"944cbd4353d191849652e26d5cd7806761de0e71e59a72861dcce5e9f3b7a7ff\": container with ID starting with 944cbd4353d191849652e26d5cd7806761de0e71e59a72861dcce5e9f3b7a7ff not found: ID does not exist" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.062639 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.155311 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51c4332f-d8d3-479f-af91-3314d7111adc-config-data\") pod \"51c4332f-d8d3-479f-af91-3314d7111adc\" (UID: \"51c4332f-d8d3-479f-af91-3314d7111adc\") " Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.155745 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c444\" (UniqueName: \"kubernetes.io/projected/51c4332f-d8d3-479f-af91-3314d7111adc-kube-api-access-9c444\") pod \"51c4332f-d8d3-479f-af91-3314d7111adc\" (UID: \"51c4332f-d8d3-479f-af91-3314d7111adc\") " Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.155814 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c4332f-d8d3-479f-af91-3314d7111adc-combined-ca-bundle\") pod \"51c4332f-d8d3-479f-af91-3314d7111adc\" (UID: \"51c4332f-d8d3-479f-af91-3314d7111adc\") " Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.158895 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51c4332f-d8d3-479f-af91-3314d7111adc-kube-api-access-9c444" (OuterVolumeSpecName: "kube-api-access-9c444") pod "51c4332f-d8d3-479f-af91-3314d7111adc" (UID: "51c4332f-d8d3-479f-af91-3314d7111adc"). InnerVolumeSpecName "kube-api-access-9c444". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.214555 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh"] Dec 05 09:46:19 crc kubenswrapper[4997]: E1205 09:46:19.215092 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79d909f6-46e3-401d-aafa-3d649b632284" containerName="nova-metadata-log" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.215114 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="79d909f6-46e3-401d-aafa-3d649b632284" containerName="nova-metadata-log" Dec 05 09:46:19 crc kubenswrapper[4997]: E1205 09:46:19.215129 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51c4332f-d8d3-479f-af91-3314d7111adc" containerName="nova-scheduler-scheduler" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.215136 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="51c4332f-d8d3-479f-af91-3314d7111adc" containerName="nova-scheduler-scheduler" Dec 05 09:46:19 crc kubenswrapper[4997]: E1205 09:46:19.215163 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79d909f6-46e3-401d-aafa-3d649b632284" containerName="nova-metadata-metadata" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.215168 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="79d909f6-46e3-401d-aafa-3d649b632284" containerName="nova-metadata-metadata" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.215389 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="51c4332f-d8d3-479f-af91-3314d7111adc" containerName="nova-scheduler-scheduler" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.215421 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="79d909f6-46e3-401d-aafa-3d649b632284" containerName="nova-metadata-log" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.215434 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="79d909f6-46e3-401d-aafa-3d649b632284" containerName="nova-metadata-metadata" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.216259 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.218896 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.219415 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.220061 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.220732 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.222873 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.224679 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.225477 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8c22d" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.235518 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh"] Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.238847 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51c4332f-d8d3-479f-af91-3314d7111adc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51c4332f-d8d3-479f-af91-3314d7111adc" (UID: "51c4332f-d8d3-479f-af91-3314d7111adc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.239173 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51c4332f-d8d3-479f-af91-3314d7111adc-config-data" (OuterVolumeSpecName: "config-data") pod "51c4332f-d8d3-479f-af91-3314d7111adc" (UID: "51c4332f-d8d3-479f-af91-3314d7111adc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.258026 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nw26v\" (UniqueName: \"kubernetes.io/projected/329e73d1-df45-481a-b897-1bde9107da2b-kube-api-access-nw26v\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.258138 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.258189 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/329e73d1-df45-481a-b897-1bde9107da2b-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.258209 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.258230 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.258257 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.258286 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.258415 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.258592 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.258632 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/329e73d1-df45-481a-b897-1bde9107da2b-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.258673 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.258901 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51c4332f-d8d3-479f-af91-3314d7111adc-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.258943 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c444\" (UniqueName: \"kubernetes.io/projected/51c4332f-d8d3-479f-af91-3314d7111adc-kube-api-access-9c444\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.258958 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c4332f-d8d3-479f-af91-3314d7111adc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.361359 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.361426 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.361509 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/329e73d1-df45-481a-b897-1bde9107da2b-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.361535 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.361566 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.361645 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nw26v\" (UniqueName: \"kubernetes.io/projected/329e73d1-df45-481a-b897-1bde9107da2b-kube-api-access-nw26v\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.361729 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.361795 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/329e73d1-df45-481a-b897-1bde9107da2b-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.361822 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.361849 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.361884 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.363325 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/329e73d1-df45-481a-b897-1bde9107da2b-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.364904 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/329e73d1-df45-481a-b897-1bde9107da2b-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.365662 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.366490 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.366521 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.366540 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.366973 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.367168 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.368270 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.369788 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.378465 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nw26v\" (UniqueName: \"kubernetes.io/projected/329e73d1-df45-481a-b897-1bde9107da2b-kube-api-access-nw26v\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: W1205 09:46:19.535095 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3180318_e875_4261_bc09_b10c13e79ca5.slice/crio-db916da1fa172731b488b5ac18ce70c70e85718ace0cc80795a77c9a425e35b0 WatchSource:0}: Error finding container db916da1fa172731b488b5ac18ce70c70e85718ace0cc80795a77c9a425e35b0: Status 404 returned error can't find the container with id db916da1fa172731b488b5ac18ce70c70e85718ace0cc80795a77c9a425e35b0 Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.536649 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.541024 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.726864 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b3180318-e875-4261-bc09-b10c13e79ca5","Type":"ContainerStarted","Data":"db916da1fa172731b488b5ac18ce70c70e85718ace0cc80795a77c9a425e35b0"} Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.732019 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.734953 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"51c4332f-d8d3-479f-af91-3314d7111adc","Type":"ContainerDied","Data":"2e3f7ef91a62475de6a1012179c9e72682747be2dfff6a7e66125482e438c969"} Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.735005 4997 scope.go:117] "RemoveContainer" containerID="35db5a225a2271f9e2cace9eeead88d9507db1e4658527d66fe51579558496fb" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.735111 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.744649 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79d909f6-46e3-401d-aafa-3d649b632284","Type":"ContainerDied","Data":"ac239a24fcdb01900d6f89c3f62788c62aa37b6dcf8c49e582522c41aa67db97"} Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.744669 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.770135 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.770212 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.791682 4997 scope.go:117] "RemoveContainer" containerID="747c13e8d23d170bcf696dca900cf69ae22620556ef8a97a6ad5111684d57f17" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.802019 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04d5f9e7-d90b-4887-af25-cbf4e8a16db0" path="/var/lib/kubelet/pods/04d5f9e7-d90b-4887-af25-cbf4e8a16db0/volumes" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.834849 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.849361 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.858422 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.860198 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.867086 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.892744 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.929295 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.966309 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.967215 4997 scope.go:117] "RemoveContainer" containerID="a78ece514e46e4b92e3bb36eb7f2ecd2ff6c5bd515c06adbe710b954d25f52f6" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.993912 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/396b6f95-2e98-4e5d-9a32-a0b6758b09db-config-data\") pod \"nova-scheduler-0\" (UID: \"396b6f95-2e98-4e5d-9a32-a0b6758b09db\") " pod="openstack/nova-scheduler-0" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.994188 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/396b6f95-2e98-4e5d-9a32-a0b6758b09db-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"396b6f95-2e98-4e5d-9a32-a0b6758b09db\") " pod="openstack/nova-scheduler-0" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.994223 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jx9l\" (UniqueName: \"kubernetes.io/projected/396b6f95-2e98-4e5d-9a32-a0b6758b09db-kube-api-access-8jx9l\") pod \"nova-scheduler-0\" (UID: \"396b6f95-2e98-4e5d-9a32-a0b6758b09db\") " pod="openstack/nova-scheduler-0" Dec 05 09:46:19 crc kubenswrapper[4997]: I1205 09:46:19.996120 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.008714 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.018542 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.020590 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.022997 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.032261 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.054236 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.056440 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.058124 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.069153 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.082042 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh"] Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.096437 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5db61aad-f627-47e0-9e6c-4336e1a723b3-logs\") pod \"nova-api-0\" (UID: \"5db61aad-f627-47e0-9e6c-4336e1a723b3\") " pod="openstack/nova-api-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.096508 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5db61aad-f627-47e0-9e6c-4336e1a723b3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5db61aad-f627-47e0-9e6c-4336e1a723b3\") " pod="openstack/nova-api-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.096568 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5db61aad-f627-47e0-9e6c-4336e1a723b3-config-data\") pod \"nova-api-0\" (UID: \"5db61aad-f627-47e0-9e6c-4336e1a723b3\") " pod="openstack/nova-api-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.096596 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/481e9fdd-91aa-4306-bd21-27fc9bbf9608-config-data\") pod \"nova-metadata-0\" (UID: \"481e9fdd-91aa-4306-bd21-27fc9bbf9608\") " pod="openstack/nova-metadata-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.096730 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x72dc\" (UniqueName: \"kubernetes.io/projected/5db61aad-f627-47e0-9e6c-4336e1a723b3-kube-api-access-x72dc\") pod \"nova-api-0\" (UID: \"5db61aad-f627-47e0-9e6c-4336e1a723b3\") " pod="openstack/nova-api-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.096881 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/481e9fdd-91aa-4306-bd21-27fc9bbf9608-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"481e9fdd-91aa-4306-bd21-27fc9bbf9608\") " pod="openstack/nova-metadata-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.096941 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxr5v\" (UniqueName: \"kubernetes.io/projected/481e9fdd-91aa-4306-bd21-27fc9bbf9608-kube-api-access-pxr5v\") pod \"nova-metadata-0\" (UID: \"481e9fdd-91aa-4306-bd21-27fc9bbf9608\") " pod="openstack/nova-metadata-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.097058 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/396b6f95-2e98-4e5d-9a32-a0b6758b09db-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"396b6f95-2e98-4e5d-9a32-a0b6758b09db\") " pod="openstack/nova-scheduler-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.097096 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jx9l\" (UniqueName: \"kubernetes.io/projected/396b6f95-2e98-4e5d-9a32-a0b6758b09db-kube-api-access-8jx9l\") pod \"nova-scheduler-0\" (UID: \"396b6f95-2e98-4e5d-9a32-a0b6758b09db\") " pod="openstack/nova-scheduler-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.098270 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/481e9fdd-91aa-4306-bd21-27fc9bbf9608-logs\") pod \"nova-metadata-0\" (UID: \"481e9fdd-91aa-4306-bd21-27fc9bbf9608\") " pod="openstack/nova-metadata-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.098480 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/396b6f95-2e98-4e5d-9a32-a0b6758b09db-config-data\") pod \"nova-scheduler-0\" (UID: \"396b6f95-2e98-4e5d-9a32-a0b6758b09db\") " pod="openstack/nova-scheduler-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.102793 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/396b6f95-2e98-4e5d-9a32-a0b6758b09db-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"396b6f95-2e98-4e5d-9a32-a0b6758b09db\") " pod="openstack/nova-scheduler-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.102832 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/396b6f95-2e98-4e5d-9a32-a0b6758b09db-config-data\") pod \"nova-scheduler-0\" (UID: \"396b6f95-2e98-4e5d-9a32-a0b6758b09db\") " pod="openstack/nova-scheduler-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.113194 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jx9l\" (UniqueName: \"kubernetes.io/projected/396b6f95-2e98-4e5d-9a32-a0b6758b09db-kube-api-access-8jx9l\") pod \"nova-scheduler-0\" (UID: \"396b6f95-2e98-4e5d-9a32-a0b6758b09db\") " pod="openstack/nova-scheduler-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.201299 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxr5v\" (UniqueName: \"kubernetes.io/projected/481e9fdd-91aa-4306-bd21-27fc9bbf9608-kube-api-access-pxr5v\") pod \"nova-metadata-0\" (UID: \"481e9fdd-91aa-4306-bd21-27fc9bbf9608\") " pod="openstack/nova-metadata-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.201427 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/481e9fdd-91aa-4306-bd21-27fc9bbf9608-logs\") pod \"nova-metadata-0\" (UID: \"481e9fdd-91aa-4306-bd21-27fc9bbf9608\") " pod="openstack/nova-metadata-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.201558 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5db61aad-f627-47e0-9e6c-4336e1a723b3-logs\") pod \"nova-api-0\" (UID: \"5db61aad-f627-47e0-9e6c-4336e1a723b3\") " pod="openstack/nova-api-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.201655 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5db61aad-f627-47e0-9e6c-4336e1a723b3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5db61aad-f627-47e0-9e6c-4336e1a723b3\") " pod="openstack/nova-api-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.201699 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5db61aad-f627-47e0-9e6c-4336e1a723b3-config-data\") pod \"nova-api-0\" (UID: \"5db61aad-f627-47e0-9e6c-4336e1a723b3\") " pod="openstack/nova-api-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.201750 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/481e9fdd-91aa-4306-bd21-27fc9bbf9608-config-data\") pod \"nova-metadata-0\" (UID: \"481e9fdd-91aa-4306-bd21-27fc9bbf9608\") " pod="openstack/nova-metadata-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.201773 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x72dc\" (UniqueName: \"kubernetes.io/projected/5db61aad-f627-47e0-9e6c-4336e1a723b3-kube-api-access-x72dc\") pod \"nova-api-0\" (UID: \"5db61aad-f627-47e0-9e6c-4336e1a723b3\") " pod="openstack/nova-api-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.201832 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/481e9fdd-91aa-4306-bd21-27fc9bbf9608-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"481e9fdd-91aa-4306-bd21-27fc9bbf9608\") " pod="openstack/nova-metadata-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.202124 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/481e9fdd-91aa-4306-bd21-27fc9bbf9608-logs\") pod \"nova-metadata-0\" (UID: \"481e9fdd-91aa-4306-bd21-27fc9bbf9608\") " pod="openstack/nova-metadata-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.205746 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/481e9fdd-91aa-4306-bd21-27fc9bbf9608-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"481e9fdd-91aa-4306-bd21-27fc9bbf9608\") " pod="openstack/nova-metadata-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.206356 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5db61aad-f627-47e0-9e6c-4336e1a723b3-logs\") pod \"nova-api-0\" (UID: \"5db61aad-f627-47e0-9e6c-4336e1a723b3\") " pod="openstack/nova-api-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.207327 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5db61aad-f627-47e0-9e6c-4336e1a723b3-config-data\") pod \"nova-api-0\" (UID: \"5db61aad-f627-47e0-9e6c-4336e1a723b3\") " pod="openstack/nova-api-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.209168 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/481e9fdd-91aa-4306-bd21-27fc9bbf9608-config-data\") pod \"nova-metadata-0\" (UID: \"481e9fdd-91aa-4306-bd21-27fc9bbf9608\") " pod="openstack/nova-metadata-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.209499 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5db61aad-f627-47e0-9e6c-4336e1a723b3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5db61aad-f627-47e0-9e6c-4336e1a723b3\") " pod="openstack/nova-api-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.220218 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxr5v\" (UniqueName: \"kubernetes.io/projected/481e9fdd-91aa-4306-bd21-27fc9bbf9608-kube-api-access-pxr5v\") pod \"nova-metadata-0\" (UID: \"481e9fdd-91aa-4306-bd21-27fc9bbf9608\") " pod="openstack/nova-metadata-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.226733 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x72dc\" (UniqueName: \"kubernetes.io/projected/5db61aad-f627-47e0-9e6c-4336e1a723b3-kube-api-access-x72dc\") pod \"nova-api-0\" (UID: \"5db61aad-f627-47e0-9e6c-4336e1a723b3\") " pod="openstack/nova-api-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.313913 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.339610 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.376845 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.764274 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" event={"ID":"329e73d1-df45-481a-b897-1bde9107da2b","Type":"ContainerStarted","Data":"a987673d5a80707f65e9900b392f91ee9b8a0d898d012d06df0c21d9adfcc35a"} Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.764586 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" event={"ID":"329e73d1-df45-481a-b897-1bde9107da2b","Type":"ContainerStarted","Data":"97d60535b245a99001202f575fda928b291991751947062c9daeb0a1880b4720"} Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.765820 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b3180318-e875-4261-bc09-b10c13e79ca5","Type":"ContainerStarted","Data":"88597d7e7d3a7daebe9408a7291fd1d3906fda17ba68fcf9d04da1a69375188e"} Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.765998 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.793957 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" podStartSLOduration=1.313195365 podStartE2EDuration="1.793938187s" podCreationTimestamp="2025-12-05 09:46:19 +0000 UTC" firstStartedPulling="2025-12-05 09:46:19.985903882 +0000 UTC m=+10280.514811133" lastFinishedPulling="2025-12-05 09:46:20.466646694 +0000 UTC m=+10280.995553955" observedRunningTime="2025-12-05 09:46:20.789785275 +0000 UTC m=+10281.318692556" watchObservedRunningTime="2025-12-05 09:46:20.793938187 +0000 UTC m=+10281.322845448" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.823099 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.825446 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.825430921 podStartE2EDuration="2.825430921s" podCreationTimestamp="2025-12-05 09:46:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:46:20.809239532 +0000 UTC m=+10281.338146793" watchObservedRunningTime="2025-12-05 09:46:20.825430921 +0000 UTC m=+10281.354338182" Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.867480 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:46:20 crc kubenswrapper[4997]: W1205 09:46:20.869306 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod481e9fdd_91aa_4306_bd21_27fc9bbf9608.slice/crio-a3c15e48fc2101e6dd2641d52b6410abf8e690d492e870481712b165e6ecb50e WatchSource:0}: Error finding container a3c15e48fc2101e6dd2641d52b6410abf8e690d492e870481712b165e6ecb50e: Status 404 returned error can't find the container with id a3c15e48fc2101e6dd2641d52b6410abf8e690d492e870481712b165e6ecb50e Dec 05 09:46:20 crc kubenswrapper[4997]: I1205 09:46:20.948596 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:46:20 crc kubenswrapper[4997]: W1205 09:46:20.959140 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5db61aad_f627_47e0_9e6c_4336e1a723b3.slice/crio-1d8d1d75d68e02cc1522be41a0df0ebdfa30690b5918e45f07db5cb30e437a67 WatchSource:0}: Error finding container 1d8d1d75d68e02cc1522be41a0df0ebdfa30690b5918e45f07db5cb30e437a67: Status 404 returned error can't find the container with id 1d8d1d75d68e02cc1522be41a0df0ebdfa30690b5918e45f07db5cb30e437a67 Dec 05 09:46:21 crc kubenswrapper[4997]: I1205 09:46:21.111852 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 05 09:46:21 crc kubenswrapper[4997]: I1205 09:46:21.762046 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51c4332f-d8d3-479f-af91-3314d7111adc" path="/var/lib/kubelet/pods/51c4332f-d8d3-479f-af91-3314d7111adc/volumes" Dec 05 09:46:21 crc kubenswrapper[4997]: I1205 09:46:21.763066 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79d909f6-46e3-401d-aafa-3d649b632284" path="/var/lib/kubelet/pods/79d909f6-46e3-401d-aafa-3d649b632284/volumes" Dec 05 09:46:21 crc kubenswrapper[4997]: I1205 09:46:21.763805 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3cbe3c8-d087-461e-a663-c0e8229640b4" path="/var/lib/kubelet/pods/f3cbe3c8-d087-461e-a663-c0e8229640b4/volumes" Dec 05 09:46:21 crc kubenswrapper[4997]: I1205 09:46:21.777854 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5db61aad-f627-47e0-9e6c-4336e1a723b3","Type":"ContainerStarted","Data":"3cc980b31ed8d20beafb380d8606fb62270fdb8aca843549de0885ba728db509"} Dec 05 09:46:21 crc kubenswrapper[4997]: I1205 09:46:21.777908 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5db61aad-f627-47e0-9e6c-4336e1a723b3","Type":"ContainerStarted","Data":"0b3c92e08d7cc615efd802e318f1d919f0766750fc701daf637381e1249b0850"} Dec 05 09:46:21 crc kubenswrapper[4997]: I1205 09:46:21.777949 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5db61aad-f627-47e0-9e6c-4336e1a723b3","Type":"ContainerStarted","Data":"1d8d1d75d68e02cc1522be41a0df0ebdfa30690b5918e45f07db5cb30e437a67"} Dec 05 09:46:21 crc kubenswrapper[4997]: I1205 09:46:21.781486 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"481e9fdd-91aa-4306-bd21-27fc9bbf9608","Type":"ContainerStarted","Data":"24151d4264ec7a89685411b856dbdc77290162d07bce92b99346e5ea27db684c"} Dec 05 09:46:21 crc kubenswrapper[4997]: I1205 09:46:21.781543 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"481e9fdd-91aa-4306-bd21-27fc9bbf9608","Type":"ContainerStarted","Data":"ceb6a25f556cec2f2fc8b07939f7076a6ecfb2431bca2c112528b904c3604001"} Dec 05 09:46:21 crc kubenswrapper[4997]: I1205 09:46:21.781559 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"481e9fdd-91aa-4306-bd21-27fc9bbf9608","Type":"ContainerStarted","Data":"a3c15e48fc2101e6dd2641d52b6410abf8e690d492e870481712b165e6ecb50e"} Dec 05 09:46:21 crc kubenswrapper[4997]: I1205 09:46:21.783798 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"396b6f95-2e98-4e5d-9a32-a0b6758b09db","Type":"ContainerStarted","Data":"c69727686af620988033c72d950e0a3ac93e78cdc2142bc9fee65c73ae5a3b8d"} Dec 05 09:46:21 crc kubenswrapper[4997]: I1205 09:46:21.783838 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"396b6f95-2e98-4e5d-9a32-a0b6758b09db","Type":"ContainerStarted","Data":"9aafa7ea512268273591deffaac0a497b18b8d97c9bebb11aad026d6f736f222"} Dec 05 09:46:21 crc kubenswrapper[4997]: I1205 09:46:21.811985 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.8119607540000002 podStartE2EDuration="2.811960754s" podCreationTimestamp="2025-12-05 09:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:46:21.800384001 +0000 UTC m=+10282.329291272" watchObservedRunningTime="2025-12-05 09:46:21.811960754 +0000 UTC m=+10282.340868015" Dec 05 09:46:21 crc kubenswrapper[4997]: I1205 09:46:21.880447 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.88041457 podStartE2EDuration="2.88041457s" podCreationTimestamp="2025-12-05 09:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:46:21.848045793 +0000 UTC m=+10282.376953064" watchObservedRunningTime="2025-12-05 09:46:21.88041457 +0000 UTC m=+10282.409321831" Dec 05 09:46:21 crc kubenswrapper[4997]: I1205 09:46:21.893884 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.893854835 podStartE2EDuration="2.893854835s" podCreationTimestamp="2025-12-05 09:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:46:21.870978054 +0000 UTC m=+10282.399885345" watchObservedRunningTime="2025-12-05 09:46:21.893854835 +0000 UTC m=+10282.422762096" Dec 05 09:46:24 crc kubenswrapper[4997]: I1205 09:46:24.096459 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 05 09:46:25 crc kubenswrapper[4997]: I1205 09:46:25.314914 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 09:46:25 crc kubenswrapper[4997]: I1205 09:46:25.340552 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 09:46:25 crc kubenswrapper[4997]: I1205 09:46:25.341062 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 09:46:30 crc kubenswrapper[4997]: I1205 09:46:30.315285 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 09:46:30 crc kubenswrapper[4997]: I1205 09:46:30.340836 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 09:46:30 crc kubenswrapper[4997]: I1205 09:46:30.340898 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 09:46:30 crc kubenswrapper[4997]: I1205 09:46:30.342936 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 09:46:30 crc kubenswrapper[4997]: I1205 09:46:30.384962 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 09:46:30 crc kubenswrapper[4997]: I1205 09:46:30.385046 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 09:46:30 crc kubenswrapper[4997]: I1205 09:46:30.910487 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 09:46:31 crc kubenswrapper[4997]: I1205 09:46:31.423893 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="481e9fdd-91aa-4306-bd21-27fc9bbf9608" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.205:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:46:31 crc kubenswrapper[4997]: I1205 09:46:31.423921 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="481e9fdd-91aa-4306-bd21-27fc9bbf9608" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.205:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:46:31 crc kubenswrapper[4997]: I1205 09:46:31.506814 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5db61aad-f627-47e0-9e6c-4336e1a723b3" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.206:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:46:31 crc kubenswrapper[4997]: I1205 09:46:31.506832 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5db61aad-f627-47e0-9e6c-4336e1a723b3" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.206:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:46:40 crc kubenswrapper[4997]: I1205 09:46:40.343009 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 09:46:40 crc kubenswrapper[4997]: I1205 09:46:40.343882 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 09:46:40 crc kubenswrapper[4997]: I1205 09:46:40.345694 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 09:46:40 crc kubenswrapper[4997]: I1205 09:46:40.381480 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 09:46:40 crc kubenswrapper[4997]: I1205 09:46:40.381598 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 09:46:40 crc kubenswrapper[4997]: I1205 09:46:40.382122 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 09:46:40 crc kubenswrapper[4997]: I1205 09:46:40.382242 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 09:46:40 crc kubenswrapper[4997]: I1205 09:46:40.385707 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 09:46:40 crc kubenswrapper[4997]: I1205 09:46:40.385767 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 09:46:40 crc kubenswrapper[4997]: I1205 09:46:40.986574 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 09:46:49 crc kubenswrapper[4997]: I1205 09:46:49.769989 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:46:49 crc kubenswrapper[4997]: I1205 09:46:49.770560 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:46:49 crc kubenswrapper[4997]: I1205 09:46:49.770648 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 09:46:49 crc kubenswrapper[4997]: I1205 09:46:49.771795 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:46:49 crc kubenswrapper[4997]: I1205 09:46:49.771902 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" gracePeriod=600 Dec 05 09:46:49 crc kubenswrapper[4997]: E1205 09:46:49.891594 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:46:50 crc kubenswrapper[4997]: I1205 09:46:50.093911 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" exitCode=0 Dec 05 09:46:50 crc kubenswrapper[4997]: I1205 09:46:50.093965 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86"} Dec 05 09:46:50 crc kubenswrapper[4997]: I1205 09:46:50.094013 4997 scope.go:117] "RemoveContainer" containerID="553255d50247709601e534c39b4889a350b0020aa3393fb72def52e3ad4252bf" Dec 05 09:46:50 crc kubenswrapper[4997]: I1205 09:46:50.094729 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:46:50 crc kubenswrapper[4997]: E1205 09:46:50.095161 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:47:00 crc kubenswrapper[4997]: I1205 09:47:00.749808 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:47:00 crc kubenswrapper[4997]: E1205 09:47:00.750806 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:47:14 crc kubenswrapper[4997]: I1205 09:47:14.749706 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:47:14 crc kubenswrapper[4997]: E1205 09:47:14.750708 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:47:29 crc kubenswrapper[4997]: I1205 09:47:29.757834 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:47:29 crc kubenswrapper[4997]: E1205 09:47:29.759370 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:47:42 crc kubenswrapper[4997]: I1205 09:47:42.749308 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:47:42 crc kubenswrapper[4997]: E1205 09:47:42.750121 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:47:56 crc kubenswrapper[4997]: I1205 09:47:56.749118 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:47:56 crc kubenswrapper[4997]: E1205 09:47:56.751003 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:48:11 crc kubenswrapper[4997]: I1205 09:48:11.749927 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:48:11 crc kubenswrapper[4997]: E1205 09:48:11.750709 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:48:18 crc kubenswrapper[4997]: I1205 09:48:18.811564 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8phqp"] Dec 05 09:48:18 crc kubenswrapper[4997]: I1205 09:48:18.815630 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8phqp" Dec 05 09:48:18 crc kubenswrapper[4997]: I1205 09:48:18.825885 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8phqp"] Dec 05 09:48:18 crc kubenswrapper[4997]: I1205 09:48:18.905683 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rz5k\" (UniqueName: \"kubernetes.io/projected/d8f232c7-354a-4226-b5a0-637bf10d598f-kube-api-access-4rz5k\") pod \"certified-operators-8phqp\" (UID: \"d8f232c7-354a-4226-b5a0-637bf10d598f\") " pod="openshift-marketplace/certified-operators-8phqp" Dec 05 09:48:18 crc kubenswrapper[4997]: I1205 09:48:18.905751 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8f232c7-354a-4226-b5a0-637bf10d598f-catalog-content\") pod \"certified-operators-8phqp\" (UID: \"d8f232c7-354a-4226-b5a0-637bf10d598f\") " pod="openshift-marketplace/certified-operators-8phqp" Dec 05 09:48:18 crc kubenswrapper[4997]: I1205 09:48:18.906011 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8f232c7-354a-4226-b5a0-637bf10d598f-utilities\") pod \"certified-operators-8phqp\" (UID: \"d8f232c7-354a-4226-b5a0-637bf10d598f\") " pod="openshift-marketplace/certified-operators-8phqp" Dec 05 09:48:19 crc kubenswrapper[4997]: I1205 09:48:19.011093 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rz5k\" (UniqueName: \"kubernetes.io/projected/d8f232c7-354a-4226-b5a0-637bf10d598f-kube-api-access-4rz5k\") pod \"certified-operators-8phqp\" (UID: \"d8f232c7-354a-4226-b5a0-637bf10d598f\") " pod="openshift-marketplace/certified-operators-8phqp" Dec 05 09:48:19 crc kubenswrapper[4997]: I1205 09:48:19.011174 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8f232c7-354a-4226-b5a0-637bf10d598f-catalog-content\") pod \"certified-operators-8phqp\" (UID: \"d8f232c7-354a-4226-b5a0-637bf10d598f\") " pod="openshift-marketplace/certified-operators-8phqp" Dec 05 09:48:19 crc kubenswrapper[4997]: I1205 09:48:19.011232 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8f232c7-354a-4226-b5a0-637bf10d598f-utilities\") pod \"certified-operators-8phqp\" (UID: \"d8f232c7-354a-4226-b5a0-637bf10d598f\") " pod="openshift-marketplace/certified-operators-8phqp" Dec 05 09:48:19 crc kubenswrapper[4997]: I1205 09:48:19.011729 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8f232c7-354a-4226-b5a0-637bf10d598f-utilities\") pod \"certified-operators-8phqp\" (UID: \"d8f232c7-354a-4226-b5a0-637bf10d598f\") " pod="openshift-marketplace/certified-operators-8phqp" Dec 05 09:48:19 crc kubenswrapper[4997]: I1205 09:48:19.012029 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8f232c7-354a-4226-b5a0-637bf10d598f-catalog-content\") pod \"certified-operators-8phqp\" (UID: \"d8f232c7-354a-4226-b5a0-637bf10d598f\") " pod="openshift-marketplace/certified-operators-8phqp" Dec 05 09:48:19 crc kubenswrapper[4997]: I1205 09:48:19.037373 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rz5k\" (UniqueName: \"kubernetes.io/projected/d8f232c7-354a-4226-b5a0-637bf10d598f-kube-api-access-4rz5k\") pod \"certified-operators-8phqp\" (UID: \"d8f232c7-354a-4226-b5a0-637bf10d598f\") " pod="openshift-marketplace/certified-operators-8phqp" Dec 05 09:48:19 crc kubenswrapper[4997]: I1205 09:48:19.150339 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8phqp" Dec 05 09:48:19 crc kubenswrapper[4997]: I1205 09:48:19.687855 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8phqp"] Dec 05 09:48:20 crc kubenswrapper[4997]: I1205 09:48:20.040215 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phqp" event={"ID":"d8f232c7-354a-4226-b5a0-637bf10d598f","Type":"ContainerStarted","Data":"8059e2567a81f28d5512c8a7f0072ae757d0a6e93423eea104eef278cf1809ce"} Dec 05 09:48:20 crc kubenswrapper[4997]: I1205 09:48:20.695948 4997 scope.go:117] "RemoveContainer" containerID="808b3c643dd1e5a337b4799df129a7e0c3d3a8aa43ecbdf7c556128a32163cb1" Dec 05 09:48:20 crc kubenswrapper[4997]: I1205 09:48:20.719103 4997 scope.go:117] "RemoveContainer" containerID="4ab675055e264ea009da7e4f9ab3bd538bf5402b34fa3253d88a3d926fa30608" Dec 05 09:48:20 crc kubenswrapper[4997]: I1205 09:48:20.754655 4997 scope.go:117] "RemoveContainer" containerID="e170f121b95b0294dab155b281cb806290b9c0e433fe3df0ad180c29627e7d0d" Dec 05 09:48:21 crc kubenswrapper[4997]: I1205 09:48:21.053778 4997 generic.go:334] "Generic (PLEG): container finished" podID="d8f232c7-354a-4226-b5a0-637bf10d598f" containerID="5f3249eb5688c7e268be9a184873841444e24811b43a36c890e8b0cfc12f9ccd" exitCode=0 Dec 05 09:48:21 crc kubenswrapper[4997]: I1205 09:48:21.053841 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phqp" event={"ID":"d8f232c7-354a-4226-b5a0-637bf10d598f","Type":"ContainerDied","Data":"5f3249eb5688c7e268be9a184873841444e24811b43a36c890e8b0cfc12f9ccd"} Dec 05 09:48:22 crc kubenswrapper[4997]: I1205 09:48:22.099993 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phqp" event={"ID":"d8f232c7-354a-4226-b5a0-637bf10d598f","Type":"ContainerStarted","Data":"cbe202cc4e8357aa24db5c1e961772386fd1e26ed6d3f2ad32f917eb65316bef"} Dec 05 09:48:23 crc kubenswrapper[4997]: I1205 09:48:23.110430 4997 generic.go:334] "Generic (PLEG): container finished" podID="d8f232c7-354a-4226-b5a0-637bf10d598f" containerID="cbe202cc4e8357aa24db5c1e961772386fd1e26ed6d3f2ad32f917eb65316bef" exitCode=0 Dec 05 09:48:23 crc kubenswrapper[4997]: I1205 09:48:23.110492 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phqp" event={"ID":"d8f232c7-354a-4226-b5a0-637bf10d598f","Type":"ContainerDied","Data":"cbe202cc4e8357aa24db5c1e961772386fd1e26ed6d3f2ad32f917eb65316bef"} Dec 05 09:48:24 crc kubenswrapper[4997]: I1205 09:48:24.124662 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phqp" event={"ID":"d8f232c7-354a-4226-b5a0-637bf10d598f","Type":"ContainerStarted","Data":"34d07acf9ac31326ff3d397c31be6ea5a2f1f2ba430f7ecb6ea7e3f21edd7348"} Dec 05 09:48:24 crc kubenswrapper[4997]: I1205 09:48:24.156480 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8phqp" podStartSLOduration=3.707831894 podStartE2EDuration="6.156449253s" podCreationTimestamp="2025-12-05 09:48:18 +0000 UTC" firstStartedPulling="2025-12-05 09:48:21.056278231 +0000 UTC m=+10401.585185522" lastFinishedPulling="2025-12-05 09:48:23.50489561 +0000 UTC m=+10404.033802881" observedRunningTime="2025-12-05 09:48:24.150582414 +0000 UTC m=+10404.679489685" watchObservedRunningTime="2025-12-05 09:48:24.156449253 +0000 UTC m=+10404.685356514" Dec 05 09:48:25 crc kubenswrapper[4997]: I1205 09:48:25.749480 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:48:25 crc kubenswrapper[4997]: E1205 09:48:25.749987 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:48:29 crc kubenswrapper[4997]: I1205 09:48:29.152013 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8phqp" Dec 05 09:48:29 crc kubenswrapper[4997]: I1205 09:48:29.152660 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8phqp" Dec 05 09:48:29 crc kubenswrapper[4997]: I1205 09:48:29.213861 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8phqp" Dec 05 09:48:29 crc kubenswrapper[4997]: I1205 09:48:29.264379 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8phqp" Dec 05 09:48:29 crc kubenswrapper[4997]: I1205 09:48:29.451523 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8phqp"] Dec 05 09:48:31 crc kubenswrapper[4997]: I1205 09:48:31.196578 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8phqp" podUID="d8f232c7-354a-4226-b5a0-637bf10d598f" containerName="registry-server" containerID="cri-o://34d07acf9ac31326ff3d397c31be6ea5a2f1f2ba430f7ecb6ea7e3f21edd7348" gracePeriod=2 Dec 05 09:48:31 crc kubenswrapper[4997]: I1205 09:48:31.697815 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8phqp" Dec 05 09:48:31 crc kubenswrapper[4997]: I1205 09:48:31.815144 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8f232c7-354a-4226-b5a0-637bf10d598f-catalog-content\") pod \"d8f232c7-354a-4226-b5a0-637bf10d598f\" (UID: \"d8f232c7-354a-4226-b5a0-637bf10d598f\") " Dec 05 09:48:31 crc kubenswrapper[4997]: I1205 09:48:31.815282 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8f232c7-354a-4226-b5a0-637bf10d598f-utilities\") pod \"d8f232c7-354a-4226-b5a0-637bf10d598f\" (UID: \"d8f232c7-354a-4226-b5a0-637bf10d598f\") " Dec 05 09:48:31 crc kubenswrapper[4997]: I1205 09:48:31.815356 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rz5k\" (UniqueName: \"kubernetes.io/projected/d8f232c7-354a-4226-b5a0-637bf10d598f-kube-api-access-4rz5k\") pod \"d8f232c7-354a-4226-b5a0-637bf10d598f\" (UID: \"d8f232c7-354a-4226-b5a0-637bf10d598f\") " Dec 05 09:48:31 crc kubenswrapper[4997]: I1205 09:48:31.817020 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8f232c7-354a-4226-b5a0-637bf10d598f-utilities" (OuterVolumeSpecName: "utilities") pod "d8f232c7-354a-4226-b5a0-637bf10d598f" (UID: "d8f232c7-354a-4226-b5a0-637bf10d598f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:48:31 crc kubenswrapper[4997]: I1205 09:48:31.821864 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8f232c7-354a-4226-b5a0-637bf10d598f-kube-api-access-4rz5k" (OuterVolumeSpecName: "kube-api-access-4rz5k") pod "d8f232c7-354a-4226-b5a0-637bf10d598f" (UID: "d8f232c7-354a-4226-b5a0-637bf10d598f"). InnerVolumeSpecName "kube-api-access-4rz5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:48:31 crc kubenswrapper[4997]: I1205 09:48:31.860746 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8f232c7-354a-4226-b5a0-637bf10d598f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d8f232c7-354a-4226-b5a0-637bf10d598f" (UID: "d8f232c7-354a-4226-b5a0-637bf10d598f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:48:31 crc kubenswrapper[4997]: I1205 09:48:31.918488 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8f232c7-354a-4226-b5a0-637bf10d598f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:48:31 crc kubenswrapper[4997]: I1205 09:48:31.919213 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8f232c7-354a-4226-b5a0-637bf10d598f-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:48:31 crc kubenswrapper[4997]: I1205 09:48:31.919223 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rz5k\" (UniqueName: \"kubernetes.io/projected/d8f232c7-354a-4226-b5a0-637bf10d598f-kube-api-access-4rz5k\") on node \"crc\" DevicePath \"\"" Dec 05 09:48:32 crc kubenswrapper[4997]: I1205 09:48:32.209508 4997 generic.go:334] "Generic (PLEG): container finished" podID="d8f232c7-354a-4226-b5a0-637bf10d598f" containerID="34d07acf9ac31326ff3d397c31be6ea5a2f1f2ba430f7ecb6ea7e3f21edd7348" exitCode=0 Dec 05 09:48:32 crc kubenswrapper[4997]: I1205 09:48:32.209549 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phqp" event={"ID":"d8f232c7-354a-4226-b5a0-637bf10d598f","Type":"ContainerDied","Data":"34d07acf9ac31326ff3d397c31be6ea5a2f1f2ba430f7ecb6ea7e3f21edd7348"} Dec 05 09:48:32 crc kubenswrapper[4997]: I1205 09:48:32.209574 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phqp" event={"ID":"d8f232c7-354a-4226-b5a0-637bf10d598f","Type":"ContainerDied","Data":"8059e2567a81f28d5512c8a7f0072ae757d0a6e93423eea104eef278cf1809ce"} Dec 05 09:48:32 crc kubenswrapper[4997]: I1205 09:48:32.209592 4997 scope.go:117] "RemoveContainer" containerID="34d07acf9ac31326ff3d397c31be6ea5a2f1f2ba430f7ecb6ea7e3f21edd7348" Dec 05 09:48:32 crc kubenswrapper[4997]: I1205 09:48:32.210646 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8phqp" Dec 05 09:48:32 crc kubenswrapper[4997]: I1205 09:48:32.238907 4997 scope.go:117] "RemoveContainer" containerID="cbe202cc4e8357aa24db5c1e961772386fd1e26ed6d3f2ad32f917eb65316bef" Dec 05 09:48:32 crc kubenswrapper[4997]: I1205 09:48:32.250782 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8phqp"] Dec 05 09:48:32 crc kubenswrapper[4997]: I1205 09:48:32.264779 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8phqp"] Dec 05 09:48:32 crc kubenswrapper[4997]: I1205 09:48:32.281137 4997 scope.go:117] "RemoveContainer" containerID="5f3249eb5688c7e268be9a184873841444e24811b43a36c890e8b0cfc12f9ccd" Dec 05 09:48:32 crc kubenswrapper[4997]: I1205 09:48:32.316668 4997 scope.go:117] "RemoveContainer" containerID="34d07acf9ac31326ff3d397c31be6ea5a2f1f2ba430f7ecb6ea7e3f21edd7348" Dec 05 09:48:32 crc kubenswrapper[4997]: E1205 09:48:32.317197 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34d07acf9ac31326ff3d397c31be6ea5a2f1f2ba430f7ecb6ea7e3f21edd7348\": container with ID starting with 34d07acf9ac31326ff3d397c31be6ea5a2f1f2ba430f7ecb6ea7e3f21edd7348 not found: ID does not exist" containerID="34d07acf9ac31326ff3d397c31be6ea5a2f1f2ba430f7ecb6ea7e3f21edd7348" Dec 05 09:48:32 crc kubenswrapper[4997]: I1205 09:48:32.317244 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34d07acf9ac31326ff3d397c31be6ea5a2f1f2ba430f7ecb6ea7e3f21edd7348"} err="failed to get container status \"34d07acf9ac31326ff3d397c31be6ea5a2f1f2ba430f7ecb6ea7e3f21edd7348\": rpc error: code = NotFound desc = could not find container \"34d07acf9ac31326ff3d397c31be6ea5a2f1f2ba430f7ecb6ea7e3f21edd7348\": container with ID starting with 34d07acf9ac31326ff3d397c31be6ea5a2f1f2ba430f7ecb6ea7e3f21edd7348 not found: ID does not exist" Dec 05 09:48:32 crc kubenswrapper[4997]: I1205 09:48:32.317272 4997 scope.go:117] "RemoveContainer" containerID="cbe202cc4e8357aa24db5c1e961772386fd1e26ed6d3f2ad32f917eb65316bef" Dec 05 09:48:32 crc kubenswrapper[4997]: E1205 09:48:32.317609 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbe202cc4e8357aa24db5c1e961772386fd1e26ed6d3f2ad32f917eb65316bef\": container with ID starting with cbe202cc4e8357aa24db5c1e961772386fd1e26ed6d3f2ad32f917eb65316bef not found: ID does not exist" containerID="cbe202cc4e8357aa24db5c1e961772386fd1e26ed6d3f2ad32f917eb65316bef" Dec 05 09:48:32 crc kubenswrapper[4997]: I1205 09:48:32.317667 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbe202cc4e8357aa24db5c1e961772386fd1e26ed6d3f2ad32f917eb65316bef"} err="failed to get container status \"cbe202cc4e8357aa24db5c1e961772386fd1e26ed6d3f2ad32f917eb65316bef\": rpc error: code = NotFound desc = could not find container \"cbe202cc4e8357aa24db5c1e961772386fd1e26ed6d3f2ad32f917eb65316bef\": container with ID starting with cbe202cc4e8357aa24db5c1e961772386fd1e26ed6d3f2ad32f917eb65316bef not found: ID does not exist" Dec 05 09:48:32 crc kubenswrapper[4997]: I1205 09:48:32.317681 4997 scope.go:117] "RemoveContainer" containerID="5f3249eb5688c7e268be9a184873841444e24811b43a36c890e8b0cfc12f9ccd" Dec 05 09:48:32 crc kubenswrapper[4997]: E1205 09:48:32.318004 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f3249eb5688c7e268be9a184873841444e24811b43a36c890e8b0cfc12f9ccd\": container with ID starting with 5f3249eb5688c7e268be9a184873841444e24811b43a36c890e8b0cfc12f9ccd not found: ID does not exist" containerID="5f3249eb5688c7e268be9a184873841444e24811b43a36c890e8b0cfc12f9ccd" Dec 05 09:48:32 crc kubenswrapper[4997]: I1205 09:48:32.318047 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f3249eb5688c7e268be9a184873841444e24811b43a36c890e8b0cfc12f9ccd"} err="failed to get container status \"5f3249eb5688c7e268be9a184873841444e24811b43a36c890e8b0cfc12f9ccd\": rpc error: code = NotFound desc = could not find container \"5f3249eb5688c7e268be9a184873841444e24811b43a36c890e8b0cfc12f9ccd\": container with ID starting with 5f3249eb5688c7e268be9a184873841444e24811b43a36c890e8b0cfc12f9ccd not found: ID does not exist" Dec 05 09:48:33 crc kubenswrapper[4997]: I1205 09:48:33.763945 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8f232c7-354a-4226-b5a0-637bf10d598f" path="/var/lib/kubelet/pods/d8f232c7-354a-4226-b5a0-637bf10d598f/volumes" Dec 05 09:48:37 crc kubenswrapper[4997]: I1205 09:48:37.749346 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:48:37 crc kubenswrapper[4997]: E1205 09:48:37.750262 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:48:52 crc kubenswrapper[4997]: I1205 09:48:52.750259 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:48:52 crc kubenswrapper[4997]: E1205 09:48:52.751059 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:49:03 crc kubenswrapper[4997]: I1205 09:49:03.536477 4997 generic.go:334] "Generic (PLEG): container finished" podID="329e73d1-df45-481a-b897-1bde9107da2b" containerID="a987673d5a80707f65e9900b392f91ee9b8a0d898d012d06df0c21d9adfcc35a" exitCode=0 Dec 05 09:49:03 crc kubenswrapper[4997]: I1205 09:49:03.536561 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" event={"ID":"329e73d1-df45-481a-b897-1bde9107da2b","Type":"ContainerDied","Data":"a987673d5a80707f65e9900b392f91ee9b8a0d898d012d06df0c21d9adfcc35a"} Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.000319 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.128112 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-migration-ssh-key-0\") pod \"329e73d1-df45-481a-b897-1bde9107da2b\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.128256 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/329e73d1-df45-481a-b897-1bde9107da2b-nova-cells-global-config-1\") pod \"329e73d1-df45-481a-b897-1bde9107da2b\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.128382 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-migration-ssh-key-1\") pod \"329e73d1-df45-481a-b897-1bde9107da2b\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.128521 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/329e73d1-df45-481a-b897-1bde9107da2b-nova-cells-global-config-0\") pod \"329e73d1-df45-481a-b897-1bde9107da2b\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.128746 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nw26v\" (UniqueName: \"kubernetes.io/projected/329e73d1-df45-481a-b897-1bde9107da2b-kube-api-access-nw26v\") pod \"329e73d1-df45-481a-b897-1bde9107da2b\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.129733 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-compute-config-1\") pod \"329e73d1-df45-481a-b897-1bde9107da2b\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.129769 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-combined-ca-bundle\") pod \"329e73d1-df45-481a-b897-1bde9107da2b\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.129862 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-ssh-key\") pod \"329e73d1-df45-481a-b897-1bde9107da2b\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.129972 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-ceph\") pod \"329e73d1-df45-481a-b897-1bde9107da2b\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.129998 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-compute-config-0\") pod \"329e73d1-df45-481a-b897-1bde9107da2b\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.130092 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-inventory\") pod \"329e73d1-df45-481a-b897-1bde9107da2b\" (UID: \"329e73d1-df45-481a-b897-1bde9107da2b\") " Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.136472 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-ceph" (OuterVolumeSpecName: "ceph") pod "329e73d1-df45-481a-b897-1bde9107da2b" (UID: "329e73d1-df45-481a-b897-1bde9107da2b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.136560 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "329e73d1-df45-481a-b897-1bde9107da2b" (UID: "329e73d1-df45-481a-b897-1bde9107da2b"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.136884 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/329e73d1-df45-481a-b897-1bde9107da2b-kube-api-access-nw26v" (OuterVolumeSpecName: "kube-api-access-nw26v") pod "329e73d1-df45-481a-b897-1bde9107da2b" (UID: "329e73d1-df45-481a-b897-1bde9107da2b"). InnerVolumeSpecName "kube-api-access-nw26v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.164947 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-inventory" (OuterVolumeSpecName: "inventory") pod "329e73d1-df45-481a-b897-1bde9107da2b" (UID: "329e73d1-df45-481a-b897-1bde9107da2b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.165470 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/329e73d1-df45-481a-b897-1bde9107da2b-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "329e73d1-df45-481a-b897-1bde9107da2b" (UID: "329e73d1-df45-481a-b897-1bde9107da2b"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.166706 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "329e73d1-df45-481a-b897-1bde9107da2b" (UID: "329e73d1-df45-481a-b897-1bde9107da2b"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.169377 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "329e73d1-df45-481a-b897-1bde9107da2b" (UID: "329e73d1-df45-481a-b897-1bde9107da2b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.170524 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/329e73d1-df45-481a-b897-1bde9107da2b-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "329e73d1-df45-481a-b897-1bde9107da2b" (UID: "329e73d1-df45-481a-b897-1bde9107da2b"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.172356 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "329e73d1-df45-481a-b897-1bde9107da2b" (UID: "329e73d1-df45-481a-b897-1bde9107da2b"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.173277 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "329e73d1-df45-481a-b897-1bde9107da2b" (UID: "329e73d1-df45-481a-b897-1bde9107da2b"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.175870 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "329e73d1-df45-481a-b897-1bde9107da2b" (UID: "329e73d1-df45-481a-b897-1bde9107da2b"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.232990 4997 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.233027 4997 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.233038 4997 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/329e73d1-df45-481a-b897-1bde9107da2b-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.233047 4997 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.233056 4997 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/329e73d1-df45-481a-b897-1bde9107da2b-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.233064 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nw26v\" (UniqueName: \"kubernetes.io/projected/329e73d1-df45-481a-b897-1bde9107da2b-kube-api-access-nw26v\") on node \"crc\" DevicePath \"\"" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.233074 4997 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.233083 4997 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.233092 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.233101 4997 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.233109 4997 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/329e73d1-df45-481a-b897-1bde9107da2b-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.558681 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" event={"ID":"329e73d1-df45-481a-b897-1bde9107da2b","Type":"ContainerDied","Data":"97d60535b245a99001202f575fda928b291991751947062c9daeb0a1880b4720"} Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.559041 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97d60535b245a99001202f575fda928b291991751947062c9daeb0a1880b4720" Dec 05 09:49:05 crc kubenswrapper[4997]: I1205 09:49:05.558730 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh" Dec 05 09:49:06 crc kubenswrapper[4997]: I1205 09:49:06.749474 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:49:06 crc kubenswrapper[4997]: E1205 09:49:06.749829 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:49:20 crc kubenswrapper[4997]: I1205 09:49:20.750120 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:49:20 crc kubenswrapper[4997]: E1205 09:49:20.750988 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:49:34 crc kubenswrapper[4997]: I1205 09:49:34.750345 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:49:34 crc kubenswrapper[4997]: E1205 09:49:34.751401 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:49:46 crc kubenswrapper[4997]: I1205 09:49:46.749425 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:49:46 crc kubenswrapper[4997]: E1205 09:49:46.750258 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:50:01 crc kubenswrapper[4997]: I1205 09:50:01.749257 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:50:01 crc kubenswrapper[4997]: E1205 09:50:01.750103 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:50:13 crc kubenswrapper[4997]: I1205 09:50:13.749406 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:50:13 crc kubenswrapper[4997]: E1205 09:50:13.750405 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:50:28 crc kubenswrapper[4997]: I1205 09:50:28.749520 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:50:28 crc kubenswrapper[4997]: E1205 09:50:28.750318 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:50:39 crc kubenswrapper[4997]: I1205 09:50:39.271202 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Dec 05 09:50:39 crc kubenswrapper[4997]: I1205 09:50:39.272014 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="ce49c282-2c15-4809-9fea-1df3e8750974" containerName="adoption" containerID="cri-o://5b48de53aeeb2c625a8c80083e172a9f3b0c470cf58770302dc3610f17279bd5" gracePeriod=30 Dec 05 09:50:41 crc kubenswrapper[4997]: I1205 09:50:41.750510 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:50:41 crc kubenswrapper[4997]: E1205 09:50:41.751258 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:50:53 crc kubenswrapper[4997]: I1205 09:50:53.749287 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:50:53 crc kubenswrapper[4997]: E1205 09:50:53.750098 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:51:08 crc kubenswrapper[4997]: I1205 09:51:08.749922 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:51:08 crc kubenswrapper[4997]: E1205 09:51:08.751974 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:51:09 crc kubenswrapper[4997]: I1205 09:51:09.777004 4997 generic.go:334] "Generic (PLEG): container finished" podID="ce49c282-2c15-4809-9fea-1df3e8750974" containerID="5b48de53aeeb2c625a8c80083e172a9f3b0c470cf58770302dc3610f17279bd5" exitCode=137 Dec 05 09:51:09 crc kubenswrapper[4997]: I1205 09:51:09.777249 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"ce49c282-2c15-4809-9fea-1df3e8750974","Type":"ContainerDied","Data":"5b48de53aeeb2c625a8c80083e172a9f3b0c470cf58770302dc3610f17279bd5"} Dec 05 09:51:09 crc kubenswrapper[4997]: I1205 09:51:09.777276 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"ce49c282-2c15-4809-9fea-1df3e8750974","Type":"ContainerDied","Data":"b1f0bc0182fbf5f6d24f64a74e445296f60390f6c735a8e331d68fdf372748e8"} Dec 05 09:51:09 crc kubenswrapper[4997]: I1205 09:51:09.777286 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1f0bc0182fbf5f6d24f64a74e445296f60390f6c735a8e331d68fdf372748e8" Dec 05 09:51:09 crc kubenswrapper[4997]: I1205 09:51:09.832255 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 05 09:51:09 crc kubenswrapper[4997]: I1205 09:51:09.978807 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5b900c62-9e5f-4417-9360-c4cb3361b7a6\") pod \"ce49c282-2c15-4809-9fea-1df3e8750974\" (UID: \"ce49c282-2c15-4809-9fea-1df3e8750974\") " Dec 05 09:51:09 crc kubenswrapper[4997]: I1205 09:51:09.979093 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnxmb\" (UniqueName: \"kubernetes.io/projected/ce49c282-2c15-4809-9fea-1df3e8750974-kube-api-access-wnxmb\") pod \"ce49c282-2c15-4809-9fea-1df3e8750974\" (UID: \"ce49c282-2c15-4809-9fea-1df3e8750974\") " Dec 05 09:51:09 crc kubenswrapper[4997]: I1205 09:51:09.984838 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce49c282-2c15-4809-9fea-1df3e8750974-kube-api-access-wnxmb" (OuterVolumeSpecName: "kube-api-access-wnxmb") pod "ce49c282-2c15-4809-9fea-1df3e8750974" (UID: "ce49c282-2c15-4809-9fea-1df3e8750974"). InnerVolumeSpecName "kube-api-access-wnxmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:51:10 crc kubenswrapper[4997]: I1205 09:51:10.001523 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5b900c62-9e5f-4417-9360-c4cb3361b7a6" (OuterVolumeSpecName: "mariadb-data") pod "ce49c282-2c15-4809-9fea-1df3e8750974" (UID: "ce49c282-2c15-4809-9fea-1df3e8750974"). InnerVolumeSpecName "pvc-5b900c62-9e5f-4417-9360-c4cb3361b7a6". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 09:51:10 crc kubenswrapper[4997]: I1205 09:51:10.082928 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnxmb\" (UniqueName: \"kubernetes.io/projected/ce49c282-2c15-4809-9fea-1df3e8750974-kube-api-access-wnxmb\") on node \"crc\" DevicePath \"\"" Dec 05 09:51:10 crc kubenswrapper[4997]: I1205 09:51:10.082984 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-5b900c62-9e5f-4417-9360-c4cb3361b7a6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5b900c62-9e5f-4417-9360-c4cb3361b7a6\") on node \"crc\" " Dec 05 09:51:10 crc kubenswrapper[4997]: I1205 09:51:10.112076 4997 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 05 09:51:10 crc kubenswrapper[4997]: I1205 09:51:10.112230 4997 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-5b900c62-9e5f-4417-9360-c4cb3361b7a6" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5b900c62-9e5f-4417-9360-c4cb3361b7a6") on node "crc" Dec 05 09:51:10 crc kubenswrapper[4997]: I1205 09:51:10.186242 4997 reconciler_common.go:293] "Volume detached for volume \"pvc-5b900c62-9e5f-4417-9360-c4cb3361b7a6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5b900c62-9e5f-4417-9360-c4cb3361b7a6\") on node \"crc\" DevicePath \"\"" Dec 05 09:51:10 crc kubenswrapper[4997]: I1205 09:51:10.785155 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 05 09:51:10 crc kubenswrapper[4997]: I1205 09:51:10.821534 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Dec 05 09:51:10 crc kubenswrapper[4997]: I1205 09:51:10.832252 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Dec 05 09:51:11 crc kubenswrapper[4997]: I1205 09:51:11.460934 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Dec 05 09:51:11 crc kubenswrapper[4997]: I1205 09:51:11.461425 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="650f03f7-065f-42d0-8dc4-9092bc28a2d4" containerName="adoption" containerID="cri-o://2889232d7b711b570b629cbe31d13490494287816ec612e7302c2aaa68fab403" gracePeriod=30 Dec 05 09:51:11 crc kubenswrapper[4997]: I1205 09:51:11.765928 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce49c282-2c15-4809-9fea-1df3e8750974" path="/var/lib/kubelet/pods/ce49c282-2c15-4809-9fea-1df3e8750974/volumes" Dec 05 09:51:20 crc kubenswrapper[4997]: I1205 09:51:20.891751 4997 scope.go:117] "RemoveContainer" containerID="5b48de53aeeb2c625a8c80083e172a9f3b0c470cf58770302dc3610f17279bd5" Dec 05 09:51:21 crc kubenswrapper[4997]: I1205 09:51:21.749642 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:51:21 crc kubenswrapper[4997]: E1205 09:51:21.750329 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:51:33 crc kubenswrapper[4997]: I1205 09:51:33.749161 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:51:33 crc kubenswrapper[4997]: E1205 09:51:33.749987 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.782419 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cw9v8"] Dec 05 09:51:38 crc kubenswrapper[4997]: E1205 09:51:38.784730 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f232c7-354a-4226-b5a0-637bf10d598f" containerName="extract-content" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.784875 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f232c7-354a-4226-b5a0-637bf10d598f" containerName="extract-content" Dec 05 09:51:38 crc kubenswrapper[4997]: E1205 09:51:38.784978 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f232c7-354a-4226-b5a0-637bf10d598f" containerName="registry-server" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.785066 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f232c7-354a-4226-b5a0-637bf10d598f" containerName="registry-server" Dec 05 09:51:38 crc kubenswrapper[4997]: E1205 09:51:38.785154 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce49c282-2c15-4809-9fea-1df3e8750974" containerName="adoption" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.785237 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce49c282-2c15-4809-9fea-1df3e8750974" containerName="adoption" Dec 05 09:51:38 crc kubenswrapper[4997]: E1205 09:51:38.785357 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f232c7-354a-4226-b5a0-637bf10d598f" containerName="extract-utilities" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.785437 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f232c7-354a-4226-b5a0-637bf10d598f" containerName="extract-utilities" Dec 05 09:51:38 crc kubenswrapper[4997]: E1205 09:51:38.785545 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="329e73d1-df45-481a-b897-1bde9107da2b" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.785647 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="329e73d1-df45-481a-b897-1bde9107da2b" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.786066 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8f232c7-354a-4226-b5a0-637bf10d598f" containerName="registry-server" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.786180 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="329e73d1-df45-481a-b897-1bde9107da2b" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.786282 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce49c282-2c15-4809-9fea-1df3e8750974" containerName="adoption" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.789557 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cw9v8" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.795029 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cw9v8"] Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.834806 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c14771d2-0a1a-4734-b551-c11227ee79f9-utilities\") pod \"community-operators-cw9v8\" (UID: \"c14771d2-0a1a-4734-b551-c11227ee79f9\") " pod="openshift-marketplace/community-operators-cw9v8" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.835179 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c14771d2-0a1a-4734-b551-c11227ee79f9-catalog-content\") pod \"community-operators-cw9v8\" (UID: \"c14771d2-0a1a-4734-b551-c11227ee79f9\") " pod="openshift-marketplace/community-operators-cw9v8" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.835726 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6qgw\" (UniqueName: \"kubernetes.io/projected/c14771d2-0a1a-4734-b551-c11227ee79f9-kube-api-access-s6qgw\") pod \"community-operators-cw9v8\" (UID: \"c14771d2-0a1a-4734-b551-c11227ee79f9\") " pod="openshift-marketplace/community-operators-cw9v8" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.937258 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6qgw\" (UniqueName: \"kubernetes.io/projected/c14771d2-0a1a-4734-b551-c11227ee79f9-kube-api-access-s6qgw\") pod \"community-operators-cw9v8\" (UID: \"c14771d2-0a1a-4734-b551-c11227ee79f9\") " pod="openshift-marketplace/community-operators-cw9v8" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.938016 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c14771d2-0a1a-4734-b551-c11227ee79f9-utilities\") pod \"community-operators-cw9v8\" (UID: \"c14771d2-0a1a-4734-b551-c11227ee79f9\") " pod="openshift-marketplace/community-operators-cw9v8" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.938437 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c14771d2-0a1a-4734-b551-c11227ee79f9-utilities\") pod \"community-operators-cw9v8\" (UID: \"c14771d2-0a1a-4734-b551-c11227ee79f9\") " pod="openshift-marketplace/community-operators-cw9v8" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.938570 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c14771d2-0a1a-4734-b551-c11227ee79f9-catalog-content\") pod \"community-operators-cw9v8\" (UID: \"c14771d2-0a1a-4734-b551-c11227ee79f9\") " pod="openshift-marketplace/community-operators-cw9v8" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.938842 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c14771d2-0a1a-4734-b551-c11227ee79f9-catalog-content\") pod \"community-operators-cw9v8\" (UID: \"c14771d2-0a1a-4734-b551-c11227ee79f9\") " pod="openshift-marketplace/community-operators-cw9v8" Dec 05 09:51:38 crc kubenswrapper[4997]: I1205 09:51:38.956786 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6qgw\" (UniqueName: \"kubernetes.io/projected/c14771d2-0a1a-4734-b551-c11227ee79f9-kube-api-access-s6qgw\") pod \"community-operators-cw9v8\" (UID: \"c14771d2-0a1a-4734-b551-c11227ee79f9\") " pod="openshift-marketplace/community-operators-cw9v8" Dec 05 09:51:39 crc kubenswrapper[4997]: I1205 09:51:39.121510 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cw9v8" Dec 05 09:51:39 crc kubenswrapper[4997]: I1205 09:51:39.694034 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cw9v8"] Dec 05 09:51:40 crc kubenswrapper[4997]: I1205 09:51:40.115250 4997 generic.go:334] "Generic (PLEG): container finished" podID="c14771d2-0a1a-4734-b551-c11227ee79f9" containerID="0931d9bd87a61f6c4b079a4a0d019592bd8a00f8997f9f204289213c4a63769b" exitCode=0 Dec 05 09:51:40 crc kubenswrapper[4997]: I1205 09:51:40.115313 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cw9v8" event={"ID":"c14771d2-0a1a-4734-b551-c11227ee79f9","Type":"ContainerDied","Data":"0931d9bd87a61f6c4b079a4a0d019592bd8a00f8997f9f204289213c4a63769b"} Dec 05 09:51:40 crc kubenswrapper[4997]: I1205 09:51:40.115371 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cw9v8" event={"ID":"c14771d2-0a1a-4734-b551-c11227ee79f9","Type":"ContainerStarted","Data":"ce1f8f811418e54fd4b09f562b54b57d368b3bf1cf50f1cf7872ca09279a80d8"} Dec 05 09:51:40 crc kubenswrapper[4997]: I1205 09:51:40.119302 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:51:41 crc kubenswrapper[4997]: I1205 09:51:41.980839 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.011481 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k77h\" (UniqueName: \"kubernetes.io/projected/650f03f7-065f-42d0-8dc4-9092bc28a2d4-kube-api-access-9k77h\") pod \"650f03f7-065f-42d0-8dc4-9092bc28a2d4\" (UID: \"650f03f7-065f-42d0-8dc4-9092bc28a2d4\") " Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.011749 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/650f03f7-065f-42d0-8dc4-9092bc28a2d4-ovn-data-cert\") pod \"650f03f7-065f-42d0-8dc4-9092bc28a2d4\" (UID: \"650f03f7-065f-42d0-8dc4-9092bc28a2d4\") " Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.012857 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b94d1-ea00-4a16-832d-09cf519ec342\") pod \"650f03f7-065f-42d0-8dc4-9092bc28a2d4\" (UID: \"650f03f7-065f-42d0-8dc4-9092bc28a2d4\") " Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.017437 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/650f03f7-065f-42d0-8dc4-9092bc28a2d4-kube-api-access-9k77h" (OuterVolumeSpecName: "kube-api-access-9k77h") pod "650f03f7-065f-42d0-8dc4-9092bc28a2d4" (UID: "650f03f7-065f-42d0-8dc4-9092bc28a2d4"). InnerVolumeSpecName "kube-api-access-9k77h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.018608 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/650f03f7-065f-42d0-8dc4-9092bc28a2d4-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "650f03f7-065f-42d0-8dc4-9092bc28a2d4" (UID: "650f03f7-065f-42d0-8dc4-9092bc28a2d4"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.029494 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b94d1-ea00-4a16-832d-09cf519ec342" (OuterVolumeSpecName: "ovn-data") pod "650f03f7-065f-42d0-8dc4-9092bc28a2d4" (UID: "650f03f7-065f-42d0-8dc4-9092bc28a2d4"). InnerVolumeSpecName "pvc-e75b94d1-ea00-4a16-832d-09cf519ec342". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.115090 4997 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/650f03f7-065f-42d0-8dc4-9092bc28a2d4-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.115146 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-e75b94d1-ea00-4a16-832d-09cf519ec342\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b94d1-ea00-4a16-832d-09cf519ec342\") on node \"crc\" " Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.115159 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k77h\" (UniqueName: \"kubernetes.io/projected/650f03f7-065f-42d0-8dc4-9092bc28a2d4-kube-api-access-9k77h\") on node \"crc\" DevicePath \"\"" Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.143277 4997 generic.go:334] "Generic (PLEG): container finished" podID="650f03f7-065f-42d0-8dc4-9092bc28a2d4" containerID="2889232d7b711b570b629cbe31d13490494287816ec612e7302c2aaa68fab403" exitCode=137 Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.143320 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"650f03f7-065f-42d0-8dc4-9092bc28a2d4","Type":"ContainerDied","Data":"2889232d7b711b570b629cbe31d13490494287816ec612e7302c2aaa68fab403"} Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.143345 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"650f03f7-065f-42d0-8dc4-9092bc28a2d4","Type":"ContainerDied","Data":"22d57cff64dd19e8c21b3bf2cf4f1e80d3b172b7ae29908ea4264aed24e9d91d"} Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.143361 4997 scope.go:117] "RemoveContainer" containerID="2889232d7b711b570b629cbe31d13490494287816ec612e7302c2aaa68fab403" Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.143503 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.148821 4997 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.149174 4997 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-e75b94d1-ea00-4a16-832d-09cf519ec342" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b94d1-ea00-4a16-832d-09cf519ec342") on node "crc" Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.167137 4997 scope.go:117] "RemoveContainer" containerID="2889232d7b711b570b629cbe31d13490494287816ec612e7302c2aaa68fab403" Dec 05 09:51:42 crc kubenswrapper[4997]: E1205 09:51:42.167668 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2889232d7b711b570b629cbe31d13490494287816ec612e7302c2aaa68fab403\": container with ID starting with 2889232d7b711b570b629cbe31d13490494287816ec612e7302c2aaa68fab403 not found: ID does not exist" containerID="2889232d7b711b570b629cbe31d13490494287816ec612e7302c2aaa68fab403" Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.167700 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2889232d7b711b570b629cbe31d13490494287816ec612e7302c2aaa68fab403"} err="failed to get container status \"2889232d7b711b570b629cbe31d13490494287816ec612e7302c2aaa68fab403\": rpc error: code = NotFound desc = could not find container \"2889232d7b711b570b629cbe31d13490494287816ec612e7302c2aaa68fab403\": container with ID starting with 2889232d7b711b570b629cbe31d13490494287816ec612e7302c2aaa68fab403 not found: ID does not exist" Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.188066 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.197992 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Dec 05 09:51:42 crc kubenswrapper[4997]: I1205 09:51:42.218269 4997 reconciler_common.go:293] "Volume detached for volume \"pvc-e75b94d1-ea00-4a16-832d-09cf519ec342\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b94d1-ea00-4a16-832d-09cf519ec342\") on node \"crc\" DevicePath \"\"" Dec 05 09:51:43 crc kubenswrapper[4997]: I1205 09:51:43.761260 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="650f03f7-065f-42d0-8dc4-9092bc28a2d4" path="/var/lib/kubelet/pods/650f03f7-065f-42d0-8dc4-9092bc28a2d4/volumes" Dec 05 09:51:45 crc kubenswrapper[4997]: I1205 09:51:45.177947 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cw9v8" event={"ID":"c14771d2-0a1a-4734-b551-c11227ee79f9","Type":"ContainerStarted","Data":"11054c02cf87c5428ae1d1b31a9559a55fdebdb770747fb7fb5e236aab63ffa6"} Dec 05 09:51:46 crc kubenswrapper[4997]: I1205 09:51:46.189368 4997 generic.go:334] "Generic (PLEG): container finished" podID="c14771d2-0a1a-4734-b551-c11227ee79f9" containerID="11054c02cf87c5428ae1d1b31a9559a55fdebdb770747fb7fb5e236aab63ffa6" exitCode=0 Dec 05 09:51:46 crc kubenswrapper[4997]: I1205 09:51:46.189699 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cw9v8" event={"ID":"c14771d2-0a1a-4734-b551-c11227ee79f9","Type":"ContainerDied","Data":"11054c02cf87c5428ae1d1b31a9559a55fdebdb770747fb7fb5e236aab63ffa6"} Dec 05 09:51:47 crc kubenswrapper[4997]: I1205 09:51:47.232465 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cw9v8" event={"ID":"c14771d2-0a1a-4734-b551-c11227ee79f9","Type":"ContainerStarted","Data":"7eac36075d8ac8da4dd453a37bc9892b814e3219db002b4e8d4cbf236739dc1c"} Dec 05 09:51:47 crc kubenswrapper[4997]: I1205 09:51:47.269027 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cw9v8" podStartSLOduration=2.7598393789999998 podStartE2EDuration="9.268996443s" podCreationTimestamp="2025-12-05 09:51:38 +0000 UTC" firstStartedPulling="2025-12-05 09:51:40.118966262 +0000 UTC m=+10600.647873533" lastFinishedPulling="2025-12-05 09:51:46.628123336 +0000 UTC m=+10607.157030597" observedRunningTime="2025-12-05 09:51:47.249247349 +0000 UTC m=+10607.778154620" watchObservedRunningTime="2025-12-05 09:51:47.268996443 +0000 UTC m=+10607.797903714" Dec 05 09:51:47 crc kubenswrapper[4997]: I1205 09:51:47.540123 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gnkzn"] Dec 05 09:51:47 crc kubenswrapper[4997]: E1205 09:51:47.540665 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="650f03f7-065f-42d0-8dc4-9092bc28a2d4" containerName="adoption" Dec 05 09:51:47 crc kubenswrapper[4997]: I1205 09:51:47.540686 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="650f03f7-065f-42d0-8dc4-9092bc28a2d4" containerName="adoption" Dec 05 09:51:47 crc kubenswrapper[4997]: I1205 09:51:47.540939 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="650f03f7-065f-42d0-8dc4-9092bc28a2d4" containerName="adoption" Dec 05 09:51:47 crc kubenswrapper[4997]: I1205 09:51:47.542698 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gnkzn" Dec 05 09:51:47 crc kubenswrapper[4997]: I1205 09:51:47.565755 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnkzn"] Dec 05 09:51:47 crc kubenswrapper[4997]: I1205 09:51:47.629214 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36383cda-5afc-46ac-8cb8-56eb4d7da279-catalog-content\") pod \"redhat-marketplace-gnkzn\" (UID: \"36383cda-5afc-46ac-8cb8-56eb4d7da279\") " pod="openshift-marketplace/redhat-marketplace-gnkzn" Dec 05 09:51:47 crc kubenswrapper[4997]: I1205 09:51:47.629276 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzznc\" (UniqueName: \"kubernetes.io/projected/36383cda-5afc-46ac-8cb8-56eb4d7da279-kube-api-access-zzznc\") pod \"redhat-marketplace-gnkzn\" (UID: \"36383cda-5afc-46ac-8cb8-56eb4d7da279\") " pod="openshift-marketplace/redhat-marketplace-gnkzn" Dec 05 09:51:47 crc kubenswrapper[4997]: I1205 09:51:47.629312 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36383cda-5afc-46ac-8cb8-56eb4d7da279-utilities\") pod \"redhat-marketplace-gnkzn\" (UID: \"36383cda-5afc-46ac-8cb8-56eb4d7da279\") " pod="openshift-marketplace/redhat-marketplace-gnkzn" Dec 05 09:51:47 crc kubenswrapper[4997]: I1205 09:51:47.731056 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36383cda-5afc-46ac-8cb8-56eb4d7da279-catalog-content\") pod \"redhat-marketplace-gnkzn\" (UID: \"36383cda-5afc-46ac-8cb8-56eb4d7da279\") " pod="openshift-marketplace/redhat-marketplace-gnkzn" Dec 05 09:51:47 crc kubenswrapper[4997]: I1205 09:51:47.731151 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzznc\" (UniqueName: \"kubernetes.io/projected/36383cda-5afc-46ac-8cb8-56eb4d7da279-kube-api-access-zzznc\") pod \"redhat-marketplace-gnkzn\" (UID: \"36383cda-5afc-46ac-8cb8-56eb4d7da279\") " pod="openshift-marketplace/redhat-marketplace-gnkzn" Dec 05 09:51:47 crc kubenswrapper[4997]: I1205 09:51:47.731211 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36383cda-5afc-46ac-8cb8-56eb4d7da279-utilities\") pod \"redhat-marketplace-gnkzn\" (UID: \"36383cda-5afc-46ac-8cb8-56eb4d7da279\") " pod="openshift-marketplace/redhat-marketplace-gnkzn" Dec 05 09:51:47 crc kubenswrapper[4997]: I1205 09:51:47.731525 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36383cda-5afc-46ac-8cb8-56eb4d7da279-catalog-content\") pod \"redhat-marketplace-gnkzn\" (UID: \"36383cda-5afc-46ac-8cb8-56eb4d7da279\") " pod="openshift-marketplace/redhat-marketplace-gnkzn" Dec 05 09:51:47 crc kubenswrapper[4997]: I1205 09:51:47.731561 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36383cda-5afc-46ac-8cb8-56eb4d7da279-utilities\") pod \"redhat-marketplace-gnkzn\" (UID: \"36383cda-5afc-46ac-8cb8-56eb4d7da279\") " pod="openshift-marketplace/redhat-marketplace-gnkzn" Dec 05 09:51:47 crc kubenswrapper[4997]: I1205 09:51:47.757652 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzznc\" (UniqueName: \"kubernetes.io/projected/36383cda-5afc-46ac-8cb8-56eb4d7da279-kube-api-access-zzznc\") pod \"redhat-marketplace-gnkzn\" (UID: \"36383cda-5afc-46ac-8cb8-56eb4d7da279\") " pod="openshift-marketplace/redhat-marketplace-gnkzn" Dec 05 09:51:47 crc kubenswrapper[4997]: I1205 09:51:47.878969 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gnkzn" Dec 05 09:51:48 crc kubenswrapper[4997]: I1205 09:51:48.418729 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnkzn"] Dec 05 09:51:48 crc kubenswrapper[4997]: I1205 09:51:48.748935 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:51:48 crc kubenswrapper[4997]: E1205 09:51:48.749436 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:51:49 crc kubenswrapper[4997]: I1205 09:51:49.122632 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cw9v8" Dec 05 09:51:49 crc kubenswrapper[4997]: I1205 09:51:49.122708 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cw9v8" Dec 05 09:51:49 crc kubenswrapper[4997]: I1205 09:51:49.177383 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cw9v8" Dec 05 09:51:49 crc kubenswrapper[4997]: I1205 09:51:49.254368 4997 generic.go:334] "Generic (PLEG): container finished" podID="36383cda-5afc-46ac-8cb8-56eb4d7da279" containerID="8d7a7fb0864d2b7e629dd0a7c17d539e9f6f61709d5d544325533224460589e6" exitCode=0 Dec 05 09:51:49 crc kubenswrapper[4997]: I1205 09:51:49.254465 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnkzn" event={"ID":"36383cda-5afc-46ac-8cb8-56eb4d7da279","Type":"ContainerDied","Data":"8d7a7fb0864d2b7e629dd0a7c17d539e9f6f61709d5d544325533224460589e6"} Dec 05 09:51:49 crc kubenswrapper[4997]: I1205 09:51:49.254724 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnkzn" event={"ID":"36383cda-5afc-46ac-8cb8-56eb4d7da279","Type":"ContainerStarted","Data":"e62f01973713ae5c4db563dc83b8e0f77e994d6670ff31266bf3a5762599d3c7"} Dec 05 09:51:51 crc kubenswrapper[4997]: I1205 09:51:51.276597 4997 generic.go:334] "Generic (PLEG): container finished" podID="36383cda-5afc-46ac-8cb8-56eb4d7da279" containerID="50dcbafdb6fcd8f91b7f67ef8b94026f86b0b6f6df95447367cc565b0bd1cb6c" exitCode=0 Dec 05 09:51:51 crc kubenswrapper[4997]: I1205 09:51:51.276712 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnkzn" event={"ID":"36383cda-5afc-46ac-8cb8-56eb4d7da279","Type":"ContainerDied","Data":"50dcbafdb6fcd8f91b7f67ef8b94026f86b0b6f6df95447367cc565b0bd1cb6c"} Dec 05 09:51:52 crc kubenswrapper[4997]: I1205 09:51:52.290697 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnkzn" event={"ID":"36383cda-5afc-46ac-8cb8-56eb4d7da279","Type":"ContainerStarted","Data":"2ac99050fa6690865c8a8de88dc9c14abe44ad2882e7fda5c7f1ede4eddd26f1"} Dec 05 09:51:52 crc kubenswrapper[4997]: I1205 09:51:52.314685 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gnkzn" podStartSLOduration=2.8948054560000003 podStartE2EDuration="5.31466652s" podCreationTimestamp="2025-12-05 09:51:47 +0000 UTC" firstStartedPulling="2025-12-05 09:51:49.257855745 +0000 UTC m=+10609.786763006" lastFinishedPulling="2025-12-05 09:51:51.677716809 +0000 UTC m=+10612.206624070" observedRunningTime="2025-12-05 09:51:52.307337382 +0000 UTC m=+10612.836244643" watchObservedRunningTime="2025-12-05 09:51:52.31466652 +0000 UTC m=+10612.843573781" Dec 05 09:51:57 crc kubenswrapper[4997]: I1205 09:51:57.880010 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gnkzn" Dec 05 09:51:57 crc kubenswrapper[4997]: I1205 09:51:57.880403 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gnkzn" Dec 05 09:51:57 crc kubenswrapper[4997]: I1205 09:51:57.925982 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gnkzn" Dec 05 09:51:58 crc kubenswrapper[4997]: I1205 09:51:58.407858 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gnkzn" Dec 05 09:51:58 crc kubenswrapper[4997]: I1205 09:51:58.459424 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnkzn"] Dec 05 09:51:59 crc kubenswrapper[4997]: I1205 09:51:59.186555 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cw9v8" Dec 05 09:52:00 crc kubenswrapper[4997]: I1205 09:52:00.395718 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gnkzn" podUID="36383cda-5afc-46ac-8cb8-56eb4d7da279" containerName="registry-server" containerID="cri-o://2ac99050fa6690865c8a8de88dc9c14abe44ad2882e7fda5c7f1ede4eddd26f1" gracePeriod=2 Dec 05 09:52:00 crc kubenswrapper[4997]: I1205 09:52:00.567045 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cw9v8"] Dec 05 09:52:00 crc kubenswrapper[4997]: I1205 09:52:00.567529 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cw9v8" podUID="c14771d2-0a1a-4734-b551-c11227ee79f9" containerName="registry-server" containerID="cri-o://7eac36075d8ac8da4dd453a37bc9892b814e3219db002b4e8d4cbf236739dc1c" gracePeriod=2 Dec 05 09:52:00 crc kubenswrapper[4997]: I1205 09:52:00.754404 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:52:00 crc kubenswrapper[4997]: I1205 09:52:00.954537 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gnkzn" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.030051 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36383cda-5afc-46ac-8cb8-56eb4d7da279-utilities\") pod \"36383cda-5afc-46ac-8cb8-56eb4d7da279\" (UID: \"36383cda-5afc-46ac-8cb8-56eb4d7da279\") " Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.030082 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzznc\" (UniqueName: \"kubernetes.io/projected/36383cda-5afc-46ac-8cb8-56eb4d7da279-kube-api-access-zzznc\") pod \"36383cda-5afc-46ac-8cb8-56eb4d7da279\" (UID: \"36383cda-5afc-46ac-8cb8-56eb4d7da279\") " Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.030135 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36383cda-5afc-46ac-8cb8-56eb4d7da279-catalog-content\") pod \"36383cda-5afc-46ac-8cb8-56eb4d7da279\" (UID: \"36383cda-5afc-46ac-8cb8-56eb4d7da279\") " Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.032167 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36383cda-5afc-46ac-8cb8-56eb4d7da279-utilities" (OuterVolumeSpecName: "utilities") pod "36383cda-5afc-46ac-8cb8-56eb4d7da279" (UID: "36383cda-5afc-46ac-8cb8-56eb4d7da279"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.036331 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36383cda-5afc-46ac-8cb8-56eb4d7da279-kube-api-access-zzznc" (OuterVolumeSpecName: "kube-api-access-zzznc") pod "36383cda-5afc-46ac-8cb8-56eb4d7da279" (UID: "36383cda-5afc-46ac-8cb8-56eb4d7da279"). InnerVolumeSpecName "kube-api-access-zzznc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.050105 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36383cda-5afc-46ac-8cb8-56eb4d7da279-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "36383cda-5afc-46ac-8cb8-56eb4d7da279" (UID: "36383cda-5afc-46ac-8cb8-56eb4d7da279"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.061897 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cw9v8" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.131727 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6qgw\" (UniqueName: \"kubernetes.io/projected/c14771d2-0a1a-4734-b551-c11227ee79f9-kube-api-access-s6qgw\") pod \"c14771d2-0a1a-4734-b551-c11227ee79f9\" (UID: \"c14771d2-0a1a-4734-b551-c11227ee79f9\") " Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.132105 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c14771d2-0a1a-4734-b551-c11227ee79f9-utilities\") pod \"c14771d2-0a1a-4734-b551-c11227ee79f9\" (UID: \"c14771d2-0a1a-4734-b551-c11227ee79f9\") " Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.132213 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c14771d2-0a1a-4734-b551-c11227ee79f9-catalog-content\") pod \"c14771d2-0a1a-4734-b551-c11227ee79f9\" (UID: \"c14771d2-0a1a-4734-b551-c11227ee79f9\") " Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.132631 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36383cda-5afc-46ac-8cb8-56eb4d7da279-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.132650 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzznc\" (UniqueName: \"kubernetes.io/projected/36383cda-5afc-46ac-8cb8-56eb4d7da279-kube-api-access-zzznc\") on node \"crc\" DevicePath \"\"" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.132658 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36383cda-5afc-46ac-8cb8-56eb4d7da279-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.132925 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c14771d2-0a1a-4734-b551-c11227ee79f9-utilities" (OuterVolumeSpecName: "utilities") pod "c14771d2-0a1a-4734-b551-c11227ee79f9" (UID: "c14771d2-0a1a-4734-b551-c11227ee79f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.135093 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c14771d2-0a1a-4734-b551-c11227ee79f9-kube-api-access-s6qgw" (OuterVolumeSpecName: "kube-api-access-s6qgw") pod "c14771d2-0a1a-4734-b551-c11227ee79f9" (UID: "c14771d2-0a1a-4734-b551-c11227ee79f9"). InnerVolumeSpecName "kube-api-access-s6qgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.193284 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c14771d2-0a1a-4734-b551-c11227ee79f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c14771d2-0a1a-4734-b551-c11227ee79f9" (UID: "c14771d2-0a1a-4734-b551-c11227ee79f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.234932 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6qgw\" (UniqueName: \"kubernetes.io/projected/c14771d2-0a1a-4734-b551-c11227ee79f9-kube-api-access-s6qgw\") on node \"crc\" DevicePath \"\"" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.234963 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c14771d2-0a1a-4734-b551-c11227ee79f9-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.234974 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c14771d2-0a1a-4734-b551-c11227ee79f9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.410350 4997 generic.go:334] "Generic (PLEG): container finished" podID="36383cda-5afc-46ac-8cb8-56eb4d7da279" containerID="2ac99050fa6690865c8a8de88dc9c14abe44ad2882e7fda5c7f1ede4eddd26f1" exitCode=0 Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.410423 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnkzn" event={"ID":"36383cda-5afc-46ac-8cb8-56eb4d7da279","Type":"ContainerDied","Data":"2ac99050fa6690865c8a8de88dc9c14abe44ad2882e7fda5c7f1ede4eddd26f1"} Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.410476 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gnkzn" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.411648 4997 scope.go:117] "RemoveContainer" containerID="2ac99050fa6690865c8a8de88dc9c14abe44ad2882e7fda5c7f1ede4eddd26f1" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.411556 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnkzn" event={"ID":"36383cda-5afc-46ac-8cb8-56eb4d7da279","Type":"ContainerDied","Data":"e62f01973713ae5c4db563dc83b8e0f77e994d6670ff31266bf3a5762599d3c7"} Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.433881 4997 generic.go:334] "Generic (PLEG): container finished" podID="c14771d2-0a1a-4734-b551-c11227ee79f9" containerID="7eac36075d8ac8da4dd453a37bc9892b814e3219db002b4e8d4cbf236739dc1c" exitCode=0 Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.433972 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cw9v8" event={"ID":"c14771d2-0a1a-4734-b551-c11227ee79f9","Type":"ContainerDied","Data":"7eac36075d8ac8da4dd453a37bc9892b814e3219db002b4e8d4cbf236739dc1c"} Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.434006 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cw9v8" event={"ID":"c14771d2-0a1a-4734-b551-c11227ee79f9","Type":"ContainerDied","Data":"ce1f8f811418e54fd4b09f562b54b57d368b3bf1cf50f1cf7872ca09279a80d8"} Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.434235 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cw9v8" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.441360 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"d10217ab4b647ccf0d7a3c346634fc388bc2f808fb062547e0c6bbe74fe6cdee"} Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.467403 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnkzn"] Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.469776 4997 scope.go:117] "RemoveContainer" containerID="50dcbafdb6fcd8f91b7f67ef8b94026f86b0b6f6df95447367cc565b0bd1cb6c" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.488521 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnkzn"] Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.518669 4997 scope.go:117] "RemoveContainer" containerID="8d7a7fb0864d2b7e629dd0a7c17d539e9f6f61709d5d544325533224460589e6" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.530899 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cw9v8"] Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.544118 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cw9v8"] Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.591782 4997 scope.go:117] "RemoveContainer" containerID="2ac99050fa6690865c8a8de88dc9c14abe44ad2882e7fda5c7f1ede4eddd26f1" Dec 05 09:52:01 crc kubenswrapper[4997]: E1205 09:52:01.592195 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ac99050fa6690865c8a8de88dc9c14abe44ad2882e7fda5c7f1ede4eddd26f1\": container with ID starting with 2ac99050fa6690865c8a8de88dc9c14abe44ad2882e7fda5c7f1ede4eddd26f1 not found: ID does not exist" containerID="2ac99050fa6690865c8a8de88dc9c14abe44ad2882e7fda5c7f1ede4eddd26f1" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.592235 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ac99050fa6690865c8a8de88dc9c14abe44ad2882e7fda5c7f1ede4eddd26f1"} err="failed to get container status \"2ac99050fa6690865c8a8de88dc9c14abe44ad2882e7fda5c7f1ede4eddd26f1\": rpc error: code = NotFound desc = could not find container \"2ac99050fa6690865c8a8de88dc9c14abe44ad2882e7fda5c7f1ede4eddd26f1\": container with ID starting with 2ac99050fa6690865c8a8de88dc9c14abe44ad2882e7fda5c7f1ede4eddd26f1 not found: ID does not exist" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.592266 4997 scope.go:117] "RemoveContainer" containerID="50dcbafdb6fcd8f91b7f67ef8b94026f86b0b6f6df95447367cc565b0bd1cb6c" Dec 05 09:52:01 crc kubenswrapper[4997]: E1205 09:52:01.592571 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50dcbafdb6fcd8f91b7f67ef8b94026f86b0b6f6df95447367cc565b0bd1cb6c\": container with ID starting with 50dcbafdb6fcd8f91b7f67ef8b94026f86b0b6f6df95447367cc565b0bd1cb6c not found: ID does not exist" containerID="50dcbafdb6fcd8f91b7f67ef8b94026f86b0b6f6df95447367cc565b0bd1cb6c" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.592606 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50dcbafdb6fcd8f91b7f67ef8b94026f86b0b6f6df95447367cc565b0bd1cb6c"} err="failed to get container status \"50dcbafdb6fcd8f91b7f67ef8b94026f86b0b6f6df95447367cc565b0bd1cb6c\": rpc error: code = NotFound desc = could not find container \"50dcbafdb6fcd8f91b7f67ef8b94026f86b0b6f6df95447367cc565b0bd1cb6c\": container with ID starting with 50dcbafdb6fcd8f91b7f67ef8b94026f86b0b6f6df95447367cc565b0bd1cb6c not found: ID does not exist" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.592653 4997 scope.go:117] "RemoveContainer" containerID="8d7a7fb0864d2b7e629dd0a7c17d539e9f6f61709d5d544325533224460589e6" Dec 05 09:52:01 crc kubenswrapper[4997]: E1205 09:52:01.592838 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d7a7fb0864d2b7e629dd0a7c17d539e9f6f61709d5d544325533224460589e6\": container with ID starting with 8d7a7fb0864d2b7e629dd0a7c17d539e9f6f61709d5d544325533224460589e6 not found: ID does not exist" containerID="8d7a7fb0864d2b7e629dd0a7c17d539e9f6f61709d5d544325533224460589e6" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.592862 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d7a7fb0864d2b7e629dd0a7c17d539e9f6f61709d5d544325533224460589e6"} err="failed to get container status \"8d7a7fb0864d2b7e629dd0a7c17d539e9f6f61709d5d544325533224460589e6\": rpc error: code = NotFound desc = could not find container \"8d7a7fb0864d2b7e629dd0a7c17d539e9f6f61709d5d544325533224460589e6\": container with ID starting with 8d7a7fb0864d2b7e629dd0a7c17d539e9f6f61709d5d544325533224460589e6 not found: ID does not exist" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.592874 4997 scope.go:117] "RemoveContainer" containerID="7eac36075d8ac8da4dd453a37bc9892b814e3219db002b4e8d4cbf236739dc1c" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.635974 4997 scope.go:117] "RemoveContainer" containerID="11054c02cf87c5428ae1d1b31a9559a55fdebdb770747fb7fb5e236aab63ffa6" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.660075 4997 scope.go:117] "RemoveContainer" containerID="0931d9bd87a61f6c4b079a4a0d019592bd8a00f8997f9f204289213c4a63769b" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.704204 4997 scope.go:117] "RemoveContainer" containerID="7eac36075d8ac8da4dd453a37bc9892b814e3219db002b4e8d4cbf236739dc1c" Dec 05 09:52:01 crc kubenswrapper[4997]: E1205 09:52:01.704642 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7eac36075d8ac8da4dd453a37bc9892b814e3219db002b4e8d4cbf236739dc1c\": container with ID starting with 7eac36075d8ac8da4dd453a37bc9892b814e3219db002b4e8d4cbf236739dc1c not found: ID does not exist" containerID="7eac36075d8ac8da4dd453a37bc9892b814e3219db002b4e8d4cbf236739dc1c" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.704668 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eac36075d8ac8da4dd453a37bc9892b814e3219db002b4e8d4cbf236739dc1c"} err="failed to get container status \"7eac36075d8ac8da4dd453a37bc9892b814e3219db002b4e8d4cbf236739dc1c\": rpc error: code = NotFound desc = could not find container \"7eac36075d8ac8da4dd453a37bc9892b814e3219db002b4e8d4cbf236739dc1c\": container with ID starting with 7eac36075d8ac8da4dd453a37bc9892b814e3219db002b4e8d4cbf236739dc1c not found: ID does not exist" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.704688 4997 scope.go:117] "RemoveContainer" containerID="11054c02cf87c5428ae1d1b31a9559a55fdebdb770747fb7fb5e236aab63ffa6" Dec 05 09:52:01 crc kubenswrapper[4997]: E1205 09:52:01.704889 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11054c02cf87c5428ae1d1b31a9559a55fdebdb770747fb7fb5e236aab63ffa6\": container with ID starting with 11054c02cf87c5428ae1d1b31a9559a55fdebdb770747fb7fb5e236aab63ffa6 not found: ID does not exist" containerID="11054c02cf87c5428ae1d1b31a9559a55fdebdb770747fb7fb5e236aab63ffa6" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.704909 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11054c02cf87c5428ae1d1b31a9559a55fdebdb770747fb7fb5e236aab63ffa6"} err="failed to get container status \"11054c02cf87c5428ae1d1b31a9559a55fdebdb770747fb7fb5e236aab63ffa6\": rpc error: code = NotFound desc = could not find container \"11054c02cf87c5428ae1d1b31a9559a55fdebdb770747fb7fb5e236aab63ffa6\": container with ID starting with 11054c02cf87c5428ae1d1b31a9559a55fdebdb770747fb7fb5e236aab63ffa6 not found: ID does not exist" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.704920 4997 scope.go:117] "RemoveContainer" containerID="0931d9bd87a61f6c4b079a4a0d019592bd8a00f8997f9f204289213c4a63769b" Dec 05 09:52:01 crc kubenswrapper[4997]: E1205 09:52:01.705091 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0931d9bd87a61f6c4b079a4a0d019592bd8a00f8997f9f204289213c4a63769b\": container with ID starting with 0931d9bd87a61f6c4b079a4a0d019592bd8a00f8997f9f204289213c4a63769b not found: ID does not exist" containerID="0931d9bd87a61f6c4b079a4a0d019592bd8a00f8997f9f204289213c4a63769b" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.705106 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0931d9bd87a61f6c4b079a4a0d019592bd8a00f8997f9f204289213c4a63769b"} err="failed to get container status \"0931d9bd87a61f6c4b079a4a0d019592bd8a00f8997f9f204289213c4a63769b\": rpc error: code = NotFound desc = could not find container \"0931d9bd87a61f6c4b079a4a0d019592bd8a00f8997f9f204289213c4a63769b\": container with ID starting with 0931d9bd87a61f6c4b079a4a0d019592bd8a00f8997f9f204289213c4a63769b not found: ID does not exist" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.760569 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36383cda-5afc-46ac-8cb8-56eb4d7da279" path="/var/lib/kubelet/pods/36383cda-5afc-46ac-8cb8-56eb4d7da279/volumes" Dec 05 09:52:01 crc kubenswrapper[4997]: I1205 09:52:01.761342 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c14771d2-0a1a-4734-b551-c11227ee79f9" path="/var/lib/kubelet/pods/c14771d2-0a1a-4734-b551-c11227ee79f9/volumes" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.688338 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f2m2h"] Dec 05 09:54:15 crc kubenswrapper[4997]: E1205 09:54:15.689748 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36383cda-5afc-46ac-8cb8-56eb4d7da279" containerName="extract-utilities" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.689788 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="36383cda-5afc-46ac-8cb8-56eb4d7da279" containerName="extract-utilities" Dec 05 09:54:15 crc kubenswrapper[4997]: E1205 09:54:15.689818 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c14771d2-0a1a-4734-b551-c11227ee79f9" containerName="extract-utilities" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.689831 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c14771d2-0a1a-4734-b551-c11227ee79f9" containerName="extract-utilities" Dec 05 09:54:15 crc kubenswrapper[4997]: E1205 09:54:15.689873 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c14771d2-0a1a-4734-b551-c11227ee79f9" containerName="extract-content" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.689887 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c14771d2-0a1a-4734-b551-c11227ee79f9" containerName="extract-content" Dec 05 09:54:15 crc kubenswrapper[4997]: E1205 09:54:15.689912 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36383cda-5afc-46ac-8cb8-56eb4d7da279" containerName="registry-server" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.689923 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="36383cda-5afc-46ac-8cb8-56eb4d7da279" containerName="registry-server" Dec 05 09:54:15 crc kubenswrapper[4997]: E1205 09:54:15.689946 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c14771d2-0a1a-4734-b551-c11227ee79f9" containerName="registry-server" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.689959 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c14771d2-0a1a-4734-b551-c11227ee79f9" containerName="registry-server" Dec 05 09:54:15 crc kubenswrapper[4997]: E1205 09:54:15.689982 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36383cda-5afc-46ac-8cb8-56eb4d7da279" containerName="extract-content" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.689991 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="36383cda-5afc-46ac-8cb8-56eb4d7da279" containerName="extract-content" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.690302 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="36383cda-5afc-46ac-8cb8-56eb4d7da279" containerName="registry-server" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.690347 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="c14771d2-0a1a-4734-b551-c11227ee79f9" containerName="registry-server" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.693191 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f2m2h" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.703758 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f2m2h"] Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.770176 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e3440fd-33af-44b1-ab26-1b70f41f2eef-catalog-content\") pod \"redhat-operators-f2m2h\" (UID: \"3e3440fd-33af-44b1-ab26-1b70f41f2eef\") " pod="openshift-marketplace/redhat-operators-f2m2h" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.770351 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e3440fd-33af-44b1-ab26-1b70f41f2eef-utilities\") pod \"redhat-operators-f2m2h\" (UID: \"3e3440fd-33af-44b1-ab26-1b70f41f2eef\") " pod="openshift-marketplace/redhat-operators-f2m2h" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.770444 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgxq8\" (UniqueName: \"kubernetes.io/projected/3e3440fd-33af-44b1-ab26-1b70f41f2eef-kube-api-access-cgxq8\") pod \"redhat-operators-f2m2h\" (UID: \"3e3440fd-33af-44b1-ab26-1b70f41f2eef\") " pod="openshift-marketplace/redhat-operators-f2m2h" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.872749 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e3440fd-33af-44b1-ab26-1b70f41f2eef-catalog-content\") pod \"redhat-operators-f2m2h\" (UID: \"3e3440fd-33af-44b1-ab26-1b70f41f2eef\") " pod="openshift-marketplace/redhat-operators-f2m2h" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.872889 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e3440fd-33af-44b1-ab26-1b70f41f2eef-utilities\") pod \"redhat-operators-f2m2h\" (UID: \"3e3440fd-33af-44b1-ab26-1b70f41f2eef\") " pod="openshift-marketplace/redhat-operators-f2m2h" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.873011 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgxq8\" (UniqueName: \"kubernetes.io/projected/3e3440fd-33af-44b1-ab26-1b70f41f2eef-kube-api-access-cgxq8\") pod \"redhat-operators-f2m2h\" (UID: \"3e3440fd-33af-44b1-ab26-1b70f41f2eef\") " pod="openshift-marketplace/redhat-operators-f2m2h" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.873397 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e3440fd-33af-44b1-ab26-1b70f41f2eef-utilities\") pod \"redhat-operators-f2m2h\" (UID: \"3e3440fd-33af-44b1-ab26-1b70f41f2eef\") " pod="openshift-marketplace/redhat-operators-f2m2h" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.874010 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e3440fd-33af-44b1-ab26-1b70f41f2eef-catalog-content\") pod \"redhat-operators-f2m2h\" (UID: \"3e3440fd-33af-44b1-ab26-1b70f41f2eef\") " pod="openshift-marketplace/redhat-operators-f2m2h" Dec 05 09:54:15 crc kubenswrapper[4997]: I1205 09:54:15.895395 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgxq8\" (UniqueName: \"kubernetes.io/projected/3e3440fd-33af-44b1-ab26-1b70f41f2eef-kube-api-access-cgxq8\") pod \"redhat-operators-f2m2h\" (UID: \"3e3440fd-33af-44b1-ab26-1b70f41f2eef\") " pod="openshift-marketplace/redhat-operators-f2m2h" Dec 05 09:54:16 crc kubenswrapper[4997]: I1205 09:54:16.030898 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f2m2h" Dec 05 09:54:16 crc kubenswrapper[4997]: I1205 09:54:16.501746 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f2m2h"] Dec 05 09:54:17 crc kubenswrapper[4997]: I1205 09:54:17.100311 4997 generic.go:334] "Generic (PLEG): container finished" podID="3e3440fd-33af-44b1-ab26-1b70f41f2eef" containerID="4a8692e4c167765458d919b844fd6edac8134e35da42fc93da0dde3ab2471dd8" exitCode=0 Dec 05 09:54:17 crc kubenswrapper[4997]: I1205 09:54:17.100407 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f2m2h" event={"ID":"3e3440fd-33af-44b1-ab26-1b70f41f2eef","Type":"ContainerDied","Data":"4a8692e4c167765458d919b844fd6edac8134e35da42fc93da0dde3ab2471dd8"} Dec 05 09:54:17 crc kubenswrapper[4997]: I1205 09:54:17.100462 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f2m2h" event={"ID":"3e3440fd-33af-44b1-ab26-1b70f41f2eef","Type":"ContainerStarted","Data":"253490dc95cec05665815e168792f8fabe40af82991612e83ffb8d1be49287dd"} Dec 05 09:54:18 crc kubenswrapper[4997]: I1205 09:54:18.112413 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f2m2h" event={"ID":"3e3440fd-33af-44b1-ab26-1b70f41f2eef","Type":"ContainerStarted","Data":"f5f9060ce6d408f65a1325679eb8d2cf9d8647c8ee5facfc9295ae1979cebaeb"} Dec 05 09:54:19 crc kubenswrapper[4997]: I1205 09:54:19.770489 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:54:19 crc kubenswrapper[4997]: I1205 09:54:19.770810 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:54:22 crc kubenswrapper[4997]: I1205 09:54:22.159789 4997 generic.go:334] "Generic (PLEG): container finished" podID="3e3440fd-33af-44b1-ab26-1b70f41f2eef" containerID="f5f9060ce6d408f65a1325679eb8d2cf9d8647c8ee5facfc9295ae1979cebaeb" exitCode=0 Dec 05 09:54:22 crc kubenswrapper[4997]: I1205 09:54:22.159870 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f2m2h" event={"ID":"3e3440fd-33af-44b1-ab26-1b70f41f2eef","Type":"ContainerDied","Data":"f5f9060ce6d408f65a1325679eb8d2cf9d8647c8ee5facfc9295ae1979cebaeb"} Dec 05 09:54:23 crc kubenswrapper[4997]: I1205 09:54:23.171184 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f2m2h" event={"ID":"3e3440fd-33af-44b1-ab26-1b70f41f2eef","Type":"ContainerStarted","Data":"f7e2fcf49a7c35a2f741ee399c62bd4fd70b4eaad9358d93e31fa53ebef119a4"} Dec 05 09:54:23 crc kubenswrapper[4997]: I1205 09:54:23.195924 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f2m2h" podStartSLOduration=2.746929302 podStartE2EDuration="8.195899232s" podCreationTimestamp="2025-12-05 09:54:15 +0000 UTC" firstStartedPulling="2025-12-05 09:54:17.105579183 +0000 UTC m=+10757.634486444" lastFinishedPulling="2025-12-05 09:54:22.554549103 +0000 UTC m=+10763.083456374" observedRunningTime="2025-12-05 09:54:23.189190172 +0000 UTC m=+10763.718097463" watchObservedRunningTime="2025-12-05 09:54:23.195899232 +0000 UTC m=+10763.724806493" Dec 05 09:54:26 crc kubenswrapper[4997]: I1205 09:54:26.031928 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f2m2h" Dec 05 09:54:26 crc kubenswrapper[4997]: I1205 09:54:26.032462 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f2m2h" Dec 05 09:54:27 crc kubenswrapper[4997]: I1205 09:54:27.088277 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f2m2h" podUID="3e3440fd-33af-44b1-ab26-1b70f41f2eef" containerName="registry-server" probeResult="failure" output=< Dec 05 09:54:27 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 09:54:27 crc kubenswrapper[4997]: > Dec 05 09:54:36 crc kubenswrapper[4997]: I1205 09:54:36.127816 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f2m2h" Dec 05 09:54:36 crc kubenswrapper[4997]: I1205 09:54:36.205563 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f2m2h" Dec 05 09:54:39 crc kubenswrapper[4997]: I1205 09:54:39.721184 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f2m2h"] Dec 05 09:54:39 crc kubenswrapper[4997]: I1205 09:54:39.722113 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f2m2h" podUID="3e3440fd-33af-44b1-ab26-1b70f41f2eef" containerName="registry-server" containerID="cri-o://f7e2fcf49a7c35a2f741ee399c62bd4fd70b4eaad9358d93e31fa53ebef119a4" gracePeriod=2 Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.242441 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f2m2h" Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.330571 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgxq8\" (UniqueName: \"kubernetes.io/projected/3e3440fd-33af-44b1-ab26-1b70f41f2eef-kube-api-access-cgxq8\") pod \"3e3440fd-33af-44b1-ab26-1b70f41f2eef\" (UID: \"3e3440fd-33af-44b1-ab26-1b70f41f2eef\") " Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.330777 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e3440fd-33af-44b1-ab26-1b70f41f2eef-catalog-content\") pod \"3e3440fd-33af-44b1-ab26-1b70f41f2eef\" (UID: \"3e3440fd-33af-44b1-ab26-1b70f41f2eef\") " Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.331156 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e3440fd-33af-44b1-ab26-1b70f41f2eef-utilities\") pod \"3e3440fd-33af-44b1-ab26-1b70f41f2eef\" (UID: \"3e3440fd-33af-44b1-ab26-1b70f41f2eef\") " Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.332933 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e3440fd-33af-44b1-ab26-1b70f41f2eef-utilities" (OuterVolumeSpecName: "utilities") pod "3e3440fd-33af-44b1-ab26-1b70f41f2eef" (UID: "3e3440fd-33af-44b1-ab26-1b70f41f2eef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.342741 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e3440fd-33af-44b1-ab26-1b70f41f2eef-kube-api-access-cgxq8" (OuterVolumeSpecName: "kube-api-access-cgxq8") pod "3e3440fd-33af-44b1-ab26-1b70f41f2eef" (UID: "3e3440fd-33af-44b1-ab26-1b70f41f2eef"). InnerVolumeSpecName "kube-api-access-cgxq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.383105 4997 generic.go:334] "Generic (PLEG): container finished" podID="3e3440fd-33af-44b1-ab26-1b70f41f2eef" containerID="f7e2fcf49a7c35a2f741ee399c62bd4fd70b4eaad9358d93e31fa53ebef119a4" exitCode=0 Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.383203 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f2m2h" Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.383203 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f2m2h" event={"ID":"3e3440fd-33af-44b1-ab26-1b70f41f2eef","Type":"ContainerDied","Data":"f7e2fcf49a7c35a2f741ee399c62bd4fd70b4eaad9358d93e31fa53ebef119a4"} Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.383347 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f2m2h" event={"ID":"3e3440fd-33af-44b1-ab26-1b70f41f2eef","Type":"ContainerDied","Data":"253490dc95cec05665815e168792f8fabe40af82991612e83ffb8d1be49287dd"} Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.383406 4997 scope.go:117] "RemoveContainer" containerID="f7e2fcf49a7c35a2f741ee399c62bd4fd70b4eaad9358d93e31fa53ebef119a4" Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.422432 4997 scope.go:117] "RemoveContainer" containerID="f5f9060ce6d408f65a1325679eb8d2cf9d8647c8ee5facfc9295ae1979cebaeb" Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.434229 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e3440fd-33af-44b1-ab26-1b70f41f2eef-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.434277 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgxq8\" (UniqueName: \"kubernetes.io/projected/3e3440fd-33af-44b1-ab26-1b70f41f2eef-kube-api-access-cgxq8\") on node \"crc\" DevicePath \"\"" Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.456083 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e3440fd-33af-44b1-ab26-1b70f41f2eef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e3440fd-33af-44b1-ab26-1b70f41f2eef" (UID: "3e3440fd-33af-44b1-ab26-1b70f41f2eef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.466592 4997 scope.go:117] "RemoveContainer" containerID="4a8692e4c167765458d919b844fd6edac8134e35da42fc93da0dde3ab2471dd8" Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.522767 4997 scope.go:117] "RemoveContainer" containerID="f7e2fcf49a7c35a2f741ee399c62bd4fd70b4eaad9358d93e31fa53ebef119a4" Dec 05 09:54:40 crc kubenswrapper[4997]: E1205 09:54:40.523398 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7e2fcf49a7c35a2f741ee399c62bd4fd70b4eaad9358d93e31fa53ebef119a4\": container with ID starting with f7e2fcf49a7c35a2f741ee399c62bd4fd70b4eaad9358d93e31fa53ebef119a4 not found: ID does not exist" containerID="f7e2fcf49a7c35a2f741ee399c62bd4fd70b4eaad9358d93e31fa53ebef119a4" Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.523439 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7e2fcf49a7c35a2f741ee399c62bd4fd70b4eaad9358d93e31fa53ebef119a4"} err="failed to get container status \"f7e2fcf49a7c35a2f741ee399c62bd4fd70b4eaad9358d93e31fa53ebef119a4\": rpc error: code = NotFound desc = could not find container \"f7e2fcf49a7c35a2f741ee399c62bd4fd70b4eaad9358d93e31fa53ebef119a4\": container with ID starting with f7e2fcf49a7c35a2f741ee399c62bd4fd70b4eaad9358d93e31fa53ebef119a4 not found: ID does not exist" Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.523482 4997 scope.go:117] "RemoveContainer" containerID="f5f9060ce6d408f65a1325679eb8d2cf9d8647c8ee5facfc9295ae1979cebaeb" Dec 05 09:54:40 crc kubenswrapper[4997]: E1205 09:54:40.523758 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5f9060ce6d408f65a1325679eb8d2cf9d8647c8ee5facfc9295ae1979cebaeb\": container with ID starting with f5f9060ce6d408f65a1325679eb8d2cf9d8647c8ee5facfc9295ae1979cebaeb not found: ID does not exist" containerID="f5f9060ce6d408f65a1325679eb8d2cf9d8647c8ee5facfc9295ae1979cebaeb" Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.523791 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5f9060ce6d408f65a1325679eb8d2cf9d8647c8ee5facfc9295ae1979cebaeb"} err="failed to get container status \"f5f9060ce6d408f65a1325679eb8d2cf9d8647c8ee5facfc9295ae1979cebaeb\": rpc error: code = NotFound desc = could not find container \"f5f9060ce6d408f65a1325679eb8d2cf9d8647c8ee5facfc9295ae1979cebaeb\": container with ID starting with f5f9060ce6d408f65a1325679eb8d2cf9d8647c8ee5facfc9295ae1979cebaeb not found: ID does not exist" Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.523811 4997 scope.go:117] "RemoveContainer" containerID="4a8692e4c167765458d919b844fd6edac8134e35da42fc93da0dde3ab2471dd8" Dec 05 09:54:40 crc kubenswrapper[4997]: E1205 09:54:40.524078 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a8692e4c167765458d919b844fd6edac8134e35da42fc93da0dde3ab2471dd8\": container with ID starting with 4a8692e4c167765458d919b844fd6edac8134e35da42fc93da0dde3ab2471dd8 not found: ID does not exist" containerID="4a8692e4c167765458d919b844fd6edac8134e35da42fc93da0dde3ab2471dd8" Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.524105 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a8692e4c167765458d919b844fd6edac8134e35da42fc93da0dde3ab2471dd8"} err="failed to get container status \"4a8692e4c167765458d919b844fd6edac8134e35da42fc93da0dde3ab2471dd8\": rpc error: code = NotFound desc = could not find container \"4a8692e4c167765458d919b844fd6edac8134e35da42fc93da0dde3ab2471dd8\": container with ID starting with 4a8692e4c167765458d919b844fd6edac8134e35da42fc93da0dde3ab2471dd8 not found: ID does not exist" Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.536725 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e3440fd-33af-44b1-ab26-1b70f41f2eef-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.727042 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f2m2h"] Dec 05 09:54:40 crc kubenswrapper[4997]: I1205 09:54:40.750558 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f2m2h"] Dec 05 09:54:41 crc kubenswrapper[4997]: I1205 09:54:41.765914 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e3440fd-33af-44b1-ab26-1b70f41f2eef" path="/var/lib/kubelet/pods/3e3440fd-33af-44b1-ab26-1b70f41f2eef/volumes" Dec 05 09:54:49 crc kubenswrapper[4997]: I1205 09:54:49.770170 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:54:49 crc kubenswrapper[4997]: I1205 09:54:49.770878 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:55:19 crc kubenswrapper[4997]: I1205 09:55:19.769869 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:55:19 crc kubenswrapper[4997]: I1205 09:55:19.771177 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:55:19 crc kubenswrapper[4997]: I1205 09:55:19.771828 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 09:55:19 crc kubenswrapper[4997]: I1205 09:55:19.773360 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d10217ab4b647ccf0d7a3c346634fc388bc2f808fb062547e0c6bbe74fe6cdee"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:55:19 crc kubenswrapper[4997]: I1205 09:55:19.773476 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://d10217ab4b647ccf0d7a3c346634fc388bc2f808fb062547e0c6bbe74fe6cdee" gracePeriod=600 Dec 05 09:55:20 crc kubenswrapper[4997]: I1205 09:55:20.873329 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="d10217ab4b647ccf0d7a3c346634fc388bc2f808fb062547e0c6bbe74fe6cdee" exitCode=0 Dec 05 09:55:20 crc kubenswrapper[4997]: I1205 09:55:20.873408 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"d10217ab4b647ccf0d7a3c346634fc388bc2f808fb062547e0c6bbe74fe6cdee"} Dec 05 09:55:20 crc kubenswrapper[4997]: I1205 09:55:20.873803 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5"} Dec 05 09:55:20 crc kubenswrapper[4997]: I1205 09:55:20.873836 4997 scope.go:117] "RemoveContainer" containerID="bc23e53cfe6fcc51e13c9ccec03b342da182369c27bb31d13d671714bf6d9a86" Dec 05 09:57:49 crc kubenswrapper[4997]: I1205 09:57:49.770444 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:57:49 crc kubenswrapper[4997]: I1205 09:57:49.771133 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:58:19 crc kubenswrapper[4997]: I1205 09:58:19.770384 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:58:19 crc kubenswrapper[4997]: I1205 09:58:19.771112 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:58:28 crc kubenswrapper[4997]: I1205 09:58:28.210884 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jvwgq"] Dec 05 09:58:28 crc kubenswrapper[4997]: E1205 09:58:28.212026 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e3440fd-33af-44b1-ab26-1b70f41f2eef" containerName="extract-content" Dec 05 09:58:28 crc kubenswrapper[4997]: I1205 09:58:28.212052 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e3440fd-33af-44b1-ab26-1b70f41f2eef" containerName="extract-content" Dec 05 09:58:28 crc kubenswrapper[4997]: E1205 09:58:28.212082 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e3440fd-33af-44b1-ab26-1b70f41f2eef" containerName="registry-server" Dec 05 09:58:28 crc kubenswrapper[4997]: I1205 09:58:28.212091 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e3440fd-33af-44b1-ab26-1b70f41f2eef" containerName="registry-server" Dec 05 09:58:28 crc kubenswrapper[4997]: E1205 09:58:28.212113 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e3440fd-33af-44b1-ab26-1b70f41f2eef" containerName="extract-utilities" Dec 05 09:58:28 crc kubenswrapper[4997]: I1205 09:58:28.212122 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e3440fd-33af-44b1-ab26-1b70f41f2eef" containerName="extract-utilities" Dec 05 09:58:28 crc kubenswrapper[4997]: I1205 09:58:28.212407 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e3440fd-33af-44b1-ab26-1b70f41f2eef" containerName="registry-server" Dec 05 09:58:28 crc kubenswrapper[4997]: I1205 09:58:28.214486 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jvwgq" Dec 05 09:58:28 crc kubenswrapper[4997]: I1205 09:58:28.234520 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jvwgq"] Dec 05 09:58:28 crc kubenswrapper[4997]: I1205 09:58:28.387933 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tssh5\" (UniqueName: \"kubernetes.io/projected/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-kube-api-access-tssh5\") pod \"certified-operators-jvwgq\" (UID: \"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4\") " pod="openshift-marketplace/certified-operators-jvwgq" Dec 05 09:58:28 crc kubenswrapper[4997]: I1205 09:58:28.388070 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-utilities\") pod \"certified-operators-jvwgq\" (UID: \"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4\") " pod="openshift-marketplace/certified-operators-jvwgq" Dec 05 09:58:28 crc kubenswrapper[4997]: I1205 09:58:28.388389 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-catalog-content\") pod \"certified-operators-jvwgq\" (UID: \"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4\") " pod="openshift-marketplace/certified-operators-jvwgq" Dec 05 09:58:28 crc kubenswrapper[4997]: I1205 09:58:28.490841 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tssh5\" (UniqueName: \"kubernetes.io/projected/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-kube-api-access-tssh5\") pod \"certified-operators-jvwgq\" (UID: \"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4\") " pod="openshift-marketplace/certified-operators-jvwgq" Dec 05 09:58:28 crc kubenswrapper[4997]: I1205 09:58:28.490952 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-utilities\") pod \"certified-operators-jvwgq\" (UID: \"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4\") " pod="openshift-marketplace/certified-operators-jvwgq" Dec 05 09:58:28 crc kubenswrapper[4997]: I1205 09:58:28.491050 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-catalog-content\") pod \"certified-operators-jvwgq\" (UID: \"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4\") " pod="openshift-marketplace/certified-operators-jvwgq" Dec 05 09:58:28 crc kubenswrapper[4997]: I1205 09:58:28.491526 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-catalog-content\") pod \"certified-operators-jvwgq\" (UID: \"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4\") " pod="openshift-marketplace/certified-operators-jvwgq" Dec 05 09:58:28 crc kubenswrapper[4997]: I1205 09:58:28.491679 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-utilities\") pod \"certified-operators-jvwgq\" (UID: \"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4\") " pod="openshift-marketplace/certified-operators-jvwgq" Dec 05 09:58:28 crc kubenswrapper[4997]: I1205 09:58:28.518593 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tssh5\" (UniqueName: \"kubernetes.io/projected/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-kube-api-access-tssh5\") pod \"certified-operators-jvwgq\" (UID: \"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4\") " pod="openshift-marketplace/certified-operators-jvwgq" Dec 05 09:58:28 crc kubenswrapper[4997]: I1205 09:58:28.538692 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jvwgq" Dec 05 09:58:29 crc kubenswrapper[4997]: I1205 09:58:29.049919 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jvwgq"] Dec 05 09:58:30 crc kubenswrapper[4997]: I1205 09:58:30.054292 4997 generic.go:334] "Generic (PLEG): container finished" podID="acfdcff2-594a-41ea-b5f4-fb43a0eee8e4" containerID="8181f647eb24a2306cdbb8decda26e61398142995b1f9f363dbc0b477e84089a" exitCode=0 Dec 05 09:58:30 crc kubenswrapper[4997]: I1205 09:58:30.054393 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jvwgq" event={"ID":"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4","Type":"ContainerDied","Data":"8181f647eb24a2306cdbb8decda26e61398142995b1f9f363dbc0b477e84089a"} Dec 05 09:58:30 crc kubenswrapper[4997]: I1205 09:58:30.055086 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jvwgq" event={"ID":"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4","Type":"ContainerStarted","Data":"b2111d05039c3d669bc6208caf0dd8f55a6798ae14a1417d853e69b99c2950ad"} Dec 05 09:58:30 crc kubenswrapper[4997]: I1205 09:58:30.057275 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:58:31 crc kubenswrapper[4997]: I1205 09:58:31.070437 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jvwgq" event={"ID":"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4","Type":"ContainerStarted","Data":"81a34552cb264103eae3917c186cbd431aa3250a1588314a4ee16d6cedbf7ecd"} Dec 05 09:58:32 crc kubenswrapper[4997]: I1205 09:58:32.087416 4997 generic.go:334] "Generic (PLEG): container finished" podID="acfdcff2-594a-41ea-b5f4-fb43a0eee8e4" containerID="81a34552cb264103eae3917c186cbd431aa3250a1588314a4ee16d6cedbf7ecd" exitCode=0 Dec 05 09:58:32 crc kubenswrapper[4997]: I1205 09:58:32.087589 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jvwgq" event={"ID":"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4","Type":"ContainerDied","Data":"81a34552cb264103eae3917c186cbd431aa3250a1588314a4ee16d6cedbf7ecd"} Dec 05 09:58:33 crc kubenswrapper[4997]: I1205 09:58:33.099264 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jvwgq" event={"ID":"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4","Type":"ContainerStarted","Data":"53c1ab36b4350d215669a2fc1c377a480a9253bbad677790d127ee9d1f2b3703"} Dec 05 09:58:33 crc kubenswrapper[4997]: I1205 09:58:33.124765 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jvwgq" podStartSLOduration=2.414795634 podStartE2EDuration="5.12473031s" podCreationTimestamp="2025-12-05 09:58:28 +0000 UTC" firstStartedPulling="2025-12-05 09:58:30.057030201 +0000 UTC m=+11010.585937462" lastFinishedPulling="2025-12-05 09:58:32.766964877 +0000 UTC m=+11013.295872138" observedRunningTime="2025-12-05 09:58:33.115660264 +0000 UTC m=+11013.644567535" watchObservedRunningTime="2025-12-05 09:58:33.12473031 +0000 UTC m=+11013.653637571" Dec 05 09:58:38 crc kubenswrapper[4997]: I1205 09:58:38.540485 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jvwgq" Dec 05 09:58:38 crc kubenswrapper[4997]: I1205 09:58:38.541289 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jvwgq" Dec 05 09:58:38 crc kubenswrapper[4997]: I1205 09:58:38.590602 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jvwgq" Dec 05 09:58:39 crc kubenswrapper[4997]: I1205 09:58:39.246809 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jvwgq" Dec 05 09:58:39 crc kubenswrapper[4997]: I1205 09:58:39.309718 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jvwgq"] Dec 05 09:58:41 crc kubenswrapper[4997]: I1205 09:58:41.196591 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jvwgq" podUID="acfdcff2-594a-41ea-b5f4-fb43a0eee8e4" containerName="registry-server" containerID="cri-o://53c1ab36b4350d215669a2fc1c377a480a9253bbad677790d127ee9d1f2b3703" gracePeriod=2 Dec 05 09:58:41 crc kubenswrapper[4997]: I1205 09:58:41.731330 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jvwgq" Dec 05 09:58:41 crc kubenswrapper[4997]: I1205 09:58:41.888577 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tssh5\" (UniqueName: \"kubernetes.io/projected/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-kube-api-access-tssh5\") pod \"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4\" (UID: \"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4\") " Dec 05 09:58:41 crc kubenswrapper[4997]: I1205 09:58:41.888821 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-catalog-content\") pod \"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4\" (UID: \"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4\") " Dec 05 09:58:41 crc kubenswrapper[4997]: I1205 09:58:41.888935 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-utilities\") pod \"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4\" (UID: \"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4\") " Dec 05 09:58:41 crc kubenswrapper[4997]: I1205 09:58:41.890801 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-utilities" (OuterVolumeSpecName: "utilities") pod "acfdcff2-594a-41ea-b5f4-fb43a0eee8e4" (UID: "acfdcff2-594a-41ea-b5f4-fb43a0eee8e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:58:41 crc kubenswrapper[4997]: I1205 09:58:41.896919 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-kube-api-access-tssh5" (OuterVolumeSpecName: "kube-api-access-tssh5") pod "acfdcff2-594a-41ea-b5f4-fb43a0eee8e4" (UID: "acfdcff2-594a-41ea-b5f4-fb43a0eee8e4"). InnerVolumeSpecName "kube-api-access-tssh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:58:41 crc kubenswrapper[4997]: I1205 09:58:41.991459 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tssh5\" (UniqueName: \"kubernetes.io/projected/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-kube-api-access-tssh5\") on node \"crc\" DevicePath \"\"" Dec 05 09:58:41 crc kubenswrapper[4997]: I1205 09:58:41.991760 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:58:42 crc kubenswrapper[4997]: I1205 09:58:42.047001 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "acfdcff2-594a-41ea-b5f4-fb43a0eee8e4" (UID: "acfdcff2-594a-41ea-b5f4-fb43a0eee8e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:58:42 crc kubenswrapper[4997]: I1205 09:58:42.094180 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:58:42 crc kubenswrapper[4997]: I1205 09:58:42.208214 4997 generic.go:334] "Generic (PLEG): container finished" podID="acfdcff2-594a-41ea-b5f4-fb43a0eee8e4" containerID="53c1ab36b4350d215669a2fc1c377a480a9253bbad677790d127ee9d1f2b3703" exitCode=0 Dec 05 09:58:42 crc kubenswrapper[4997]: I1205 09:58:42.208317 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jvwgq" event={"ID":"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4","Type":"ContainerDied","Data":"53c1ab36b4350d215669a2fc1c377a480a9253bbad677790d127ee9d1f2b3703"} Dec 05 09:58:42 crc kubenswrapper[4997]: I1205 09:58:42.208603 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jvwgq" event={"ID":"acfdcff2-594a-41ea-b5f4-fb43a0eee8e4","Type":"ContainerDied","Data":"b2111d05039c3d669bc6208caf0dd8f55a6798ae14a1417d853e69b99c2950ad"} Dec 05 09:58:42 crc kubenswrapper[4997]: I1205 09:58:42.208350 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jvwgq" Dec 05 09:58:42 crc kubenswrapper[4997]: I1205 09:58:42.208644 4997 scope.go:117] "RemoveContainer" containerID="53c1ab36b4350d215669a2fc1c377a480a9253bbad677790d127ee9d1f2b3703" Dec 05 09:58:42 crc kubenswrapper[4997]: I1205 09:58:42.231228 4997 scope.go:117] "RemoveContainer" containerID="81a34552cb264103eae3917c186cbd431aa3250a1588314a4ee16d6cedbf7ecd" Dec 05 09:58:42 crc kubenswrapper[4997]: I1205 09:58:42.249305 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jvwgq"] Dec 05 09:58:42 crc kubenswrapper[4997]: I1205 09:58:42.258158 4997 scope.go:117] "RemoveContainer" containerID="8181f647eb24a2306cdbb8decda26e61398142995b1f9f363dbc0b477e84089a" Dec 05 09:58:42 crc kubenswrapper[4997]: I1205 09:58:42.260432 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jvwgq"] Dec 05 09:58:42 crc kubenswrapper[4997]: I1205 09:58:42.314745 4997 scope.go:117] "RemoveContainer" containerID="53c1ab36b4350d215669a2fc1c377a480a9253bbad677790d127ee9d1f2b3703" Dec 05 09:58:42 crc kubenswrapper[4997]: E1205 09:58:42.315399 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53c1ab36b4350d215669a2fc1c377a480a9253bbad677790d127ee9d1f2b3703\": container with ID starting with 53c1ab36b4350d215669a2fc1c377a480a9253bbad677790d127ee9d1f2b3703 not found: ID does not exist" containerID="53c1ab36b4350d215669a2fc1c377a480a9253bbad677790d127ee9d1f2b3703" Dec 05 09:58:42 crc kubenswrapper[4997]: I1205 09:58:42.315444 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53c1ab36b4350d215669a2fc1c377a480a9253bbad677790d127ee9d1f2b3703"} err="failed to get container status \"53c1ab36b4350d215669a2fc1c377a480a9253bbad677790d127ee9d1f2b3703\": rpc error: code = NotFound desc = could not find container \"53c1ab36b4350d215669a2fc1c377a480a9253bbad677790d127ee9d1f2b3703\": container with ID starting with 53c1ab36b4350d215669a2fc1c377a480a9253bbad677790d127ee9d1f2b3703 not found: ID does not exist" Dec 05 09:58:42 crc kubenswrapper[4997]: I1205 09:58:42.315488 4997 scope.go:117] "RemoveContainer" containerID="81a34552cb264103eae3917c186cbd431aa3250a1588314a4ee16d6cedbf7ecd" Dec 05 09:58:42 crc kubenswrapper[4997]: E1205 09:58:42.315875 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81a34552cb264103eae3917c186cbd431aa3250a1588314a4ee16d6cedbf7ecd\": container with ID starting with 81a34552cb264103eae3917c186cbd431aa3250a1588314a4ee16d6cedbf7ecd not found: ID does not exist" containerID="81a34552cb264103eae3917c186cbd431aa3250a1588314a4ee16d6cedbf7ecd" Dec 05 09:58:42 crc kubenswrapper[4997]: I1205 09:58:42.315925 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81a34552cb264103eae3917c186cbd431aa3250a1588314a4ee16d6cedbf7ecd"} err="failed to get container status \"81a34552cb264103eae3917c186cbd431aa3250a1588314a4ee16d6cedbf7ecd\": rpc error: code = NotFound desc = could not find container \"81a34552cb264103eae3917c186cbd431aa3250a1588314a4ee16d6cedbf7ecd\": container with ID starting with 81a34552cb264103eae3917c186cbd431aa3250a1588314a4ee16d6cedbf7ecd not found: ID does not exist" Dec 05 09:58:42 crc kubenswrapper[4997]: I1205 09:58:42.315944 4997 scope.go:117] "RemoveContainer" containerID="8181f647eb24a2306cdbb8decda26e61398142995b1f9f363dbc0b477e84089a" Dec 05 09:58:42 crc kubenswrapper[4997]: E1205 09:58:42.316481 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8181f647eb24a2306cdbb8decda26e61398142995b1f9f363dbc0b477e84089a\": container with ID starting with 8181f647eb24a2306cdbb8decda26e61398142995b1f9f363dbc0b477e84089a not found: ID does not exist" containerID="8181f647eb24a2306cdbb8decda26e61398142995b1f9f363dbc0b477e84089a" Dec 05 09:58:42 crc kubenswrapper[4997]: I1205 09:58:42.316531 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8181f647eb24a2306cdbb8decda26e61398142995b1f9f363dbc0b477e84089a"} err="failed to get container status \"8181f647eb24a2306cdbb8decda26e61398142995b1f9f363dbc0b477e84089a\": rpc error: code = NotFound desc = could not find container \"8181f647eb24a2306cdbb8decda26e61398142995b1f9f363dbc0b477e84089a\": container with ID starting with 8181f647eb24a2306cdbb8decda26e61398142995b1f9f363dbc0b477e84089a not found: ID does not exist" Dec 05 09:58:43 crc kubenswrapper[4997]: I1205 09:58:43.764763 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acfdcff2-594a-41ea-b5f4-fb43a0eee8e4" path="/var/lib/kubelet/pods/acfdcff2-594a-41ea-b5f4-fb43a0eee8e4/volumes" Dec 05 09:58:49 crc kubenswrapper[4997]: I1205 09:58:49.770194 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:58:49 crc kubenswrapper[4997]: I1205 09:58:49.770794 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:58:49 crc kubenswrapper[4997]: I1205 09:58:49.771003 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 09:58:49 crc kubenswrapper[4997]: I1205 09:58:49.771948 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:58:49 crc kubenswrapper[4997]: I1205 09:58:49.772302 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" gracePeriod=600 Dec 05 09:58:49 crc kubenswrapper[4997]: E1205 09:58:49.895961 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:58:50 crc kubenswrapper[4997]: I1205 09:58:50.312066 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" exitCode=0 Dec 05 09:58:50 crc kubenswrapper[4997]: I1205 09:58:50.312128 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5"} Dec 05 09:58:50 crc kubenswrapper[4997]: I1205 09:58:50.312539 4997 scope.go:117] "RemoveContainer" containerID="d10217ab4b647ccf0d7a3c346634fc388bc2f808fb062547e0c6bbe74fe6cdee" Dec 05 09:58:50 crc kubenswrapper[4997]: I1205 09:58:50.313507 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 09:58:50 crc kubenswrapper[4997]: E1205 09:58:50.314169 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:59:01 crc kubenswrapper[4997]: I1205 09:59:01.749539 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 09:59:01 crc kubenswrapper[4997]: E1205 09:59:01.750932 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:59:15 crc kubenswrapper[4997]: I1205 09:59:15.750405 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 09:59:15 crc kubenswrapper[4997]: E1205 09:59:15.753724 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:59:30 crc kubenswrapper[4997]: I1205 09:59:30.748974 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 09:59:30 crc kubenswrapper[4997]: E1205 09:59:30.749595 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:59:41 crc kubenswrapper[4997]: I1205 09:59:41.750673 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 09:59:41 crc kubenswrapper[4997]: E1205 09:59:41.752481 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 09:59:56 crc kubenswrapper[4997]: I1205 09:59:56.748513 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 09:59:56 crc kubenswrapper[4997]: E1205 09:59:56.749485 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.165551 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn"] Dec 05 10:00:00 crc kubenswrapper[4997]: E1205 10:00:00.167591 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acfdcff2-594a-41ea-b5f4-fb43a0eee8e4" containerName="registry-server" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.168055 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="acfdcff2-594a-41ea-b5f4-fb43a0eee8e4" containerName="registry-server" Dec 05 10:00:00 crc kubenswrapper[4997]: E1205 10:00:00.168157 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acfdcff2-594a-41ea-b5f4-fb43a0eee8e4" containerName="extract-content" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.168222 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="acfdcff2-594a-41ea-b5f4-fb43a0eee8e4" containerName="extract-content" Dec 05 10:00:00 crc kubenswrapper[4997]: E1205 10:00:00.168291 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acfdcff2-594a-41ea-b5f4-fb43a0eee8e4" containerName="extract-utilities" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.168357 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="acfdcff2-594a-41ea-b5f4-fb43a0eee8e4" containerName="extract-utilities" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.168740 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="acfdcff2-594a-41ea-b5f4-fb43a0eee8e4" containerName="registry-server" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.169845 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.172463 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.174373 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.178820 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn"] Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.204816 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ca62626-28f6-4973-bdfd-102f6d290bd2-config-volume\") pod \"collect-profiles-29415480-8zmcn\" (UID: \"4ca62626-28f6-4973-bdfd-102f6d290bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.205004 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wpn2\" (UniqueName: \"kubernetes.io/projected/4ca62626-28f6-4973-bdfd-102f6d290bd2-kube-api-access-4wpn2\") pod \"collect-profiles-29415480-8zmcn\" (UID: \"4ca62626-28f6-4973-bdfd-102f6d290bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.205037 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ca62626-28f6-4973-bdfd-102f6d290bd2-secret-volume\") pod \"collect-profiles-29415480-8zmcn\" (UID: \"4ca62626-28f6-4973-bdfd-102f6d290bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.306882 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wpn2\" (UniqueName: \"kubernetes.io/projected/4ca62626-28f6-4973-bdfd-102f6d290bd2-kube-api-access-4wpn2\") pod \"collect-profiles-29415480-8zmcn\" (UID: \"4ca62626-28f6-4973-bdfd-102f6d290bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.306945 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ca62626-28f6-4973-bdfd-102f6d290bd2-secret-volume\") pod \"collect-profiles-29415480-8zmcn\" (UID: \"4ca62626-28f6-4973-bdfd-102f6d290bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.307031 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ca62626-28f6-4973-bdfd-102f6d290bd2-config-volume\") pod \"collect-profiles-29415480-8zmcn\" (UID: \"4ca62626-28f6-4973-bdfd-102f6d290bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.307890 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ca62626-28f6-4973-bdfd-102f6d290bd2-config-volume\") pod \"collect-profiles-29415480-8zmcn\" (UID: \"4ca62626-28f6-4973-bdfd-102f6d290bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.312389 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ca62626-28f6-4973-bdfd-102f6d290bd2-secret-volume\") pod \"collect-profiles-29415480-8zmcn\" (UID: \"4ca62626-28f6-4973-bdfd-102f6d290bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.324833 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wpn2\" (UniqueName: \"kubernetes.io/projected/4ca62626-28f6-4973-bdfd-102f6d290bd2-kube-api-access-4wpn2\") pod \"collect-profiles-29415480-8zmcn\" (UID: \"4ca62626-28f6-4973-bdfd-102f6d290bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.504359 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn" Dec 05 10:00:00 crc kubenswrapper[4997]: I1205 10:00:00.957230 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn"] Dec 05 10:00:01 crc kubenswrapper[4997]: I1205 10:00:01.155078 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn" event={"ID":"4ca62626-28f6-4973-bdfd-102f6d290bd2","Type":"ContainerStarted","Data":"fe686f81ee59eb6f54070fcb9d9fb5170b3552fed29ae33185d0754a032de36a"} Dec 05 10:00:01 crc kubenswrapper[4997]: I1205 10:00:01.155392 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn" event={"ID":"4ca62626-28f6-4973-bdfd-102f6d290bd2","Type":"ContainerStarted","Data":"06cd2a5543f712833901e5b4703f3d1825ceb22ab46fb68e69bf1f2f36e35e06"} Dec 05 10:00:02 crc kubenswrapper[4997]: I1205 10:00:02.166802 4997 generic.go:334] "Generic (PLEG): container finished" podID="4ca62626-28f6-4973-bdfd-102f6d290bd2" containerID="fe686f81ee59eb6f54070fcb9d9fb5170b3552fed29ae33185d0754a032de36a" exitCode=0 Dec 05 10:00:02 crc kubenswrapper[4997]: I1205 10:00:02.166906 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn" event={"ID":"4ca62626-28f6-4973-bdfd-102f6d290bd2","Type":"ContainerDied","Data":"fe686f81ee59eb6f54070fcb9d9fb5170b3552fed29ae33185d0754a032de36a"} Dec 05 10:00:03 crc kubenswrapper[4997]: I1205 10:00:03.527864 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn" Dec 05 10:00:03 crc kubenswrapper[4997]: I1205 10:00:03.674766 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ca62626-28f6-4973-bdfd-102f6d290bd2-secret-volume\") pod \"4ca62626-28f6-4973-bdfd-102f6d290bd2\" (UID: \"4ca62626-28f6-4973-bdfd-102f6d290bd2\") " Dec 05 10:00:03 crc kubenswrapper[4997]: I1205 10:00:03.674843 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ca62626-28f6-4973-bdfd-102f6d290bd2-config-volume\") pod \"4ca62626-28f6-4973-bdfd-102f6d290bd2\" (UID: \"4ca62626-28f6-4973-bdfd-102f6d290bd2\") " Dec 05 10:00:03 crc kubenswrapper[4997]: I1205 10:00:03.675157 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wpn2\" (UniqueName: \"kubernetes.io/projected/4ca62626-28f6-4973-bdfd-102f6d290bd2-kube-api-access-4wpn2\") pod \"4ca62626-28f6-4973-bdfd-102f6d290bd2\" (UID: \"4ca62626-28f6-4973-bdfd-102f6d290bd2\") " Dec 05 10:00:03 crc kubenswrapper[4997]: I1205 10:00:03.675574 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ca62626-28f6-4973-bdfd-102f6d290bd2-config-volume" (OuterVolumeSpecName: "config-volume") pod "4ca62626-28f6-4973-bdfd-102f6d290bd2" (UID: "4ca62626-28f6-4973-bdfd-102f6d290bd2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:00:03 crc kubenswrapper[4997]: I1205 10:00:03.676489 4997 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ca62626-28f6-4973-bdfd-102f6d290bd2-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 10:00:03 crc kubenswrapper[4997]: I1205 10:00:03.681132 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ca62626-28f6-4973-bdfd-102f6d290bd2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4ca62626-28f6-4973-bdfd-102f6d290bd2" (UID: "4ca62626-28f6-4973-bdfd-102f6d290bd2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:00:03 crc kubenswrapper[4997]: I1205 10:00:03.681556 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ca62626-28f6-4973-bdfd-102f6d290bd2-kube-api-access-4wpn2" (OuterVolumeSpecName: "kube-api-access-4wpn2") pod "4ca62626-28f6-4973-bdfd-102f6d290bd2" (UID: "4ca62626-28f6-4973-bdfd-102f6d290bd2"). InnerVolumeSpecName "kube-api-access-4wpn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:00:03 crc kubenswrapper[4997]: I1205 10:00:03.778448 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wpn2\" (UniqueName: \"kubernetes.io/projected/4ca62626-28f6-4973-bdfd-102f6d290bd2-kube-api-access-4wpn2\") on node \"crc\" DevicePath \"\"" Dec 05 10:00:03 crc kubenswrapper[4997]: I1205 10:00:03.778485 4997 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ca62626-28f6-4973-bdfd-102f6d290bd2-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 10:00:04 crc kubenswrapper[4997]: I1205 10:00:04.187671 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn" event={"ID":"4ca62626-28f6-4973-bdfd-102f6d290bd2","Type":"ContainerDied","Data":"06cd2a5543f712833901e5b4703f3d1825ceb22ab46fb68e69bf1f2f36e35e06"} Dec 05 10:00:04 crc kubenswrapper[4997]: I1205 10:00:04.188033 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06cd2a5543f712833901e5b4703f3d1825ceb22ab46fb68e69bf1f2f36e35e06" Dec 05 10:00:04 crc kubenswrapper[4997]: I1205 10:00:04.188090 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-8zmcn" Dec 05 10:00:04 crc kubenswrapper[4997]: I1205 10:00:04.612499 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr"] Dec 05 10:00:04 crc kubenswrapper[4997]: I1205 10:00:04.624912 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415435-5nqrr"] Dec 05 10:00:05 crc kubenswrapper[4997]: I1205 10:00:05.761278 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1" path="/var/lib/kubelet/pods/f57dfa0d-1ca4-4b36-b2d5-91ab5be616a1/volumes" Dec 05 10:00:10 crc kubenswrapper[4997]: I1205 10:00:10.749881 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:00:10 crc kubenswrapper[4997]: E1205 10:00:10.751051 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:00:21 crc kubenswrapper[4997]: I1205 10:00:21.246771 4997 scope.go:117] "RemoveContainer" containerID="46f6bdd664b048dd2d1854cdcdbc93d413ce0f5a3f1e90c9f0de703378393c59" Dec 05 10:00:23 crc kubenswrapper[4997]: I1205 10:00:23.750278 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:00:23 crc kubenswrapper[4997]: E1205 10:00:23.751306 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:00:37 crc kubenswrapper[4997]: I1205 10:00:37.749011 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:00:37 crc kubenswrapper[4997]: E1205 10:00:37.749882 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:00:48 crc kubenswrapper[4997]: I1205 10:00:48.749644 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:00:48 crc kubenswrapper[4997]: E1205 10:00:48.750782 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.161976 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29415481-fnjnr"] Dec 05 10:01:00 crc kubenswrapper[4997]: E1205 10:01:00.164713 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ca62626-28f6-4973-bdfd-102f6d290bd2" containerName="collect-profiles" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.164856 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca62626-28f6-4973-bdfd-102f6d290bd2" containerName="collect-profiles" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.165249 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ca62626-28f6-4973-bdfd-102f6d290bd2" containerName="collect-profiles" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.166674 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415481-fnjnr" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.184495 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29415481-fnjnr"] Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.231475 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-fernet-keys\") pod \"keystone-cron-29415481-fnjnr\" (UID: \"08b4289a-1b86-4057-860a-4f5b4bd157b1\") " pod="openstack/keystone-cron-29415481-fnjnr" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.231924 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs48r\" (UniqueName: \"kubernetes.io/projected/08b4289a-1b86-4057-860a-4f5b4bd157b1-kube-api-access-zs48r\") pod \"keystone-cron-29415481-fnjnr\" (UID: \"08b4289a-1b86-4057-860a-4f5b4bd157b1\") " pod="openstack/keystone-cron-29415481-fnjnr" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.232017 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-config-data\") pod \"keystone-cron-29415481-fnjnr\" (UID: \"08b4289a-1b86-4057-860a-4f5b4bd157b1\") " pod="openstack/keystone-cron-29415481-fnjnr" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.232073 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-combined-ca-bundle\") pod \"keystone-cron-29415481-fnjnr\" (UID: \"08b4289a-1b86-4057-860a-4f5b4bd157b1\") " pod="openstack/keystone-cron-29415481-fnjnr" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.333486 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-fernet-keys\") pod \"keystone-cron-29415481-fnjnr\" (UID: \"08b4289a-1b86-4057-860a-4f5b4bd157b1\") " pod="openstack/keystone-cron-29415481-fnjnr" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.333657 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs48r\" (UniqueName: \"kubernetes.io/projected/08b4289a-1b86-4057-860a-4f5b4bd157b1-kube-api-access-zs48r\") pod \"keystone-cron-29415481-fnjnr\" (UID: \"08b4289a-1b86-4057-860a-4f5b4bd157b1\") " pod="openstack/keystone-cron-29415481-fnjnr" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.333730 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-config-data\") pod \"keystone-cron-29415481-fnjnr\" (UID: \"08b4289a-1b86-4057-860a-4f5b4bd157b1\") " pod="openstack/keystone-cron-29415481-fnjnr" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.333766 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-combined-ca-bundle\") pod \"keystone-cron-29415481-fnjnr\" (UID: \"08b4289a-1b86-4057-860a-4f5b4bd157b1\") " pod="openstack/keystone-cron-29415481-fnjnr" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.340134 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-combined-ca-bundle\") pod \"keystone-cron-29415481-fnjnr\" (UID: \"08b4289a-1b86-4057-860a-4f5b4bd157b1\") " pod="openstack/keystone-cron-29415481-fnjnr" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.343445 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-fernet-keys\") pod \"keystone-cron-29415481-fnjnr\" (UID: \"08b4289a-1b86-4057-860a-4f5b4bd157b1\") " pod="openstack/keystone-cron-29415481-fnjnr" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.350210 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-config-data\") pod \"keystone-cron-29415481-fnjnr\" (UID: \"08b4289a-1b86-4057-860a-4f5b4bd157b1\") " pod="openstack/keystone-cron-29415481-fnjnr" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.350344 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs48r\" (UniqueName: \"kubernetes.io/projected/08b4289a-1b86-4057-860a-4f5b4bd157b1-kube-api-access-zs48r\") pod \"keystone-cron-29415481-fnjnr\" (UID: \"08b4289a-1b86-4057-860a-4f5b4bd157b1\") " pod="openstack/keystone-cron-29415481-fnjnr" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.495432 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415481-fnjnr" Dec 05 10:01:00 crc kubenswrapper[4997]: I1205 10:01:00.955158 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29415481-fnjnr"] Dec 05 10:01:01 crc kubenswrapper[4997]: I1205 10:01:01.894650 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415481-fnjnr" event={"ID":"08b4289a-1b86-4057-860a-4f5b4bd157b1","Type":"ContainerStarted","Data":"d08fabc10f5931aca1cb235ba6dd83a4a7b63ef0f31e5569d89706f9a66a6513"} Dec 05 10:01:01 crc kubenswrapper[4997]: I1205 10:01:01.894951 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415481-fnjnr" event={"ID":"08b4289a-1b86-4057-860a-4f5b4bd157b1","Type":"ContainerStarted","Data":"3dfb73c3e995fd4c2f99d4d3c4f4e425fed20a574f4265005934bf202cba3fc0"} Dec 05 10:01:01 crc kubenswrapper[4997]: I1205 10:01:01.909979 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29415481-fnjnr" podStartSLOduration=1.90996298 podStartE2EDuration="1.90996298s" podCreationTimestamp="2025-12-05 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:01:01.908555132 +0000 UTC m=+11162.437462403" watchObservedRunningTime="2025-12-05 10:01:01.90996298 +0000 UTC m=+11162.438870241" Dec 05 10:01:02 crc kubenswrapper[4997]: I1205 10:01:02.749666 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:01:02 crc kubenswrapper[4997]: E1205 10:01:02.750012 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:01:03 crc kubenswrapper[4997]: I1205 10:01:03.912877 4997 generic.go:334] "Generic (PLEG): container finished" podID="08b4289a-1b86-4057-860a-4f5b4bd157b1" containerID="d08fabc10f5931aca1cb235ba6dd83a4a7b63ef0f31e5569d89706f9a66a6513" exitCode=0 Dec 05 10:01:03 crc kubenswrapper[4997]: I1205 10:01:03.912973 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415481-fnjnr" event={"ID":"08b4289a-1b86-4057-860a-4f5b4bd157b1","Type":"ContainerDied","Data":"d08fabc10f5931aca1cb235ba6dd83a4a7b63ef0f31e5569d89706f9a66a6513"} Dec 05 10:01:05 crc kubenswrapper[4997]: I1205 10:01:05.350112 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415481-fnjnr" Dec 05 10:01:05 crc kubenswrapper[4997]: I1205 10:01:05.452635 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zs48r\" (UniqueName: \"kubernetes.io/projected/08b4289a-1b86-4057-860a-4f5b4bd157b1-kube-api-access-zs48r\") pod \"08b4289a-1b86-4057-860a-4f5b4bd157b1\" (UID: \"08b4289a-1b86-4057-860a-4f5b4bd157b1\") " Dec 05 10:01:05 crc kubenswrapper[4997]: I1205 10:01:05.452774 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-fernet-keys\") pod \"08b4289a-1b86-4057-860a-4f5b4bd157b1\" (UID: \"08b4289a-1b86-4057-860a-4f5b4bd157b1\") " Dec 05 10:01:05 crc kubenswrapper[4997]: I1205 10:01:05.452805 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-config-data\") pod \"08b4289a-1b86-4057-860a-4f5b4bd157b1\" (UID: \"08b4289a-1b86-4057-860a-4f5b4bd157b1\") " Dec 05 10:01:05 crc kubenswrapper[4997]: I1205 10:01:05.453006 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-combined-ca-bundle\") pod \"08b4289a-1b86-4057-860a-4f5b4bd157b1\" (UID: \"08b4289a-1b86-4057-860a-4f5b4bd157b1\") " Dec 05 10:01:05 crc kubenswrapper[4997]: I1205 10:01:05.460004 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08b4289a-1b86-4057-860a-4f5b4bd157b1-kube-api-access-zs48r" (OuterVolumeSpecName: "kube-api-access-zs48r") pod "08b4289a-1b86-4057-860a-4f5b4bd157b1" (UID: "08b4289a-1b86-4057-860a-4f5b4bd157b1"). InnerVolumeSpecName "kube-api-access-zs48r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:01:05 crc kubenswrapper[4997]: I1205 10:01:05.468931 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "08b4289a-1b86-4057-860a-4f5b4bd157b1" (UID: "08b4289a-1b86-4057-860a-4f5b4bd157b1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:01:05 crc kubenswrapper[4997]: I1205 10:01:05.494067 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08b4289a-1b86-4057-860a-4f5b4bd157b1" (UID: "08b4289a-1b86-4057-860a-4f5b4bd157b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:01:05 crc kubenswrapper[4997]: I1205 10:01:05.520676 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-config-data" (OuterVolumeSpecName: "config-data") pod "08b4289a-1b86-4057-860a-4f5b4bd157b1" (UID: "08b4289a-1b86-4057-860a-4f5b4bd157b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:01:05 crc kubenswrapper[4997]: I1205 10:01:05.556246 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zs48r\" (UniqueName: \"kubernetes.io/projected/08b4289a-1b86-4057-860a-4f5b4bd157b1-kube-api-access-zs48r\") on node \"crc\" DevicePath \"\"" Dec 05 10:01:05 crc kubenswrapper[4997]: I1205 10:01:05.556488 4997 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 10:01:05 crc kubenswrapper[4997]: I1205 10:01:05.556558 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 10:01:05 crc kubenswrapper[4997]: I1205 10:01:05.556643 4997 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b4289a-1b86-4057-860a-4f5b4bd157b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 10:01:05 crc kubenswrapper[4997]: I1205 10:01:05.935597 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415481-fnjnr" event={"ID":"08b4289a-1b86-4057-860a-4f5b4bd157b1","Type":"ContainerDied","Data":"3dfb73c3e995fd4c2f99d4d3c4f4e425fed20a574f4265005934bf202cba3fc0"} Dec 05 10:01:05 crc kubenswrapper[4997]: I1205 10:01:05.935683 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415481-fnjnr" Dec 05 10:01:05 crc kubenswrapper[4997]: I1205 10:01:05.935697 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3dfb73c3e995fd4c2f99d4d3c4f4e425fed20a574f4265005934bf202cba3fc0" Dec 05 10:01:15 crc kubenswrapper[4997]: I1205 10:01:15.749694 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:01:15 crc kubenswrapper[4997]: E1205 10:01:15.750916 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:01:27 crc kubenswrapper[4997]: I1205 10:01:27.749374 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:01:27 crc kubenswrapper[4997]: E1205 10:01:27.750206 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:01:38 crc kubenswrapper[4997]: I1205 10:01:38.749094 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:01:38 crc kubenswrapper[4997]: E1205 10:01:38.749910 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:01:50 crc kubenswrapper[4997]: I1205 10:01:50.346033 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w72lv"] Dec 05 10:01:50 crc kubenswrapper[4997]: E1205 10:01:50.348332 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08b4289a-1b86-4057-860a-4f5b4bd157b1" containerName="keystone-cron" Dec 05 10:01:50 crc kubenswrapper[4997]: I1205 10:01:50.348367 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b4289a-1b86-4057-860a-4f5b4bd157b1" containerName="keystone-cron" Dec 05 10:01:50 crc kubenswrapper[4997]: I1205 10:01:50.348606 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="08b4289a-1b86-4057-860a-4f5b4bd157b1" containerName="keystone-cron" Dec 05 10:01:50 crc kubenswrapper[4997]: I1205 10:01:50.350386 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w72lv" Dec 05 10:01:50 crc kubenswrapper[4997]: I1205 10:01:50.372078 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a112e0f3-589f-434c-ac2b-a5510b0b7659-utilities\") pod \"redhat-marketplace-w72lv\" (UID: \"a112e0f3-589f-434c-ac2b-a5510b0b7659\") " pod="openshift-marketplace/redhat-marketplace-w72lv" Dec 05 10:01:50 crc kubenswrapper[4997]: I1205 10:01:50.372168 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a112e0f3-589f-434c-ac2b-a5510b0b7659-catalog-content\") pod \"redhat-marketplace-w72lv\" (UID: \"a112e0f3-589f-434c-ac2b-a5510b0b7659\") " pod="openshift-marketplace/redhat-marketplace-w72lv" Dec 05 10:01:50 crc kubenswrapper[4997]: I1205 10:01:50.372234 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49thk\" (UniqueName: \"kubernetes.io/projected/a112e0f3-589f-434c-ac2b-a5510b0b7659-kube-api-access-49thk\") pod \"redhat-marketplace-w72lv\" (UID: \"a112e0f3-589f-434c-ac2b-a5510b0b7659\") " pod="openshift-marketplace/redhat-marketplace-w72lv" Dec 05 10:01:50 crc kubenswrapper[4997]: I1205 10:01:50.395293 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w72lv"] Dec 05 10:01:50 crc kubenswrapper[4997]: I1205 10:01:50.476473 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a112e0f3-589f-434c-ac2b-a5510b0b7659-utilities\") pod \"redhat-marketplace-w72lv\" (UID: \"a112e0f3-589f-434c-ac2b-a5510b0b7659\") " pod="openshift-marketplace/redhat-marketplace-w72lv" Dec 05 10:01:50 crc kubenswrapper[4997]: I1205 10:01:50.476567 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a112e0f3-589f-434c-ac2b-a5510b0b7659-catalog-content\") pod \"redhat-marketplace-w72lv\" (UID: \"a112e0f3-589f-434c-ac2b-a5510b0b7659\") " pod="openshift-marketplace/redhat-marketplace-w72lv" Dec 05 10:01:50 crc kubenswrapper[4997]: I1205 10:01:50.476648 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49thk\" (UniqueName: \"kubernetes.io/projected/a112e0f3-589f-434c-ac2b-a5510b0b7659-kube-api-access-49thk\") pod \"redhat-marketplace-w72lv\" (UID: \"a112e0f3-589f-434c-ac2b-a5510b0b7659\") " pod="openshift-marketplace/redhat-marketplace-w72lv" Dec 05 10:01:50 crc kubenswrapper[4997]: I1205 10:01:50.477521 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a112e0f3-589f-434c-ac2b-a5510b0b7659-utilities\") pod \"redhat-marketplace-w72lv\" (UID: \"a112e0f3-589f-434c-ac2b-a5510b0b7659\") " pod="openshift-marketplace/redhat-marketplace-w72lv" Dec 05 10:01:50 crc kubenswrapper[4997]: I1205 10:01:50.480763 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a112e0f3-589f-434c-ac2b-a5510b0b7659-catalog-content\") pod \"redhat-marketplace-w72lv\" (UID: \"a112e0f3-589f-434c-ac2b-a5510b0b7659\") " pod="openshift-marketplace/redhat-marketplace-w72lv" Dec 05 10:01:50 crc kubenswrapper[4997]: I1205 10:01:50.506677 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49thk\" (UniqueName: \"kubernetes.io/projected/a112e0f3-589f-434c-ac2b-a5510b0b7659-kube-api-access-49thk\") pod \"redhat-marketplace-w72lv\" (UID: \"a112e0f3-589f-434c-ac2b-a5510b0b7659\") " pod="openshift-marketplace/redhat-marketplace-w72lv" Dec 05 10:01:50 crc kubenswrapper[4997]: I1205 10:01:50.677410 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w72lv" Dec 05 10:01:51 crc kubenswrapper[4997]: I1205 10:01:51.146501 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w72lv"] Dec 05 10:01:51 crc kubenswrapper[4997]: I1205 10:01:51.715961 4997 generic.go:334] "Generic (PLEG): container finished" podID="a112e0f3-589f-434c-ac2b-a5510b0b7659" containerID="c4aa62473db0d16f025f8818f1098e2f7501d3b0faffef20298256c6227f31e2" exitCode=0 Dec 05 10:01:51 crc kubenswrapper[4997]: I1205 10:01:51.716086 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w72lv" event={"ID":"a112e0f3-589f-434c-ac2b-a5510b0b7659","Type":"ContainerDied","Data":"c4aa62473db0d16f025f8818f1098e2f7501d3b0faffef20298256c6227f31e2"} Dec 05 10:01:51 crc kubenswrapper[4997]: I1205 10:01:51.716245 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w72lv" event={"ID":"a112e0f3-589f-434c-ac2b-a5510b0b7659","Type":"ContainerStarted","Data":"99a972a629cfb6ab5a3988b25a720caf7922d641b01dd3d19bb2219663a27322"} Dec 05 10:01:52 crc kubenswrapper[4997]: I1205 10:01:52.729996 4997 generic.go:334] "Generic (PLEG): container finished" podID="a112e0f3-589f-434c-ac2b-a5510b0b7659" containerID="9a3fa0ab43759214dc2beb1ed29ec69502a7ad0a57dc4896c2f88b4f2bccb7d8" exitCode=0 Dec 05 10:01:52 crc kubenswrapper[4997]: I1205 10:01:52.730049 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w72lv" event={"ID":"a112e0f3-589f-434c-ac2b-a5510b0b7659","Type":"ContainerDied","Data":"9a3fa0ab43759214dc2beb1ed29ec69502a7ad0a57dc4896c2f88b4f2bccb7d8"} Dec 05 10:01:52 crc kubenswrapper[4997]: I1205 10:01:52.753173 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:01:52 crc kubenswrapper[4997]: E1205 10:01:52.753470 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:01:53 crc kubenswrapper[4997]: I1205 10:01:53.745008 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w72lv" event={"ID":"a112e0f3-589f-434c-ac2b-a5510b0b7659","Type":"ContainerStarted","Data":"e8aad41438bb0cb7f6a5aa629e75ae82851685c93a41c19dd0f643e0d58a037f"} Dec 05 10:01:53 crc kubenswrapper[4997]: I1205 10:01:53.769992 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w72lv" podStartSLOduration=2.354541498 podStartE2EDuration="3.769969426s" podCreationTimestamp="2025-12-05 10:01:50 +0000 UTC" firstStartedPulling="2025-12-05 10:01:51.719280972 +0000 UTC m=+11212.248188253" lastFinishedPulling="2025-12-05 10:01:53.13470892 +0000 UTC m=+11213.663616181" observedRunningTime="2025-12-05 10:01:53.764749694 +0000 UTC m=+11214.293656955" watchObservedRunningTime="2025-12-05 10:01:53.769969426 +0000 UTC m=+11214.298876687" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.626071 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.628719 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.631442 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.631508 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.631648 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-qsjdm" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.632568 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.642542 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.783582 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8688d151-9d21-4107-86cd-f9cee40d08b8-config-data\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.783688 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8688d151-9d21-4107-86cd-f9cee40d08b8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.783764 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn9zx\" (UniqueName: \"kubernetes.io/projected/8688d151-9d21-4107-86cd-f9cee40d08b8-kube-api-access-zn9zx\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.783804 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.783879 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.784337 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8688d151-9d21-4107-86cd-f9cee40d08b8-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.784608 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.784704 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.784735 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8688d151-9d21-4107-86cd-f9cee40d08b8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.886424 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.886494 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.886519 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8688d151-9d21-4107-86cd-f9cee40d08b8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.886573 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8688d151-9d21-4107-86cd-f9cee40d08b8-config-data\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.886607 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8688d151-9d21-4107-86cd-f9cee40d08b8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.886671 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn9zx\" (UniqueName: \"kubernetes.io/projected/8688d151-9d21-4107-86cd-f9cee40d08b8-kube-api-access-zn9zx\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.886886 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.886964 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.887068 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8688d151-9d21-4107-86cd-f9cee40d08b8-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.887835 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8688d151-9d21-4107-86cd-f9cee40d08b8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.887980 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.888121 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8688d151-9d21-4107-86cd-f9cee40d08b8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.889443 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8688d151-9d21-4107-86cd-f9cee40d08b8-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.889607 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.892890 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.900475 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8688d151-9d21-4107-86cd-f9cee40d08b8-config-data\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.901210 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.901343 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.901539 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.910271 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn9zx\" (UniqueName: \"kubernetes.io/projected/8688d151-9d21-4107-86cd-f9cee40d08b8-kube-api-access-zn9zx\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.920367 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " pod="openstack/tempest-tests-tempest" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.957862 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-qsjdm" Dec 05 10:01:59 crc kubenswrapper[4997]: I1205 10:01:59.966441 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 05 10:02:00 crc kubenswrapper[4997]: I1205 10:02:00.436137 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 05 10:02:00 crc kubenswrapper[4997]: I1205 10:02:00.678007 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w72lv" Dec 05 10:02:00 crc kubenswrapper[4997]: I1205 10:02:00.678070 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w72lv" Dec 05 10:02:00 crc kubenswrapper[4997]: I1205 10:02:00.748385 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w72lv" Dec 05 10:02:00 crc kubenswrapper[4997]: I1205 10:02:00.823637 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8688d151-9d21-4107-86cd-f9cee40d08b8","Type":"ContainerStarted","Data":"9617e0c85ed24cd7886c6da10dd3b69b803d3f3cd30ecd4a3085d87a5323fb87"} Dec 05 10:02:00 crc kubenswrapper[4997]: I1205 10:02:00.895861 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w72lv" Dec 05 10:02:01 crc kubenswrapper[4997]: I1205 10:02:01.002401 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w72lv"] Dec 05 10:02:02 crc kubenswrapper[4997]: I1205 10:02:02.840589 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w72lv" podUID="a112e0f3-589f-434c-ac2b-a5510b0b7659" containerName="registry-server" containerID="cri-o://e8aad41438bb0cb7f6a5aa629e75ae82851685c93a41c19dd0f643e0d58a037f" gracePeriod=2 Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.452471 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w72lv" Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.563998 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a112e0f3-589f-434c-ac2b-a5510b0b7659-catalog-content\") pod \"a112e0f3-589f-434c-ac2b-a5510b0b7659\" (UID: \"a112e0f3-589f-434c-ac2b-a5510b0b7659\") " Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.564100 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49thk\" (UniqueName: \"kubernetes.io/projected/a112e0f3-589f-434c-ac2b-a5510b0b7659-kube-api-access-49thk\") pod \"a112e0f3-589f-434c-ac2b-a5510b0b7659\" (UID: \"a112e0f3-589f-434c-ac2b-a5510b0b7659\") " Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.564142 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a112e0f3-589f-434c-ac2b-a5510b0b7659-utilities\") pod \"a112e0f3-589f-434c-ac2b-a5510b0b7659\" (UID: \"a112e0f3-589f-434c-ac2b-a5510b0b7659\") " Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.565663 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a112e0f3-589f-434c-ac2b-a5510b0b7659-utilities" (OuterVolumeSpecName: "utilities") pod "a112e0f3-589f-434c-ac2b-a5510b0b7659" (UID: "a112e0f3-589f-434c-ac2b-a5510b0b7659"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.575023 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a112e0f3-589f-434c-ac2b-a5510b0b7659-kube-api-access-49thk" (OuterVolumeSpecName: "kube-api-access-49thk") pod "a112e0f3-589f-434c-ac2b-a5510b0b7659" (UID: "a112e0f3-589f-434c-ac2b-a5510b0b7659"). InnerVolumeSpecName "kube-api-access-49thk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.593830 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a112e0f3-589f-434c-ac2b-a5510b0b7659-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a112e0f3-589f-434c-ac2b-a5510b0b7659" (UID: "a112e0f3-589f-434c-ac2b-a5510b0b7659"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.667014 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a112e0f3-589f-434c-ac2b-a5510b0b7659-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.667362 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49thk\" (UniqueName: \"kubernetes.io/projected/a112e0f3-589f-434c-ac2b-a5510b0b7659-kube-api-access-49thk\") on node \"crc\" DevicePath \"\"" Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.667375 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a112e0f3-589f-434c-ac2b-a5510b0b7659-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.870132 4997 generic.go:334] "Generic (PLEG): container finished" podID="a112e0f3-589f-434c-ac2b-a5510b0b7659" containerID="e8aad41438bb0cb7f6a5aa629e75ae82851685c93a41c19dd0f643e0d58a037f" exitCode=0 Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.870177 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w72lv" event={"ID":"a112e0f3-589f-434c-ac2b-a5510b0b7659","Type":"ContainerDied","Data":"e8aad41438bb0cb7f6a5aa629e75ae82851685c93a41c19dd0f643e0d58a037f"} Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.870210 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w72lv" event={"ID":"a112e0f3-589f-434c-ac2b-a5510b0b7659","Type":"ContainerDied","Data":"99a972a629cfb6ab5a3988b25a720caf7922d641b01dd3d19bb2219663a27322"} Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.870234 4997 scope.go:117] "RemoveContainer" containerID="e8aad41438bb0cb7f6a5aa629e75ae82851685c93a41c19dd0f643e0d58a037f" Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.870229 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w72lv" Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.905715 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w72lv"] Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.906971 4997 scope.go:117] "RemoveContainer" containerID="9a3fa0ab43759214dc2beb1ed29ec69502a7ad0a57dc4896c2f88b4f2bccb7d8" Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.923284 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w72lv"] Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.933013 4997 scope.go:117] "RemoveContainer" containerID="c4aa62473db0d16f025f8818f1098e2f7501d3b0faffef20298256c6227f31e2" Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.995909 4997 scope.go:117] "RemoveContainer" containerID="e8aad41438bb0cb7f6a5aa629e75ae82851685c93a41c19dd0f643e0d58a037f" Dec 05 10:02:03 crc kubenswrapper[4997]: E1205 10:02:03.996508 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8aad41438bb0cb7f6a5aa629e75ae82851685c93a41c19dd0f643e0d58a037f\": container with ID starting with e8aad41438bb0cb7f6a5aa629e75ae82851685c93a41c19dd0f643e0d58a037f not found: ID does not exist" containerID="e8aad41438bb0cb7f6a5aa629e75ae82851685c93a41c19dd0f643e0d58a037f" Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.996576 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8aad41438bb0cb7f6a5aa629e75ae82851685c93a41c19dd0f643e0d58a037f"} err="failed to get container status \"e8aad41438bb0cb7f6a5aa629e75ae82851685c93a41c19dd0f643e0d58a037f\": rpc error: code = NotFound desc = could not find container \"e8aad41438bb0cb7f6a5aa629e75ae82851685c93a41c19dd0f643e0d58a037f\": container with ID starting with e8aad41438bb0cb7f6a5aa629e75ae82851685c93a41c19dd0f643e0d58a037f not found: ID does not exist" Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.996608 4997 scope.go:117] "RemoveContainer" containerID="9a3fa0ab43759214dc2beb1ed29ec69502a7ad0a57dc4896c2f88b4f2bccb7d8" Dec 05 10:02:03 crc kubenswrapper[4997]: E1205 10:02:03.998298 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a3fa0ab43759214dc2beb1ed29ec69502a7ad0a57dc4896c2f88b4f2bccb7d8\": container with ID starting with 9a3fa0ab43759214dc2beb1ed29ec69502a7ad0a57dc4896c2f88b4f2bccb7d8 not found: ID does not exist" containerID="9a3fa0ab43759214dc2beb1ed29ec69502a7ad0a57dc4896c2f88b4f2bccb7d8" Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.998364 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a3fa0ab43759214dc2beb1ed29ec69502a7ad0a57dc4896c2f88b4f2bccb7d8"} err="failed to get container status \"9a3fa0ab43759214dc2beb1ed29ec69502a7ad0a57dc4896c2f88b4f2bccb7d8\": rpc error: code = NotFound desc = could not find container \"9a3fa0ab43759214dc2beb1ed29ec69502a7ad0a57dc4896c2f88b4f2bccb7d8\": container with ID starting with 9a3fa0ab43759214dc2beb1ed29ec69502a7ad0a57dc4896c2f88b4f2bccb7d8 not found: ID does not exist" Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.998397 4997 scope.go:117] "RemoveContainer" containerID="c4aa62473db0d16f025f8818f1098e2f7501d3b0faffef20298256c6227f31e2" Dec 05 10:02:03 crc kubenswrapper[4997]: E1205 10:02:03.999196 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4aa62473db0d16f025f8818f1098e2f7501d3b0faffef20298256c6227f31e2\": container with ID starting with c4aa62473db0d16f025f8818f1098e2f7501d3b0faffef20298256c6227f31e2 not found: ID does not exist" containerID="c4aa62473db0d16f025f8818f1098e2f7501d3b0faffef20298256c6227f31e2" Dec 05 10:02:03 crc kubenswrapper[4997]: I1205 10:02:03.999323 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4aa62473db0d16f025f8818f1098e2f7501d3b0faffef20298256c6227f31e2"} err="failed to get container status \"c4aa62473db0d16f025f8818f1098e2f7501d3b0faffef20298256c6227f31e2\": rpc error: code = NotFound desc = could not find container \"c4aa62473db0d16f025f8818f1098e2f7501d3b0faffef20298256c6227f31e2\": container with ID starting with c4aa62473db0d16f025f8818f1098e2f7501d3b0faffef20298256c6227f31e2 not found: ID does not exist" Dec 05 10:02:04 crc kubenswrapper[4997]: I1205 10:02:04.750241 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:02:04 crc kubenswrapper[4997]: E1205 10:02:04.750890 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:02:05 crc kubenswrapper[4997]: I1205 10:02:05.766821 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a112e0f3-589f-434c-ac2b-a5510b0b7659" path="/var/lib/kubelet/pods/a112e0f3-589f-434c-ac2b-a5510b0b7659/volumes" Dec 05 10:02:07 crc kubenswrapper[4997]: I1205 10:02:07.025248 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pttpb"] Dec 05 10:02:07 crc kubenswrapper[4997]: E1205 10:02:07.028112 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a112e0f3-589f-434c-ac2b-a5510b0b7659" containerName="extract-utilities" Dec 05 10:02:07 crc kubenswrapper[4997]: I1205 10:02:07.028160 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a112e0f3-589f-434c-ac2b-a5510b0b7659" containerName="extract-utilities" Dec 05 10:02:07 crc kubenswrapper[4997]: E1205 10:02:07.028220 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a112e0f3-589f-434c-ac2b-a5510b0b7659" containerName="extract-content" Dec 05 10:02:07 crc kubenswrapper[4997]: I1205 10:02:07.028230 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a112e0f3-589f-434c-ac2b-a5510b0b7659" containerName="extract-content" Dec 05 10:02:07 crc kubenswrapper[4997]: E1205 10:02:07.028248 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a112e0f3-589f-434c-ac2b-a5510b0b7659" containerName="registry-server" Dec 05 10:02:07 crc kubenswrapper[4997]: I1205 10:02:07.028254 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a112e0f3-589f-434c-ac2b-a5510b0b7659" containerName="registry-server" Dec 05 10:02:07 crc kubenswrapper[4997]: I1205 10:02:07.029036 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a112e0f3-589f-434c-ac2b-a5510b0b7659" containerName="registry-server" Dec 05 10:02:07 crc kubenswrapper[4997]: I1205 10:02:07.034511 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pttpb" Dec 05 10:02:07 crc kubenswrapper[4997]: I1205 10:02:07.055051 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pttpb"] Dec 05 10:02:07 crc kubenswrapper[4997]: I1205 10:02:07.172879 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxhkc\" (UniqueName: \"kubernetes.io/projected/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-kube-api-access-qxhkc\") pod \"community-operators-pttpb\" (UID: \"7b8c2fba-f0ae-4bb1-92ce-6145d918c386\") " pod="openshift-marketplace/community-operators-pttpb" Dec 05 10:02:07 crc kubenswrapper[4997]: I1205 10:02:07.173049 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-catalog-content\") pod \"community-operators-pttpb\" (UID: \"7b8c2fba-f0ae-4bb1-92ce-6145d918c386\") " pod="openshift-marketplace/community-operators-pttpb" Dec 05 10:02:07 crc kubenswrapper[4997]: I1205 10:02:07.173112 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-utilities\") pod \"community-operators-pttpb\" (UID: \"7b8c2fba-f0ae-4bb1-92ce-6145d918c386\") " pod="openshift-marketplace/community-operators-pttpb" Dec 05 10:02:07 crc kubenswrapper[4997]: I1205 10:02:07.276122 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-catalog-content\") pod \"community-operators-pttpb\" (UID: \"7b8c2fba-f0ae-4bb1-92ce-6145d918c386\") " pod="openshift-marketplace/community-operators-pttpb" Dec 05 10:02:07 crc kubenswrapper[4997]: I1205 10:02:07.276248 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-utilities\") pod \"community-operators-pttpb\" (UID: \"7b8c2fba-f0ae-4bb1-92ce-6145d918c386\") " pod="openshift-marketplace/community-operators-pttpb" Dec 05 10:02:07 crc kubenswrapper[4997]: I1205 10:02:07.276518 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxhkc\" (UniqueName: \"kubernetes.io/projected/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-kube-api-access-qxhkc\") pod \"community-operators-pttpb\" (UID: \"7b8c2fba-f0ae-4bb1-92ce-6145d918c386\") " pod="openshift-marketplace/community-operators-pttpb" Dec 05 10:02:07 crc kubenswrapper[4997]: I1205 10:02:07.276939 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-catalog-content\") pod \"community-operators-pttpb\" (UID: \"7b8c2fba-f0ae-4bb1-92ce-6145d918c386\") " pod="openshift-marketplace/community-operators-pttpb" Dec 05 10:02:07 crc kubenswrapper[4997]: I1205 10:02:07.277334 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-utilities\") pod \"community-operators-pttpb\" (UID: \"7b8c2fba-f0ae-4bb1-92ce-6145d918c386\") " pod="openshift-marketplace/community-operators-pttpb" Dec 05 10:02:07 crc kubenswrapper[4997]: I1205 10:02:07.311815 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxhkc\" (UniqueName: \"kubernetes.io/projected/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-kube-api-access-qxhkc\") pod \"community-operators-pttpb\" (UID: \"7b8c2fba-f0ae-4bb1-92ce-6145d918c386\") " pod="openshift-marketplace/community-operators-pttpb" Dec 05 10:02:07 crc kubenswrapper[4997]: I1205 10:02:07.373006 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pttpb" Dec 05 10:02:13 crc kubenswrapper[4997]: I1205 10:02:13.572432 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pttpb"] Dec 05 10:02:14 crc kubenswrapper[4997]: I1205 10:02:14.040839 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pttpb" event={"ID":"7b8c2fba-f0ae-4bb1-92ce-6145d918c386","Type":"ContainerDied","Data":"4f65ea53e1ce3bd53cb1dd396f4e1ead8c0ab28bf21ea8b1b03072d8964c1847"} Dec 05 10:02:14 crc kubenswrapper[4997]: I1205 10:02:14.040764 4997 generic.go:334] "Generic (PLEG): container finished" podID="7b8c2fba-f0ae-4bb1-92ce-6145d918c386" containerID="4f65ea53e1ce3bd53cb1dd396f4e1ead8c0ab28bf21ea8b1b03072d8964c1847" exitCode=0 Dec 05 10:02:14 crc kubenswrapper[4997]: I1205 10:02:14.041384 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pttpb" event={"ID":"7b8c2fba-f0ae-4bb1-92ce-6145d918c386","Type":"ContainerStarted","Data":"bedbc1758ab6ae8f3d819d4fb041f0249dda407d359788061e7084a5d41e6900"} Dec 05 10:02:15 crc kubenswrapper[4997]: I1205 10:02:15.064029 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pttpb" event={"ID":"7b8c2fba-f0ae-4bb1-92ce-6145d918c386","Type":"ContainerStarted","Data":"86626590a7fbc31b03515b7e5cd423ec5f65560fbf0711f7789169ac65707ffd"} Dec 05 10:02:16 crc kubenswrapper[4997]: I1205 10:02:16.092490 4997 generic.go:334] "Generic (PLEG): container finished" podID="7b8c2fba-f0ae-4bb1-92ce-6145d918c386" containerID="86626590a7fbc31b03515b7e5cd423ec5f65560fbf0711f7789169ac65707ffd" exitCode=0 Dec 05 10:02:16 crc kubenswrapper[4997]: I1205 10:02:16.092561 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pttpb" event={"ID":"7b8c2fba-f0ae-4bb1-92ce-6145d918c386","Type":"ContainerDied","Data":"86626590a7fbc31b03515b7e5cd423ec5f65560fbf0711f7789169ac65707ffd"} Dec 05 10:02:16 crc kubenswrapper[4997]: I1205 10:02:16.753668 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:02:16 crc kubenswrapper[4997]: E1205 10:02:16.754723 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:02:18 crc kubenswrapper[4997]: I1205 10:02:18.122928 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pttpb" event={"ID":"7b8c2fba-f0ae-4bb1-92ce-6145d918c386","Type":"ContainerStarted","Data":"22ff3a476c6b7462546f2d3fc3771772a9782f2b810f366254c8135b9fda772f"} Dec 05 10:02:18 crc kubenswrapper[4997]: I1205 10:02:18.164915 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pttpb" podStartSLOduration=9.121073643 podStartE2EDuration="12.164884667s" podCreationTimestamp="2025-12-05 10:02:06 +0000 UTC" firstStartedPulling="2025-12-05 10:02:14.043205411 +0000 UTC m=+11234.572112682" lastFinishedPulling="2025-12-05 10:02:17.087016445 +0000 UTC m=+11237.615923706" observedRunningTime="2025-12-05 10:02:18.151397442 +0000 UTC m=+11238.680304733" watchObservedRunningTime="2025-12-05 10:02:18.164884667 +0000 UTC m=+11238.693791928" Dec 05 10:02:27 crc kubenswrapper[4997]: I1205 10:02:27.375290 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pttpb" Dec 05 10:02:27 crc kubenswrapper[4997]: I1205 10:02:27.375855 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pttpb" Dec 05 10:02:27 crc kubenswrapper[4997]: I1205 10:02:27.446440 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pttpb" Dec 05 10:02:28 crc kubenswrapper[4997]: I1205 10:02:28.319980 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pttpb" Dec 05 10:02:28 crc kubenswrapper[4997]: I1205 10:02:28.385420 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pttpb"] Dec 05 10:02:29 crc kubenswrapper[4997]: I1205 10:02:29.758554 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:02:29 crc kubenswrapper[4997]: E1205 10:02:29.760079 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:02:30 crc kubenswrapper[4997]: I1205 10:02:30.310192 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pttpb" podUID="7b8c2fba-f0ae-4bb1-92ce-6145d918c386" containerName="registry-server" containerID="cri-o://22ff3a476c6b7462546f2d3fc3771772a9782f2b810f366254c8135b9fda772f" gracePeriod=2 Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.045999 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pttpb" Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.141399 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-utilities\") pod \"7b8c2fba-f0ae-4bb1-92ce-6145d918c386\" (UID: \"7b8c2fba-f0ae-4bb1-92ce-6145d918c386\") " Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.141497 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-catalog-content\") pod \"7b8c2fba-f0ae-4bb1-92ce-6145d918c386\" (UID: \"7b8c2fba-f0ae-4bb1-92ce-6145d918c386\") " Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.141725 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxhkc\" (UniqueName: \"kubernetes.io/projected/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-kube-api-access-qxhkc\") pod \"7b8c2fba-f0ae-4bb1-92ce-6145d918c386\" (UID: \"7b8c2fba-f0ae-4bb1-92ce-6145d918c386\") " Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.142342 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-utilities" (OuterVolumeSpecName: "utilities") pod "7b8c2fba-f0ae-4bb1-92ce-6145d918c386" (UID: "7b8c2fba-f0ae-4bb1-92ce-6145d918c386"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.150705 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-kube-api-access-qxhkc" (OuterVolumeSpecName: "kube-api-access-qxhkc") pod "7b8c2fba-f0ae-4bb1-92ce-6145d918c386" (UID: "7b8c2fba-f0ae-4bb1-92ce-6145d918c386"). InnerVolumeSpecName "kube-api-access-qxhkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.214918 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b8c2fba-f0ae-4bb1-92ce-6145d918c386" (UID: "7b8c2fba-f0ae-4bb1-92ce-6145d918c386"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.246361 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.246733 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.246760 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxhkc\" (UniqueName: \"kubernetes.io/projected/7b8c2fba-f0ae-4bb1-92ce-6145d918c386-kube-api-access-qxhkc\") on node \"crc\" DevicePath \"\"" Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.328600 4997 generic.go:334] "Generic (PLEG): container finished" podID="7b8c2fba-f0ae-4bb1-92ce-6145d918c386" containerID="22ff3a476c6b7462546f2d3fc3771772a9782f2b810f366254c8135b9fda772f" exitCode=0 Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.328673 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pttpb" event={"ID":"7b8c2fba-f0ae-4bb1-92ce-6145d918c386","Type":"ContainerDied","Data":"22ff3a476c6b7462546f2d3fc3771772a9782f2b810f366254c8135b9fda772f"} Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.328815 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pttpb" event={"ID":"7b8c2fba-f0ae-4bb1-92ce-6145d918c386","Type":"ContainerDied","Data":"bedbc1758ab6ae8f3d819d4fb041f0249dda407d359788061e7084a5d41e6900"} Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.328838 4997 scope.go:117] "RemoveContainer" containerID="22ff3a476c6b7462546f2d3fc3771772a9782f2b810f366254c8135b9fda772f" Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.328880 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pttpb" Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.376103 4997 scope.go:117] "RemoveContainer" containerID="86626590a7fbc31b03515b7e5cd423ec5f65560fbf0711f7789169ac65707ffd" Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.376641 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pttpb"] Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.394253 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pttpb"] Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.421593 4997 scope.go:117] "RemoveContainer" containerID="4f65ea53e1ce3bd53cb1dd396f4e1ead8c0ab28bf21ea8b1b03072d8964c1847" Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.471432 4997 scope.go:117] "RemoveContainer" containerID="22ff3a476c6b7462546f2d3fc3771772a9782f2b810f366254c8135b9fda772f" Dec 05 10:02:31 crc kubenswrapper[4997]: E1205 10:02:31.472222 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22ff3a476c6b7462546f2d3fc3771772a9782f2b810f366254c8135b9fda772f\": container with ID starting with 22ff3a476c6b7462546f2d3fc3771772a9782f2b810f366254c8135b9fda772f not found: ID does not exist" containerID="22ff3a476c6b7462546f2d3fc3771772a9782f2b810f366254c8135b9fda772f" Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.472312 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22ff3a476c6b7462546f2d3fc3771772a9782f2b810f366254c8135b9fda772f"} err="failed to get container status \"22ff3a476c6b7462546f2d3fc3771772a9782f2b810f366254c8135b9fda772f\": rpc error: code = NotFound desc = could not find container \"22ff3a476c6b7462546f2d3fc3771772a9782f2b810f366254c8135b9fda772f\": container with ID starting with 22ff3a476c6b7462546f2d3fc3771772a9782f2b810f366254c8135b9fda772f not found: ID does not exist" Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.472360 4997 scope.go:117] "RemoveContainer" containerID="86626590a7fbc31b03515b7e5cd423ec5f65560fbf0711f7789169ac65707ffd" Dec 05 10:02:31 crc kubenswrapper[4997]: E1205 10:02:31.472918 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86626590a7fbc31b03515b7e5cd423ec5f65560fbf0711f7789169ac65707ffd\": container with ID starting with 86626590a7fbc31b03515b7e5cd423ec5f65560fbf0711f7789169ac65707ffd not found: ID does not exist" containerID="86626590a7fbc31b03515b7e5cd423ec5f65560fbf0711f7789169ac65707ffd" Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.472992 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86626590a7fbc31b03515b7e5cd423ec5f65560fbf0711f7789169ac65707ffd"} err="failed to get container status \"86626590a7fbc31b03515b7e5cd423ec5f65560fbf0711f7789169ac65707ffd\": rpc error: code = NotFound desc = could not find container \"86626590a7fbc31b03515b7e5cd423ec5f65560fbf0711f7789169ac65707ffd\": container with ID starting with 86626590a7fbc31b03515b7e5cd423ec5f65560fbf0711f7789169ac65707ffd not found: ID does not exist" Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.473033 4997 scope.go:117] "RemoveContainer" containerID="4f65ea53e1ce3bd53cb1dd396f4e1ead8c0ab28bf21ea8b1b03072d8964c1847" Dec 05 10:02:31 crc kubenswrapper[4997]: E1205 10:02:31.473320 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f65ea53e1ce3bd53cb1dd396f4e1ead8c0ab28bf21ea8b1b03072d8964c1847\": container with ID starting with 4f65ea53e1ce3bd53cb1dd396f4e1ead8c0ab28bf21ea8b1b03072d8964c1847 not found: ID does not exist" containerID="4f65ea53e1ce3bd53cb1dd396f4e1ead8c0ab28bf21ea8b1b03072d8964c1847" Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.473353 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f65ea53e1ce3bd53cb1dd396f4e1ead8c0ab28bf21ea8b1b03072d8964c1847"} err="failed to get container status \"4f65ea53e1ce3bd53cb1dd396f4e1ead8c0ab28bf21ea8b1b03072d8964c1847\": rpc error: code = NotFound desc = could not find container \"4f65ea53e1ce3bd53cb1dd396f4e1ead8c0ab28bf21ea8b1b03072d8964c1847\": container with ID starting with 4f65ea53e1ce3bd53cb1dd396f4e1ead8c0ab28bf21ea8b1b03072d8964c1847 not found: ID does not exist" Dec 05 10:02:31 crc kubenswrapper[4997]: I1205 10:02:31.764020 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b8c2fba-f0ae-4bb1-92ce-6145d918c386" path="/var/lib/kubelet/pods/7b8c2fba-f0ae-4bb1-92ce-6145d918c386/volumes" Dec 05 10:02:44 crc kubenswrapper[4997]: I1205 10:02:44.749946 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:02:44 crc kubenswrapper[4997]: E1205 10:02:44.750743 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:02:55 crc kubenswrapper[4997]: I1205 10:02:55.750784 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:02:55 crc kubenswrapper[4997]: E1205 10:02:55.751443 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:02:57 crc kubenswrapper[4997]: E1205 10:02:57.762997 4997 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:65066e8ca260a75886ae57f157049605" Dec 05 10:02:57 crc kubenswrapper[4997]: E1205 10:02:57.763493 4997 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:65066e8ca260a75886ae57f157049605" Dec 05 10:02:57 crc kubenswrapper[4997]: E1205 10:02:57.763649 4997 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:65066e8ca260a75886ae57f157049605,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zn9zx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(8688d151-9d21-4107-86cd-f9cee40d08b8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 10:02:57 crc kubenswrapper[4997]: E1205 10:02:57.765089 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="8688d151-9d21-4107-86cd-f9cee40d08b8" Dec 05 10:02:58 crc kubenswrapper[4997]: E1205 10:02:58.677718 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:65066e8ca260a75886ae57f157049605\\\"\"" pod="openstack/tempest-tests-tempest" podUID="8688d151-9d21-4107-86cd-f9cee40d08b8" Dec 05 10:03:08 crc kubenswrapper[4997]: I1205 10:03:08.750378 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:03:08 crc kubenswrapper[4997]: E1205 10:03:08.751795 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:03:12 crc kubenswrapper[4997]: I1205 10:03:12.980821 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 05 10:03:14 crc kubenswrapper[4997]: I1205 10:03:14.867080 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8688d151-9d21-4107-86cd-f9cee40d08b8","Type":"ContainerStarted","Data":"936834a8577f4301dd95bb05bf37471800043ac7dedb18f11e7f25a128566bad"} Dec 05 10:03:14 crc kubenswrapper[4997]: I1205 10:03:14.897402 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.363327899 podStartE2EDuration="1m16.897377417s" podCreationTimestamp="2025-12-05 10:01:58 +0000 UTC" firstStartedPulling="2025-12-05 10:02:00.443681678 +0000 UTC m=+11220.972588939" lastFinishedPulling="2025-12-05 10:03:12.977731196 +0000 UTC m=+11293.506638457" observedRunningTime="2025-12-05 10:03:14.886666677 +0000 UTC m=+11295.415573958" watchObservedRunningTime="2025-12-05 10:03:14.897377417 +0000 UTC m=+11295.426284678" Dec 05 10:03:22 crc kubenswrapper[4997]: I1205 10:03:22.749765 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:03:22 crc kubenswrapper[4997]: E1205 10:03:22.750783 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:03:37 crc kubenswrapper[4997]: I1205 10:03:37.749818 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:03:37 crc kubenswrapper[4997]: E1205 10:03:37.750669 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:03:49 crc kubenswrapper[4997]: I1205 10:03:49.757257 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:03:49 crc kubenswrapper[4997]: E1205 10:03:49.758168 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:04:01 crc kubenswrapper[4997]: I1205 10:04:01.749035 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:04:02 crc kubenswrapper[4997]: I1205 10:04:02.463546 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"9a58315196354db15ade60cdec84e4c423cf75d724ed764e78dac051b6d4a462"} Dec 05 10:04:32 crc kubenswrapper[4997]: I1205 10:04:32.465374 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f5zf6"] Dec 05 10:04:32 crc kubenswrapper[4997]: E1205 10:04:32.467097 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b8c2fba-f0ae-4bb1-92ce-6145d918c386" containerName="extract-content" Dec 05 10:04:32 crc kubenswrapper[4997]: I1205 10:04:32.467118 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b8c2fba-f0ae-4bb1-92ce-6145d918c386" containerName="extract-content" Dec 05 10:04:32 crc kubenswrapper[4997]: E1205 10:04:32.467143 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b8c2fba-f0ae-4bb1-92ce-6145d918c386" containerName="registry-server" Dec 05 10:04:32 crc kubenswrapper[4997]: I1205 10:04:32.467149 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b8c2fba-f0ae-4bb1-92ce-6145d918c386" containerName="registry-server" Dec 05 10:04:32 crc kubenswrapper[4997]: E1205 10:04:32.467208 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b8c2fba-f0ae-4bb1-92ce-6145d918c386" containerName="extract-utilities" Dec 05 10:04:32 crc kubenswrapper[4997]: I1205 10:04:32.467215 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b8c2fba-f0ae-4bb1-92ce-6145d918c386" containerName="extract-utilities" Dec 05 10:04:32 crc kubenswrapper[4997]: I1205 10:04:32.467522 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b8c2fba-f0ae-4bb1-92ce-6145d918c386" containerName="registry-server" Dec 05 10:04:32 crc kubenswrapper[4997]: I1205 10:04:32.476310 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f5zf6" Dec 05 10:04:32 crc kubenswrapper[4997]: I1205 10:04:32.487995 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f5zf6"] Dec 05 10:04:32 crc kubenswrapper[4997]: I1205 10:04:32.525426 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1d14bd5-3277-4bd5-a623-a899113f63c5-utilities\") pod \"redhat-operators-f5zf6\" (UID: \"c1d14bd5-3277-4bd5-a623-a899113f63c5\") " pod="openshift-marketplace/redhat-operators-f5zf6" Dec 05 10:04:32 crc kubenswrapper[4997]: I1205 10:04:32.525501 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1d14bd5-3277-4bd5-a623-a899113f63c5-catalog-content\") pod \"redhat-operators-f5zf6\" (UID: \"c1d14bd5-3277-4bd5-a623-a899113f63c5\") " pod="openshift-marketplace/redhat-operators-f5zf6" Dec 05 10:04:32 crc kubenswrapper[4997]: I1205 10:04:32.525573 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvwdw\" (UniqueName: \"kubernetes.io/projected/c1d14bd5-3277-4bd5-a623-a899113f63c5-kube-api-access-pvwdw\") pod \"redhat-operators-f5zf6\" (UID: \"c1d14bd5-3277-4bd5-a623-a899113f63c5\") " pod="openshift-marketplace/redhat-operators-f5zf6" Dec 05 10:04:32 crc kubenswrapper[4997]: I1205 10:04:32.627717 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvwdw\" (UniqueName: \"kubernetes.io/projected/c1d14bd5-3277-4bd5-a623-a899113f63c5-kube-api-access-pvwdw\") pod \"redhat-operators-f5zf6\" (UID: \"c1d14bd5-3277-4bd5-a623-a899113f63c5\") " pod="openshift-marketplace/redhat-operators-f5zf6" Dec 05 10:04:32 crc kubenswrapper[4997]: I1205 10:04:32.627910 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1d14bd5-3277-4bd5-a623-a899113f63c5-utilities\") pod \"redhat-operators-f5zf6\" (UID: \"c1d14bd5-3277-4bd5-a623-a899113f63c5\") " pod="openshift-marketplace/redhat-operators-f5zf6" Dec 05 10:04:32 crc kubenswrapper[4997]: I1205 10:04:32.627953 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1d14bd5-3277-4bd5-a623-a899113f63c5-catalog-content\") pod \"redhat-operators-f5zf6\" (UID: \"c1d14bd5-3277-4bd5-a623-a899113f63c5\") " pod="openshift-marketplace/redhat-operators-f5zf6" Dec 05 10:04:32 crc kubenswrapper[4997]: I1205 10:04:32.628939 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1d14bd5-3277-4bd5-a623-a899113f63c5-utilities\") pod \"redhat-operators-f5zf6\" (UID: \"c1d14bd5-3277-4bd5-a623-a899113f63c5\") " pod="openshift-marketplace/redhat-operators-f5zf6" Dec 05 10:04:32 crc kubenswrapper[4997]: I1205 10:04:32.634794 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1d14bd5-3277-4bd5-a623-a899113f63c5-catalog-content\") pod \"redhat-operators-f5zf6\" (UID: \"c1d14bd5-3277-4bd5-a623-a899113f63c5\") " pod="openshift-marketplace/redhat-operators-f5zf6" Dec 05 10:04:32 crc kubenswrapper[4997]: I1205 10:04:32.665494 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvwdw\" (UniqueName: \"kubernetes.io/projected/c1d14bd5-3277-4bd5-a623-a899113f63c5-kube-api-access-pvwdw\") pod \"redhat-operators-f5zf6\" (UID: \"c1d14bd5-3277-4bd5-a623-a899113f63c5\") " pod="openshift-marketplace/redhat-operators-f5zf6" Dec 05 10:04:32 crc kubenswrapper[4997]: I1205 10:04:32.820978 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f5zf6" Dec 05 10:04:33 crc kubenswrapper[4997]: I1205 10:04:33.605572 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f5zf6"] Dec 05 10:04:33 crc kubenswrapper[4997]: I1205 10:04:33.877658 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5zf6" event={"ID":"c1d14bd5-3277-4bd5-a623-a899113f63c5","Type":"ContainerStarted","Data":"2204fc9d7525fcfd0d5fccfbeac344d852f46d981ea5cce7ad5100d9d959ab55"} Dec 05 10:04:34 crc kubenswrapper[4997]: I1205 10:04:34.892000 4997 generic.go:334] "Generic (PLEG): container finished" podID="c1d14bd5-3277-4bd5-a623-a899113f63c5" containerID="93f7378ae293d06aa074a1c52337451ce5afce2e7deeb19639e80677b4b8cbe1" exitCode=0 Dec 05 10:04:34 crc kubenswrapper[4997]: I1205 10:04:34.893805 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5zf6" event={"ID":"c1d14bd5-3277-4bd5-a623-a899113f63c5","Type":"ContainerDied","Data":"93f7378ae293d06aa074a1c52337451ce5afce2e7deeb19639e80677b4b8cbe1"} Dec 05 10:04:34 crc kubenswrapper[4997]: I1205 10:04:34.897551 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:04:35 crc kubenswrapper[4997]: I1205 10:04:35.906365 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5zf6" event={"ID":"c1d14bd5-3277-4bd5-a623-a899113f63c5","Type":"ContainerStarted","Data":"19057ada048ebdcaf45377f1715685dd7c5a6f0b4e5f9adaafb95a84bf9a8e34"} Dec 05 10:04:39 crc kubenswrapper[4997]: I1205 10:04:39.960347 4997 generic.go:334] "Generic (PLEG): container finished" podID="c1d14bd5-3277-4bd5-a623-a899113f63c5" containerID="19057ada048ebdcaf45377f1715685dd7c5a6f0b4e5f9adaafb95a84bf9a8e34" exitCode=0 Dec 05 10:04:39 crc kubenswrapper[4997]: I1205 10:04:39.960439 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5zf6" event={"ID":"c1d14bd5-3277-4bd5-a623-a899113f63c5","Type":"ContainerDied","Data":"19057ada048ebdcaf45377f1715685dd7c5a6f0b4e5f9adaafb95a84bf9a8e34"} Dec 05 10:04:40 crc kubenswrapper[4997]: I1205 10:04:40.974883 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5zf6" event={"ID":"c1d14bd5-3277-4bd5-a623-a899113f63c5","Type":"ContainerStarted","Data":"6d947d303bff605be2173e6663955d3c51f7852c80212addbb5c4362668a709d"} Dec 05 10:04:41 crc kubenswrapper[4997]: I1205 10:04:41.003265 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f5zf6" podStartSLOduration=3.395361514 podStartE2EDuration="9.00323529s" podCreationTimestamp="2025-12-05 10:04:32 +0000 UTC" firstStartedPulling="2025-12-05 10:04:34.896180107 +0000 UTC m=+11375.425087368" lastFinishedPulling="2025-12-05 10:04:40.504053883 +0000 UTC m=+11381.032961144" observedRunningTime="2025-12-05 10:04:40.993899617 +0000 UTC m=+11381.522806898" watchObservedRunningTime="2025-12-05 10:04:41.00323529 +0000 UTC m=+11381.532142561" Dec 05 10:04:42 crc kubenswrapper[4997]: I1205 10:04:42.821356 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f5zf6" Dec 05 10:04:42 crc kubenswrapper[4997]: I1205 10:04:42.821966 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f5zf6" Dec 05 10:04:43 crc kubenswrapper[4997]: I1205 10:04:43.888339 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f5zf6" podUID="c1d14bd5-3277-4bd5-a623-a899113f63c5" containerName="registry-server" probeResult="failure" output=< Dec 05 10:04:43 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 10:04:43 crc kubenswrapper[4997]: > Dec 05 10:04:52 crc kubenswrapper[4997]: I1205 10:04:52.877056 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f5zf6" Dec 05 10:04:52 crc kubenswrapper[4997]: I1205 10:04:52.941304 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f5zf6" Dec 05 10:04:53 crc kubenswrapper[4997]: I1205 10:04:53.127017 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f5zf6"] Dec 05 10:04:54 crc kubenswrapper[4997]: I1205 10:04:54.119606 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f5zf6" podUID="c1d14bd5-3277-4bd5-a623-a899113f63c5" containerName="registry-server" containerID="cri-o://6d947d303bff605be2173e6663955d3c51f7852c80212addbb5c4362668a709d" gracePeriod=2 Dec 05 10:04:54 crc kubenswrapper[4997]: I1205 10:04:54.813128 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f5zf6" Dec 05 10:04:54 crc kubenswrapper[4997]: I1205 10:04:54.912305 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1d14bd5-3277-4bd5-a623-a899113f63c5-catalog-content\") pod \"c1d14bd5-3277-4bd5-a623-a899113f63c5\" (UID: \"c1d14bd5-3277-4bd5-a623-a899113f63c5\") " Dec 05 10:04:54 crc kubenswrapper[4997]: I1205 10:04:54.912466 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1d14bd5-3277-4bd5-a623-a899113f63c5-utilities\") pod \"c1d14bd5-3277-4bd5-a623-a899113f63c5\" (UID: \"c1d14bd5-3277-4bd5-a623-a899113f63c5\") " Dec 05 10:04:54 crc kubenswrapper[4997]: I1205 10:04:54.912626 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvwdw\" (UniqueName: \"kubernetes.io/projected/c1d14bd5-3277-4bd5-a623-a899113f63c5-kube-api-access-pvwdw\") pod \"c1d14bd5-3277-4bd5-a623-a899113f63c5\" (UID: \"c1d14bd5-3277-4bd5-a623-a899113f63c5\") " Dec 05 10:04:54 crc kubenswrapper[4997]: I1205 10:04:54.916812 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1d14bd5-3277-4bd5-a623-a899113f63c5-utilities" (OuterVolumeSpecName: "utilities") pod "c1d14bd5-3277-4bd5-a623-a899113f63c5" (UID: "c1d14bd5-3277-4bd5-a623-a899113f63c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:04:54 crc kubenswrapper[4997]: I1205 10:04:54.930900 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1d14bd5-3277-4bd5-a623-a899113f63c5-kube-api-access-pvwdw" (OuterVolumeSpecName: "kube-api-access-pvwdw") pod "c1d14bd5-3277-4bd5-a623-a899113f63c5" (UID: "c1d14bd5-3277-4bd5-a623-a899113f63c5"). InnerVolumeSpecName "kube-api-access-pvwdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.016466 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvwdw\" (UniqueName: \"kubernetes.io/projected/c1d14bd5-3277-4bd5-a623-a899113f63c5-kube-api-access-pvwdw\") on node \"crc\" DevicePath \"\"" Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.016508 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1d14bd5-3277-4bd5-a623-a899113f63c5-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.039753 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1d14bd5-3277-4bd5-a623-a899113f63c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c1d14bd5-3277-4bd5-a623-a899113f63c5" (UID: "c1d14bd5-3277-4bd5-a623-a899113f63c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.117924 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1d14bd5-3277-4bd5-a623-a899113f63c5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.132387 4997 generic.go:334] "Generic (PLEG): container finished" podID="c1d14bd5-3277-4bd5-a623-a899113f63c5" containerID="6d947d303bff605be2173e6663955d3c51f7852c80212addbb5c4362668a709d" exitCode=0 Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.132434 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f5zf6" Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.132441 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5zf6" event={"ID":"c1d14bd5-3277-4bd5-a623-a899113f63c5","Type":"ContainerDied","Data":"6d947d303bff605be2173e6663955d3c51f7852c80212addbb5c4362668a709d"} Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.132484 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5zf6" event={"ID":"c1d14bd5-3277-4bd5-a623-a899113f63c5","Type":"ContainerDied","Data":"2204fc9d7525fcfd0d5fccfbeac344d852f46d981ea5cce7ad5100d9d959ab55"} Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.132506 4997 scope.go:117] "RemoveContainer" containerID="6d947d303bff605be2173e6663955d3c51f7852c80212addbb5c4362668a709d" Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.161904 4997 scope.go:117] "RemoveContainer" containerID="19057ada048ebdcaf45377f1715685dd7c5a6f0b4e5f9adaafb95a84bf9a8e34" Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.171833 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f5zf6"] Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.190832 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f5zf6"] Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.198448 4997 scope.go:117] "RemoveContainer" containerID="93f7378ae293d06aa074a1c52337451ce5afce2e7deeb19639e80677b4b8cbe1" Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.240187 4997 scope.go:117] "RemoveContainer" containerID="6d947d303bff605be2173e6663955d3c51f7852c80212addbb5c4362668a709d" Dec 05 10:04:55 crc kubenswrapper[4997]: E1205 10:04:55.240786 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d947d303bff605be2173e6663955d3c51f7852c80212addbb5c4362668a709d\": container with ID starting with 6d947d303bff605be2173e6663955d3c51f7852c80212addbb5c4362668a709d not found: ID does not exist" containerID="6d947d303bff605be2173e6663955d3c51f7852c80212addbb5c4362668a709d" Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.240851 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d947d303bff605be2173e6663955d3c51f7852c80212addbb5c4362668a709d"} err="failed to get container status \"6d947d303bff605be2173e6663955d3c51f7852c80212addbb5c4362668a709d\": rpc error: code = NotFound desc = could not find container \"6d947d303bff605be2173e6663955d3c51f7852c80212addbb5c4362668a709d\": container with ID starting with 6d947d303bff605be2173e6663955d3c51f7852c80212addbb5c4362668a709d not found: ID does not exist" Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.240888 4997 scope.go:117] "RemoveContainer" containerID="19057ada048ebdcaf45377f1715685dd7c5a6f0b4e5f9adaafb95a84bf9a8e34" Dec 05 10:04:55 crc kubenswrapper[4997]: E1205 10:04:55.242196 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19057ada048ebdcaf45377f1715685dd7c5a6f0b4e5f9adaafb95a84bf9a8e34\": container with ID starting with 19057ada048ebdcaf45377f1715685dd7c5a6f0b4e5f9adaafb95a84bf9a8e34 not found: ID does not exist" containerID="19057ada048ebdcaf45377f1715685dd7c5a6f0b4e5f9adaafb95a84bf9a8e34" Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.242234 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19057ada048ebdcaf45377f1715685dd7c5a6f0b4e5f9adaafb95a84bf9a8e34"} err="failed to get container status \"19057ada048ebdcaf45377f1715685dd7c5a6f0b4e5f9adaafb95a84bf9a8e34\": rpc error: code = NotFound desc = could not find container \"19057ada048ebdcaf45377f1715685dd7c5a6f0b4e5f9adaafb95a84bf9a8e34\": container with ID starting with 19057ada048ebdcaf45377f1715685dd7c5a6f0b4e5f9adaafb95a84bf9a8e34 not found: ID does not exist" Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.242255 4997 scope.go:117] "RemoveContainer" containerID="93f7378ae293d06aa074a1c52337451ce5afce2e7deeb19639e80677b4b8cbe1" Dec 05 10:04:55 crc kubenswrapper[4997]: E1205 10:04:55.242581 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93f7378ae293d06aa074a1c52337451ce5afce2e7deeb19639e80677b4b8cbe1\": container with ID starting with 93f7378ae293d06aa074a1c52337451ce5afce2e7deeb19639e80677b4b8cbe1 not found: ID does not exist" containerID="93f7378ae293d06aa074a1c52337451ce5afce2e7deeb19639e80677b4b8cbe1" Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.242634 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93f7378ae293d06aa074a1c52337451ce5afce2e7deeb19639e80677b4b8cbe1"} err="failed to get container status \"93f7378ae293d06aa074a1c52337451ce5afce2e7deeb19639e80677b4b8cbe1\": rpc error: code = NotFound desc = could not find container \"93f7378ae293d06aa074a1c52337451ce5afce2e7deeb19639e80677b4b8cbe1\": container with ID starting with 93f7378ae293d06aa074a1c52337451ce5afce2e7deeb19639e80677b4b8cbe1 not found: ID does not exist" Dec 05 10:04:55 crc kubenswrapper[4997]: I1205 10:04:55.761718 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1d14bd5-3277-4bd5-a623-a899113f63c5" path="/var/lib/kubelet/pods/c1d14bd5-3277-4bd5-a623-a899113f63c5/volumes" Dec 05 10:06:19 crc kubenswrapper[4997]: I1205 10:06:19.770933 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:06:19 crc kubenswrapper[4997]: I1205 10:06:19.771517 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:06:49 crc kubenswrapper[4997]: I1205 10:06:49.769909 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:06:49 crc kubenswrapper[4997]: I1205 10:06:49.770549 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:07:19 crc kubenswrapper[4997]: I1205 10:07:19.770999 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:07:19 crc kubenswrapper[4997]: I1205 10:07:19.771653 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:07:19 crc kubenswrapper[4997]: I1205 10:07:19.771723 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 10:07:19 crc kubenswrapper[4997]: I1205 10:07:19.772769 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9a58315196354db15ade60cdec84e4c423cf75d724ed764e78dac051b6d4a462"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 10:07:19 crc kubenswrapper[4997]: I1205 10:07:19.772839 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://9a58315196354db15ade60cdec84e4c423cf75d724ed764e78dac051b6d4a462" gracePeriod=600 Dec 05 10:07:20 crc kubenswrapper[4997]: I1205 10:07:20.835868 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="9a58315196354db15ade60cdec84e4c423cf75d724ed764e78dac051b6d4a462" exitCode=0 Dec 05 10:07:20 crc kubenswrapper[4997]: I1205 10:07:20.836347 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"9a58315196354db15ade60cdec84e4c423cf75d724ed764e78dac051b6d4a462"} Dec 05 10:07:20 crc kubenswrapper[4997]: I1205 10:07:20.836382 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24"} Dec 05 10:07:20 crc kubenswrapper[4997]: I1205 10:07:20.836398 4997 scope.go:117] "RemoveContainer" containerID="e70376d17d707f2e538e3e589c0b0bfab1a2cb028aaddcfe218996dca90569c5" Dec 05 10:09:49 crc kubenswrapper[4997]: I1205 10:09:49.770140 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:09:49 crc kubenswrapper[4997]: I1205 10:09:49.770705 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:09:52 crc kubenswrapper[4997]: I1205 10:09:52.397121 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ddcsh"] Dec 05 10:09:52 crc kubenswrapper[4997]: E1205 10:09:52.398045 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1d14bd5-3277-4bd5-a623-a899113f63c5" containerName="extract-content" Dec 05 10:09:52 crc kubenswrapper[4997]: I1205 10:09:52.398066 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1d14bd5-3277-4bd5-a623-a899113f63c5" containerName="extract-content" Dec 05 10:09:52 crc kubenswrapper[4997]: E1205 10:09:52.398118 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1d14bd5-3277-4bd5-a623-a899113f63c5" containerName="registry-server" Dec 05 10:09:52 crc kubenswrapper[4997]: I1205 10:09:52.398126 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1d14bd5-3277-4bd5-a623-a899113f63c5" containerName="registry-server" Dec 05 10:09:52 crc kubenswrapper[4997]: E1205 10:09:52.398138 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1d14bd5-3277-4bd5-a623-a899113f63c5" containerName="extract-utilities" Dec 05 10:09:52 crc kubenswrapper[4997]: I1205 10:09:52.398146 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1d14bd5-3277-4bd5-a623-a899113f63c5" containerName="extract-utilities" Dec 05 10:09:52 crc kubenswrapper[4997]: I1205 10:09:52.398408 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1d14bd5-3277-4bd5-a623-a899113f63c5" containerName="registry-server" Dec 05 10:09:52 crc kubenswrapper[4997]: I1205 10:09:52.400435 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ddcsh" Dec 05 10:09:52 crc kubenswrapper[4997]: I1205 10:09:52.410541 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ddcsh"] Dec 05 10:09:52 crc kubenswrapper[4997]: I1205 10:09:52.438964 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-catalog-content\") pod \"certified-operators-ddcsh\" (UID: \"7b69efc7-1c53-4b8b-8a66-5a239c0c75db\") " pod="openshift-marketplace/certified-operators-ddcsh" Dec 05 10:09:52 crc kubenswrapper[4997]: I1205 10:09:52.439076 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pwml\" (UniqueName: \"kubernetes.io/projected/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-kube-api-access-5pwml\") pod \"certified-operators-ddcsh\" (UID: \"7b69efc7-1c53-4b8b-8a66-5a239c0c75db\") " pod="openshift-marketplace/certified-operators-ddcsh" Dec 05 10:09:52 crc kubenswrapper[4997]: I1205 10:09:52.439107 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-utilities\") pod \"certified-operators-ddcsh\" (UID: \"7b69efc7-1c53-4b8b-8a66-5a239c0c75db\") " pod="openshift-marketplace/certified-operators-ddcsh" Dec 05 10:09:52 crc kubenswrapper[4997]: I1205 10:09:52.540965 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-catalog-content\") pod \"certified-operators-ddcsh\" (UID: \"7b69efc7-1c53-4b8b-8a66-5a239c0c75db\") " pod="openshift-marketplace/certified-operators-ddcsh" Dec 05 10:09:52 crc kubenswrapper[4997]: I1205 10:09:52.541129 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pwml\" (UniqueName: \"kubernetes.io/projected/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-kube-api-access-5pwml\") pod \"certified-operators-ddcsh\" (UID: \"7b69efc7-1c53-4b8b-8a66-5a239c0c75db\") " pod="openshift-marketplace/certified-operators-ddcsh" Dec 05 10:09:52 crc kubenswrapper[4997]: I1205 10:09:52.541166 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-utilities\") pod \"certified-operators-ddcsh\" (UID: \"7b69efc7-1c53-4b8b-8a66-5a239c0c75db\") " pod="openshift-marketplace/certified-operators-ddcsh" Dec 05 10:09:52 crc kubenswrapper[4997]: I1205 10:09:52.541829 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-utilities\") pod \"certified-operators-ddcsh\" (UID: \"7b69efc7-1c53-4b8b-8a66-5a239c0c75db\") " pod="openshift-marketplace/certified-operators-ddcsh" Dec 05 10:09:52 crc kubenswrapper[4997]: I1205 10:09:52.542988 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-catalog-content\") pod \"certified-operators-ddcsh\" (UID: \"7b69efc7-1c53-4b8b-8a66-5a239c0c75db\") " pod="openshift-marketplace/certified-operators-ddcsh" Dec 05 10:09:52 crc kubenswrapper[4997]: I1205 10:09:52.564870 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pwml\" (UniqueName: \"kubernetes.io/projected/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-kube-api-access-5pwml\") pod \"certified-operators-ddcsh\" (UID: \"7b69efc7-1c53-4b8b-8a66-5a239c0c75db\") " pod="openshift-marketplace/certified-operators-ddcsh" Dec 05 10:09:52 crc kubenswrapper[4997]: I1205 10:09:52.732142 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ddcsh" Dec 05 10:09:53 crc kubenswrapper[4997]: I1205 10:09:53.327389 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ddcsh"] Dec 05 10:09:53 crc kubenswrapper[4997]: I1205 10:09:53.805508 4997 generic.go:334] "Generic (PLEG): container finished" podID="7b69efc7-1c53-4b8b-8a66-5a239c0c75db" containerID="8a81edc23b468ab28d91a13d1ab30c1c2095ab68dd6aa90b684e02627b55ae06" exitCode=0 Dec 05 10:09:53 crc kubenswrapper[4997]: I1205 10:09:53.805703 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddcsh" event={"ID":"7b69efc7-1c53-4b8b-8a66-5a239c0c75db","Type":"ContainerDied","Data":"8a81edc23b468ab28d91a13d1ab30c1c2095ab68dd6aa90b684e02627b55ae06"} Dec 05 10:09:53 crc kubenswrapper[4997]: I1205 10:09:53.805892 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddcsh" event={"ID":"7b69efc7-1c53-4b8b-8a66-5a239c0c75db","Type":"ContainerStarted","Data":"e52792750048ab83dde56e16ce41b59a73169a50569c815a916127211e47e911"} Dec 05 10:09:53 crc kubenswrapper[4997]: I1205 10:09:53.807962 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:09:54 crc kubenswrapper[4997]: I1205 10:09:54.818328 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddcsh" event={"ID":"7b69efc7-1c53-4b8b-8a66-5a239c0c75db","Type":"ContainerStarted","Data":"2b70e5e74f769850d3be587db5789e6249241d56c2742a9a13f1c1a25b38b715"} Dec 05 10:09:56 crc kubenswrapper[4997]: I1205 10:09:56.857265 4997 generic.go:334] "Generic (PLEG): container finished" podID="7b69efc7-1c53-4b8b-8a66-5a239c0c75db" containerID="2b70e5e74f769850d3be587db5789e6249241d56c2742a9a13f1c1a25b38b715" exitCode=0 Dec 05 10:09:56 crc kubenswrapper[4997]: I1205 10:09:56.857578 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddcsh" event={"ID":"7b69efc7-1c53-4b8b-8a66-5a239c0c75db","Type":"ContainerDied","Data":"2b70e5e74f769850d3be587db5789e6249241d56c2742a9a13f1c1a25b38b715"} Dec 05 10:09:57 crc kubenswrapper[4997]: I1205 10:09:57.872179 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddcsh" event={"ID":"7b69efc7-1c53-4b8b-8a66-5a239c0c75db","Type":"ContainerStarted","Data":"ede82f2b1d670a34326b66833dabaae1803ec6612370d665ee8923257bc22402"} Dec 05 10:09:57 crc kubenswrapper[4997]: I1205 10:09:57.906195 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ddcsh" podStartSLOduration=2.326719389 podStartE2EDuration="5.906177283s" podCreationTimestamp="2025-12-05 10:09:52 +0000 UTC" firstStartedPulling="2025-12-05 10:09:53.807649705 +0000 UTC m=+11694.336556976" lastFinishedPulling="2025-12-05 10:09:57.387107609 +0000 UTC m=+11697.916014870" observedRunningTime="2025-12-05 10:09:57.889365518 +0000 UTC m=+11698.418272839" watchObservedRunningTime="2025-12-05 10:09:57.906177283 +0000 UTC m=+11698.435084544" Dec 05 10:10:02 crc kubenswrapper[4997]: I1205 10:10:02.732341 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ddcsh" Dec 05 10:10:02 crc kubenswrapper[4997]: I1205 10:10:02.732968 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ddcsh" Dec 05 10:10:02 crc kubenswrapper[4997]: I1205 10:10:02.785117 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ddcsh" Dec 05 10:10:02 crc kubenswrapper[4997]: I1205 10:10:02.995095 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ddcsh" Dec 05 10:10:06 crc kubenswrapper[4997]: I1205 10:10:06.130539 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ddcsh"] Dec 05 10:10:06 crc kubenswrapper[4997]: I1205 10:10:06.132553 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ddcsh" podUID="7b69efc7-1c53-4b8b-8a66-5a239c0c75db" containerName="registry-server" containerID="cri-o://ede82f2b1d670a34326b66833dabaae1803ec6612370d665ee8923257bc22402" gracePeriod=2 Dec 05 10:10:06 crc kubenswrapper[4997]: I1205 10:10:06.984912 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ddcsh" Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.013721 4997 generic.go:334] "Generic (PLEG): container finished" podID="7b69efc7-1c53-4b8b-8a66-5a239c0c75db" containerID="ede82f2b1d670a34326b66833dabaae1803ec6612370d665ee8923257bc22402" exitCode=0 Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.013771 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddcsh" event={"ID":"7b69efc7-1c53-4b8b-8a66-5a239c0c75db","Type":"ContainerDied","Data":"ede82f2b1d670a34326b66833dabaae1803ec6612370d665ee8923257bc22402"} Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.013799 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddcsh" event={"ID":"7b69efc7-1c53-4b8b-8a66-5a239c0c75db","Type":"ContainerDied","Data":"e52792750048ab83dde56e16ce41b59a73169a50569c815a916127211e47e911"} Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.013817 4997 scope.go:117] "RemoveContainer" containerID="ede82f2b1d670a34326b66833dabaae1803ec6612370d665ee8923257bc22402" Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.013837 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ddcsh" Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.040938 4997 scope.go:117] "RemoveContainer" containerID="2b70e5e74f769850d3be587db5789e6249241d56c2742a9a13f1c1a25b38b715" Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.067490 4997 scope.go:117] "RemoveContainer" containerID="8a81edc23b468ab28d91a13d1ab30c1c2095ab68dd6aa90b684e02627b55ae06" Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.072344 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-catalog-content\") pod \"7b69efc7-1c53-4b8b-8a66-5a239c0c75db\" (UID: \"7b69efc7-1c53-4b8b-8a66-5a239c0c75db\") " Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.072543 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-utilities\") pod \"7b69efc7-1c53-4b8b-8a66-5a239c0c75db\" (UID: \"7b69efc7-1c53-4b8b-8a66-5a239c0c75db\") " Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.075501 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pwml\" (UniqueName: \"kubernetes.io/projected/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-kube-api-access-5pwml\") pod \"7b69efc7-1c53-4b8b-8a66-5a239c0c75db\" (UID: \"7b69efc7-1c53-4b8b-8a66-5a239c0c75db\") " Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.075730 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-utilities" (OuterVolumeSpecName: "utilities") pod "7b69efc7-1c53-4b8b-8a66-5a239c0c75db" (UID: "7b69efc7-1c53-4b8b-8a66-5a239c0c75db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.076260 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.084053 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-kube-api-access-5pwml" (OuterVolumeSpecName: "kube-api-access-5pwml") pod "7b69efc7-1c53-4b8b-8a66-5a239c0c75db" (UID: "7b69efc7-1c53-4b8b-8a66-5a239c0c75db"). InnerVolumeSpecName "kube-api-access-5pwml". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.123290 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b69efc7-1c53-4b8b-8a66-5a239c0c75db" (UID: "7b69efc7-1c53-4b8b-8a66-5a239c0c75db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.160184 4997 scope.go:117] "RemoveContainer" containerID="ede82f2b1d670a34326b66833dabaae1803ec6612370d665ee8923257bc22402" Dec 05 10:10:07 crc kubenswrapper[4997]: E1205 10:10:07.162326 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ede82f2b1d670a34326b66833dabaae1803ec6612370d665ee8923257bc22402\": container with ID starting with ede82f2b1d670a34326b66833dabaae1803ec6612370d665ee8923257bc22402 not found: ID does not exist" containerID="ede82f2b1d670a34326b66833dabaae1803ec6612370d665ee8923257bc22402" Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.162364 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ede82f2b1d670a34326b66833dabaae1803ec6612370d665ee8923257bc22402"} err="failed to get container status \"ede82f2b1d670a34326b66833dabaae1803ec6612370d665ee8923257bc22402\": rpc error: code = NotFound desc = could not find container \"ede82f2b1d670a34326b66833dabaae1803ec6612370d665ee8923257bc22402\": container with ID starting with ede82f2b1d670a34326b66833dabaae1803ec6612370d665ee8923257bc22402 not found: ID does not exist" Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.162410 4997 scope.go:117] "RemoveContainer" containerID="2b70e5e74f769850d3be587db5789e6249241d56c2742a9a13f1c1a25b38b715" Dec 05 10:10:07 crc kubenswrapper[4997]: E1205 10:10:07.162910 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b70e5e74f769850d3be587db5789e6249241d56c2742a9a13f1c1a25b38b715\": container with ID starting with 2b70e5e74f769850d3be587db5789e6249241d56c2742a9a13f1c1a25b38b715 not found: ID does not exist" containerID="2b70e5e74f769850d3be587db5789e6249241d56c2742a9a13f1c1a25b38b715" Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.163047 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b70e5e74f769850d3be587db5789e6249241d56c2742a9a13f1c1a25b38b715"} err="failed to get container status \"2b70e5e74f769850d3be587db5789e6249241d56c2742a9a13f1c1a25b38b715\": rpc error: code = NotFound desc = could not find container \"2b70e5e74f769850d3be587db5789e6249241d56c2742a9a13f1c1a25b38b715\": container with ID starting with 2b70e5e74f769850d3be587db5789e6249241d56c2742a9a13f1c1a25b38b715 not found: ID does not exist" Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.163128 4997 scope.go:117] "RemoveContainer" containerID="8a81edc23b468ab28d91a13d1ab30c1c2095ab68dd6aa90b684e02627b55ae06" Dec 05 10:10:07 crc kubenswrapper[4997]: E1205 10:10:07.163484 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a81edc23b468ab28d91a13d1ab30c1c2095ab68dd6aa90b684e02627b55ae06\": container with ID starting with 8a81edc23b468ab28d91a13d1ab30c1c2095ab68dd6aa90b684e02627b55ae06 not found: ID does not exist" containerID="8a81edc23b468ab28d91a13d1ab30c1c2095ab68dd6aa90b684e02627b55ae06" Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.163521 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a81edc23b468ab28d91a13d1ab30c1c2095ab68dd6aa90b684e02627b55ae06"} err="failed to get container status \"8a81edc23b468ab28d91a13d1ab30c1c2095ab68dd6aa90b684e02627b55ae06\": rpc error: code = NotFound desc = could not find container \"8a81edc23b468ab28d91a13d1ab30c1c2095ab68dd6aa90b684e02627b55ae06\": container with ID starting with 8a81edc23b468ab28d91a13d1ab30c1c2095ab68dd6aa90b684e02627b55ae06 not found: ID does not exist" Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.178728 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pwml\" (UniqueName: \"kubernetes.io/projected/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-kube-api-access-5pwml\") on node \"crc\" DevicePath \"\"" Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.178766 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b69efc7-1c53-4b8b-8a66-5a239c0c75db-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.364730 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ddcsh"] Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.377870 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ddcsh"] Dec 05 10:10:07 crc kubenswrapper[4997]: I1205 10:10:07.763587 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b69efc7-1c53-4b8b-8a66-5a239c0c75db" path="/var/lib/kubelet/pods/7b69efc7-1c53-4b8b-8a66-5a239c0c75db/volumes" Dec 05 10:10:19 crc kubenswrapper[4997]: I1205 10:10:19.769777 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:10:19 crc kubenswrapper[4997]: I1205 10:10:19.772220 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:10:49 crc kubenswrapper[4997]: I1205 10:10:49.770380 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:10:49 crc kubenswrapper[4997]: I1205 10:10:49.771382 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:10:49 crc kubenswrapper[4997]: I1205 10:10:49.771453 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 10:10:49 crc kubenswrapper[4997]: I1205 10:10:49.772706 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 10:10:49 crc kubenswrapper[4997]: I1205 10:10:49.772817 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" gracePeriod=600 Dec 05 10:10:49 crc kubenswrapper[4997]: E1205 10:10:49.900524 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:10:50 crc kubenswrapper[4997]: I1205 10:10:50.513797 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" exitCode=0 Dec 05 10:10:50 crc kubenswrapper[4997]: I1205 10:10:50.513874 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24"} Dec 05 10:10:50 crc kubenswrapper[4997]: I1205 10:10:50.515637 4997 scope.go:117] "RemoveContainer" containerID="9a58315196354db15ade60cdec84e4c423cf75d724ed764e78dac051b6d4a462" Dec 05 10:10:50 crc kubenswrapper[4997]: I1205 10:10:50.516307 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:10:50 crc kubenswrapper[4997]: E1205 10:10:50.516725 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:11:04 crc kubenswrapper[4997]: I1205 10:11:04.749390 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:11:04 crc kubenswrapper[4997]: E1205 10:11:04.750206 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:11:18 crc kubenswrapper[4997]: I1205 10:11:18.749875 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:11:18 crc kubenswrapper[4997]: E1205 10:11:18.751165 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:11:31 crc kubenswrapper[4997]: I1205 10:11:31.749136 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:11:31 crc kubenswrapper[4997]: E1205 10:11:31.749923 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:11:42 crc kubenswrapper[4997]: I1205 10:11:42.750274 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:11:42 crc kubenswrapper[4997]: E1205 10:11:42.751413 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:11:54 crc kubenswrapper[4997]: I1205 10:11:54.749557 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:11:54 crc kubenswrapper[4997]: E1205 10:11:54.750207 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:12:05 crc kubenswrapper[4997]: I1205 10:12:05.749952 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:12:05 crc kubenswrapper[4997]: E1205 10:12:05.750813 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:12:17 crc kubenswrapper[4997]: I1205 10:12:17.749143 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:12:17 crc kubenswrapper[4997]: E1205 10:12:17.750892 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:12:31 crc kubenswrapper[4997]: I1205 10:12:31.749268 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:12:31 crc kubenswrapper[4997]: E1205 10:12:31.750108 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:12:46 crc kubenswrapper[4997]: I1205 10:12:46.750006 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:12:46 crc kubenswrapper[4997]: E1205 10:12:46.750986 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:13:00 crc kubenswrapper[4997]: I1205 10:13:00.750305 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:13:00 crc kubenswrapper[4997]: E1205 10:13:00.751476 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:13:05 crc kubenswrapper[4997]: I1205 10:13:05.276706 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6qj8b"] Dec 05 10:13:05 crc kubenswrapper[4997]: E1205 10:13:05.277897 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b69efc7-1c53-4b8b-8a66-5a239c0c75db" containerName="registry-server" Dec 05 10:13:05 crc kubenswrapper[4997]: I1205 10:13:05.277925 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b69efc7-1c53-4b8b-8a66-5a239c0c75db" containerName="registry-server" Dec 05 10:13:05 crc kubenswrapper[4997]: E1205 10:13:05.277948 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b69efc7-1c53-4b8b-8a66-5a239c0c75db" containerName="extract-content" Dec 05 10:13:05 crc kubenswrapper[4997]: I1205 10:13:05.277956 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b69efc7-1c53-4b8b-8a66-5a239c0c75db" containerName="extract-content" Dec 05 10:13:05 crc kubenswrapper[4997]: E1205 10:13:05.277974 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b69efc7-1c53-4b8b-8a66-5a239c0c75db" containerName="extract-utilities" Dec 05 10:13:05 crc kubenswrapper[4997]: I1205 10:13:05.277983 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b69efc7-1c53-4b8b-8a66-5a239c0c75db" containerName="extract-utilities" Dec 05 10:13:05 crc kubenswrapper[4997]: I1205 10:13:05.278244 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b69efc7-1c53-4b8b-8a66-5a239c0c75db" containerName="registry-server" Dec 05 10:13:05 crc kubenswrapper[4997]: I1205 10:13:05.280149 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6qj8b" Dec 05 10:13:05 crc kubenswrapper[4997]: I1205 10:13:05.283331 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6qj8b"] Dec 05 10:13:05 crc kubenswrapper[4997]: I1205 10:13:05.354755 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18f00986-d79c-428c-99c6-3a30f1aa1442-utilities\") pod \"redhat-marketplace-6qj8b\" (UID: \"18f00986-d79c-428c-99c6-3a30f1aa1442\") " pod="openshift-marketplace/redhat-marketplace-6qj8b" Dec 05 10:13:05 crc kubenswrapper[4997]: I1205 10:13:05.354841 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xdb6\" (UniqueName: \"kubernetes.io/projected/18f00986-d79c-428c-99c6-3a30f1aa1442-kube-api-access-8xdb6\") pod \"redhat-marketplace-6qj8b\" (UID: \"18f00986-d79c-428c-99c6-3a30f1aa1442\") " pod="openshift-marketplace/redhat-marketplace-6qj8b" Dec 05 10:13:05 crc kubenswrapper[4997]: I1205 10:13:05.354879 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18f00986-d79c-428c-99c6-3a30f1aa1442-catalog-content\") pod \"redhat-marketplace-6qj8b\" (UID: \"18f00986-d79c-428c-99c6-3a30f1aa1442\") " pod="openshift-marketplace/redhat-marketplace-6qj8b" Dec 05 10:13:05 crc kubenswrapper[4997]: I1205 10:13:05.456443 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18f00986-d79c-428c-99c6-3a30f1aa1442-utilities\") pod \"redhat-marketplace-6qj8b\" (UID: \"18f00986-d79c-428c-99c6-3a30f1aa1442\") " pod="openshift-marketplace/redhat-marketplace-6qj8b" Dec 05 10:13:05 crc kubenswrapper[4997]: I1205 10:13:05.456538 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xdb6\" (UniqueName: \"kubernetes.io/projected/18f00986-d79c-428c-99c6-3a30f1aa1442-kube-api-access-8xdb6\") pod \"redhat-marketplace-6qj8b\" (UID: \"18f00986-d79c-428c-99c6-3a30f1aa1442\") " pod="openshift-marketplace/redhat-marketplace-6qj8b" Dec 05 10:13:05 crc kubenswrapper[4997]: I1205 10:13:05.456578 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18f00986-d79c-428c-99c6-3a30f1aa1442-catalog-content\") pod \"redhat-marketplace-6qj8b\" (UID: \"18f00986-d79c-428c-99c6-3a30f1aa1442\") " pod="openshift-marketplace/redhat-marketplace-6qj8b" Dec 05 10:13:05 crc kubenswrapper[4997]: I1205 10:13:05.457379 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18f00986-d79c-428c-99c6-3a30f1aa1442-catalog-content\") pod \"redhat-marketplace-6qj8b\" (UID: \"18f00986-d79c-428c-99c6-3a30f1aa1442\") " pod="openshift-marketplace/redhat-marketplace-6qj8b" Dec 05 10:13:05 crc kubenswrapper[4997]: I1205 10:13:05.457693 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18f00986-d79c-428c-99c6-3a30f1aa1442-utilities\") pod \"redhat-marketplace-6qj8b\" (UID: \"18f00986-d79c-428c-99c6-3a30f1aa1442\") " pod="openshift-marketplace/redhat-marketplace-6qj8b" Dec 05 10:13:05 crc kubenswrapper[4997]: I1205 10:13:05.482323 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xdb6\" (UniqueName: \"kubernetes.io/projected/18f00986-d79c-428c-99c6-3a30f1aa1442-kube-api-access-8xdb6\") pod \"redhat-marketplace-6qj8b\" (UID: \"18f00986-d79c-428c-99c6-3a30f1aa1442\") " pod="openshift-marketplace/redhat-marketplace-6qj8b" Dec 05 10:13:05 crc kubenswrapper[4997]: I1205 10:13:05.625273 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6qj8b" Dec 05 10:13:06 crc kubenswrapper[4997]: I1205 10:13:06.267305 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6qj8b"] Dec 05 10:13:07 crc kubenswrapper[4997]: I1205 10:13:07.072531 4997 generic.go:334] "Generic (PLEG): container finished" podID="18f00986-d79c-428c-99c6-3a30f1aa1442" containerID="61eaf22c809ae6b7082fc1a71d37dd844fa4fb37aea8c6feda6834b6e457cd92" exitCode=0 Dec 05 10:13:07 crc kubenswrapper[4997]: I1205 10:13:07.072760 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6qj8b" event={"ID":"18f00986-d79c-428c-99c6-3a30f1aa1442","Type":"ContainerDied","Data":"61eaf22c809ae6b7082fc1a71d37dd844fa4fb37aea8c6feda6834b6e457cd92"} Dec 05 10:13:07 crc kubenswrapper[4997]: I1205 10:13:07.072818 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6qj8b" event={"ID":"18f00986-d79c-428c-99c6-3a30f1aa1442","Type":"ContainerStarted","Data":"a0a284e9ee2cef4bccdfe926ed63997f3c25ee611cfd5e3b6a864adc7fc1b71f"} Dec 05 10:13:08 crc kubenswrapper[4997]: I1205 10:13:08.088714 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6qj8b" event={"ID":"18f00986-d79c-428c-99c6-3a30f1aa1442","Type":"ContainerStarted","Data":"083c4327db990099e64bac3edb094b8cc907369c06ae6a3910e63658f88f706e"} Dec 05 10:13:09 crc kubenswrapper[4997]: I1205 10:13:09.110565 4997 generic.go:334] "Generic (PLEG): container finished" podID="18f00986-d79c-428c-99c6-3a30f1aa1442" containerID="083c4327db990099e64bac3edb094b8cc907369c06ae6a3910e63658f88f706e" exitCode=0 Dec 05 10:13:09 crc kubenswrapper[4997]: I1205 10:13:09.110668 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6qj8b" event={"ID":"18f00986-d79c-428c-99c6-3a30f1aa1442","Type":"ContainerDied","Data":"083c4327db990099e64bac3edb094b8cc907369c06ae6a3910e63658f88f706e"} Dec 05 10:13:10 crc kubenswrapper[4997]: I1205 10:13:10.129136 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6qj8b" event={"ID":"18f00986-d79c-428c-99c6-3a30f1aa1442","Type":"ContainerStarted","Data":"6792f171268941c09db09bb4b5ca62bbf792d13626bd4dda6580d9336e927fe4"} Dec 05 10:13:10 crc kubenswrapper[4997]: I1205 10:13:10.150587 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6qj8b" podStartSLOduration=2.688385487 podStartE2EDuration="5.150568577s" podCreationTimestamp="2025-12-05 10:13:05 +0000 UTC" firstStartedPulling="2025-12-05 10:13:07.074872323 +0000 UTC m=+11887.603779584" lastFinishedPulling="2025-12-05 10:13:09.537055413 +0000 UTC m=+11890.065962674" observedRunningTime="2025-12-05 10:13:10.146792665 +0000 UTC m=+11890.675699936" watchObservedRunningTime="2025-12-05 10:13:10.150568577 +0000 UTC m=+11890.679475838" Dec 05 10:13:15 crc kubenswrapper[4997]: I1205 10:13:15.628223 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6qj8b" Dec 05 10:13:15 crc kubenswrapper[4997]: I1205 10:13:15.629237 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6qj8b" Dec 05 10:13:15 crc kubenswrapper[4997]: I1205 10:13:15.677157 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6qj8b" Dec 05 10:13:15 crc kubenswrapper[4997]: I1205 10:13:15.749651 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:13:15 crc kubenswrapper[4997]: E1205 10:13:15.749996 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:13:16 crc kubenswrapper[4997]: I1205 10:13:16.243060 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6qj8b" Dec 05 10:13:16 crc kubenswrapper[4997]: I1205 10:13:16.309440 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6qj8b"] Dec 05 10:13:18 crc kubenswrapper[4997]: I1205 10:13:18.211324 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6qj8b" podUID="18f00986-d79c-428c-99c6-3a30f1aa1442" containerName="registry-server" containerID="cri-o://6792f171268941c09db09bb4b5ca62bbf792d13626bd4dda6580d9336e927fe4" gracePeriod=2 Dec 05 10:13:19 crc kubenswrapper[4997]: I1205 10:13:19.225427 4997 generic.go:334] "Generic (PLEG): container finished" podID="18f00986-d79c-428c-99c6-3a30f1aa1442" containerID="6792f171268941c09db09bb4b5ca62bbf792d13626bd4dda6580d9336e927fe4" exitCode=0 Dec 05 10:13:19 crc kubenswrapper[4997]: I1205 10:13:19.226040 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6qj8b" event={"ID":"18f00986-d79c-428c-99c6-3a30f1aa1442","Type":"ContainerDied","Data":"6792f171268941c09db09bb4b5ca62bbf792d13626bd4dda6580d9336e927fe4"} Dec 05 10:13:19 crc kubenswrapper[4997]: I1205 10:13:19.448712 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6qj8b" Dec 05 10:13:19 crc kubenswrapper[4997]: I1205 10:13:19.569542 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xdb6\" (UniqueName: \"kubernetes.io/projected/18f00986-d79c-428c-99c6-3a30f1aa1442-kube-api-access-8xdb6\") pod \"18f00986-d79c-428c-99c6-3a30f1aa1442\" (UID: \"18f00986-d79c-428c-99c6-3a30f1aa1442\") " Dec 05 10:13:19 crc kubenswrapper[4997]: I1205 10:13:19.569948 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18f00986-d79c-428c-99c6-3a30f1aa1442-catalog-content\") pod \"18f00986-d79c-428c-99c6-3a30f1aa1442\" (UID: \"18f00986-d79c-428c-99c6-3a30f1aa1442\") " Dec 05 10:13:19 crc kubenswrapper[4997]: I1205 10:13:19.570165 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18f00986-d79c-428c-99c6-3a30f1aa1442-utilities\") pod \"18f00986-d79c-428c-99c6-3a30f1aa1442\" (UID: \"18f00986-d79c-428c-99c6-3a30f1aa1442\") " Dec 05 10:13:19 crc kubenswrapper[4997]: I1205 10:13:19.571053 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18f00986-d79c-428c-99c6-3a30f1aa1442-utilities" (OuterVolumeSpecName: "utilities") pod "18f00986-d79c-428c-99c6-3a30f1aa1442" (UID: "18f00986-d79c-428c-99c6-3a30f1aa1442"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:13:19 crc kubenswrapper[4997]: I1205 10:13:19.572421 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18f00986-d79c-428c-99c6-3a30f1aa1442-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:13:19 crc kubenswrapper[4997]: I1205 10:13:19.578725 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18f00986-d79c-428c-99c6-3a30f1aa1442-kube-api-access-8xdb6" (OuterVolumeSpecName: "kube-api-access-8xdb6") pod "18f00986-d79c-428c-99c6-3a30f1aa1442" (UID: "18f00986-d79c-428c-99c6-3a30f1aa1442"). InnerVolumeSpecName "kube-api-access-8xdb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:13:19 crc kubenswrapper[4997]: I1205 10:13:19.603191 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18f00986-d79c-428c-99c6-3a30f1aa1442-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18f00986-d79c-428c-99c6-3a30f1aa1442" (UID: "18f00986-d79c-428c-99c6-3a30f1aa1442"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:13:19 crc kubenswrapper[4997]: I1205 10:13:19.675093 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xdb6\" (UniqueName: \"kubernetes.io/projected/18f00986-d79c-428c-99c6-3a30f1aa1442-kube-api-access-8xdb6\") on node \"crc\" DevicePath \"\"" Dec 05 10:13:19 crc kubenswrapper[4997]: I1205 10:13:19.675810 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18f00986-d79c-428c-99c6-3a30f1aa1442-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:13:20 crc kubenswrapper[4997]: I1205 10:13:20.242548 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6qj8b" event={"ID":"18f00986-d79c-428c-99c6-3a30f1aa1442","Type":"ContainerDied","Data":"a0a284e9ee2cef4bccdfe926ed63997f3c25ee611cfd5e3b6a864adc7fc1b71f"} Dec 05 10:13:20 crc kubenswrapper[4997]: I1205 10:13:20.242604 4997 scope.go:117] "RemoveContainer" containerID="6792f171268941c09db09bb4b5ca62bbf792d13626bd4dda6580d9336e927fe4" Dec 05 10:13:20 crc kubenswrapper[4997]: I1205 10:13:20.242736 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6qj8b" Dec 05 10:13:20 crc kubenswrapper[4997]: I1205 10:13:20.269498 4997 scope.go:117] "RemoveContainer" containerID="083c4327db990099e64bac3edb094b8cc907369c06ae6a3910e63658f88f706e" Dec 05 10:13:20 crc kubenswrapper[4997]: I1205 10:13:20.275982 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6qj8b"] Dec 05 10:13:20 crc kubenswrapper[4997]: I1205 10:13:20.289039 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6qj8b"] Dec 05 10:13:20 crc kubenswrapper[4997]: I1205 10:13:20.307848 4997 scope.go:117] "RemoveContainer" containerID="61eaf22c809ae6b7082fc1a71d37dd844fa4fb37aea8c6feda6834b6e457cd92" Dec 05 10:13:21 crc kubenswrapper[4997]: I1205 10:13:21.764148 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18f00986-d79c-428c-99c6-3a30f1aa1442" path="/var/lib/kubelet/pods/18f00986-d79c-428c-99c6-3a30f1aa1442/volumes" Dec 05 10:13:23 crc kubenswrapper[4997]: I1205 10:13:23.248340 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9nlgw"] Dec 05 10:13:23 crc kubenswrapper[4997]: E1205 10:13:23.249194 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18f00986-d79c-428c-99c6-3a30f1aa1442" containerName="extract-utilities" Dec 05 10:13:23 crc kubenswrapper[4997]: I1205 10:13:23.249209 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="18f00986-d79c-428c-99c6-3a30f1aa1442" containerName="extract-utilities" Dec 05 10:13:23 crc kubenswrapper[4997]: E1205 10:13:23.249222 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18f00986-d79c-428c-99c6-3a30f1aa1442" containerName="registry-server" Dec 05 10:13:23 crc kubenswrapper[4997]: I1205 10:13:23.249228 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="18f00986-d79c-428c-99c6-3a30f1aa1442" containerName="registry-server" Dec 05 10:13:23 crc kubenswrapper[4997]: E1205 10:13:23.249243 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18f00986-d79c-428c-99c6-3a30f1aa1442" containerName="extract-content" Dec 05 10:13:23 crc kubenswrapper[4997]: I1205 10:13:23.249248 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="18f00986-d79c-428c-99c6-3a30f1aa1442" containerName="extract-content" Dec 05 10:13:23 crc kubenswrapper[4997]: I1205 10:13:23.249447 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="18f00986-d79c-428c-99c6-3a30f1aa1442" containerName="registry-server" Dec 05 10:13:23 crc kubenswrapper[4997]: I1205 10:13:23.252276 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9nlgw" Dec 05 10:13:23 crc kubenswrapper[4997]: I1205 10:13:23.268967 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9nlgw"] Dec 05 10:13:23 crc kubenswrapper[4997]: I1205 10:13:23.355198 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-utilities\") pod \"community-operators-9nlgw\" (UID: \"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed\") " pod="openshift-marketplace/community-operators-9nlgw" Dec 05 10:13:23 crc kubenswrapper[4997]: I1205 10:13:23.355289 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-catalog-content\") pod \"community-operators-9nlgw\" (UID: \"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed\") " pod="openshift-marketplace/community-operators-9nlgw" Dec 05 10:13:23 crc kubenswrapper[4997]: I1205 10:13:23.355371 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnvl4\" (UniqueName: \"kubernetes.io/projected/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-kube-api-access-mnvl4\") pod \"community-operators-9nlgw\" (UID: \"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed\") " pod="openshift-marketplace/community-operators-9nlgw" Dec 05 10:13:23 crc kubenswrapper[4997]: I1205 10:13:23.457197 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-utilities\") pod \"community-operators-9nlgw\" (UID: \"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed\") " pod="openshift-marketplace/community-operators-9nlgw" Dec 05 10:13:23 crc kubenswrapper[4997]: I1205 10:13:23.457269 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-catalog-content\") pod \"community-operators-9nlgw\" (UID: \"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed\") " pod="openshift-marketplace/community-operators-9nlgw" Dec 05 10:13:23 crc kubenswrapper[4997]: I1205 10:13:23.457342 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnvl4\" (UniqueName: \"kubernetes.io/projected/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-kube-api-access-mnvl4\") pod \"community-operators-9nlgw\" (UID: \"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed\") " pod="openshift-marketplace/community-operators-9nlgw" Dec 05 10:13:23 crc kubenswrapper[4997]: I1205 10:13:23.457746 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-utilities\") pod \"community-operators-9nlgw\" (UID: \"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed\") " pod="openshift-marketplace/community-operators-9nlgw" Dec 05 10:13:23 crc kubenswrapper[4997]: I1205 10:13:23.458139 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-catalog-content\") pod \"community-operators-9nlgw\" (UID: \"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed\") " pod="openshift-marketplace/community-operators-9nlgw" Dec 05 10:13:23 crc kubenswrapper[4997]: I1205 10:13:23.498576 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnvl4\" (UniqueName: \"kubernetes.io/projected/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-kube-api-access-mnvl4\") pod \"community-operators-9nlgw\" (UID: \"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed\") " pod="openshift-marketplace/community-operators-9nlgw" Dec 05 10:13:23 crc kubenswrapper[4997]: I1205 10:13:23.579042 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9nlgw" Dec 05 10:13:24 crc kubenswrapper[4997]: I1205 10:13:24.318847 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9nlgw"] Dec 05 10:13:25 crc kubenswrapper[4997]: I1205 10:13:25.316549 4997 generic.go:334] "Generic (PLEG): container finished" podID="6e981f9f-0caf-490c-a9c2-eb6f0e7609ed" containerID="7ee6e92496bfa9ef055f436a6c168171eeab63623f10da7635d2718877c1ec84" exitCode=0 Dec 05 10:13:25 crc kubenswrapper[4997]: I1205 10:13:25.316673 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9nlgw" event={"ID":"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed","Type":"ContainerDied","Data":"7ee6e92496bfa9ef055f436a6c168171eeab63623f10da7635d2718877c1ec84"} Dec 05 10:13:25 crc kubenswrapper[4997]: I1205 10:13:25.317255 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9nlgw" event={"ID":"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed","Type":"ContainerStarted","Data":"998a30f8b3b33668d69a6e41ea89e3b5834f605b885c35c0f3b1747a0ce41ba5"} Dec 05 10:13:27 crc kubenswrapper[4997]: I1205 10:13:27.348997 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9nlgw" event={"ID":"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed","Type":"ContainerStarted","Data":"75d855f8293646353533a48c413e4568eb046f6d8e1b0643beb63691dc222ad5"} Dec 05 10:13:27 crc kubenswrapper[4997]: I1205 10:13:27.749896 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:13:27 crc kubenswrapper[4997]: E1205 10:13:27.750425 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:13:28 crc kubenswrapper[4997]: I1205 10:13:28.360705 4997 generic.go:334] "Generic (PLEG): container finished" podID="6e981f9f-0caf-490c-a9c2-eb6f0e7609ed" containerID="75d855f8293646353533a48c413e4568eb046f6d8e1b0643beb63691dc222ad5" exitCode=0 Dec 05 10:13:28 crc kubenswrapper[4997]: I1205 10:13:28.360747 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9nlgw" event={"ID":"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed","Type":"ContainerDied","Data":"75d855f8293646353533a48c413e4568eb046f6d8e1b0643beb63691dc222ad5"} Dec 05 10:13:29 crc kubenswrapper[4997]: I1205 10:13:29.374916 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9nlgw" event={"ID":"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed","Type":"ContainerStarted","Data":"0ffc3605a015c9ad6c24e020ab1e1f8ab44c79a8a394bb795e83b562431f3bb9"} Dec 05 10:13:29 crc kubenswrapper[4997]: I1205 10:13:29.414328 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9nlgw" podStartSLOduration=2.950704178 podStartE2EDuration="6.414302972s" podCreationTimestamp="2025-12-05 10:13:23 +0000 UTC" firstStartedPulling="2025-12-05 10:13:25.325797255 +0000 UTC m=+11905.854704516" lastFinishedPulling="2025-12-05 10:13:28.789396049 +0000 UTC m=+11909.318303310" observedRunningTime="2025-12-05 10:13:29.404368023 +0000 UTC m=+11909.933275274" watchObservedRunningTime="2025-12-05 10:13:29.414302972 +0000 UTC m=+11909.943210233" Dec 05 10:13:33 crc kubenswrapper[4997]: I1205 10:13:33.579251 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9nlgw" Dec 05 10:13:33 crc kubenswrapper[4997]: I1205 10:13:33.579882 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9nlgw" Dec 05 10:13:33 crc kubenswrapper[4997]: I1205 10:13:33.642465 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9nlgw" Dec 05 10:13:34 crc kubenswrapper[4997]: I1205 10:13:34.526589 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9nlgw" Dec 05 10:13:34 crc kubenswrapper[4997]: I1205 10:13:34.590926 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9nlgw"] Dec 05 10:13:36 crc kubenswrapper[4997]: I1205 10:13:36.501678 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9nlgw" podUID="6e981f9f-0caf-490c-a9c2-eb6f0e7609ed" containerName="registry-server" containerID="cri-o://0ffc3605a015c9ad6c24e020ab1e1f8ab44c79a8a394bb795e83b562431f3bb9" gracePeriod=2 Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.193670 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9nlgw" Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.262459 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-utilities\") pod \"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed\" (UID: \"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed\") " Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.262742 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-catalog-content\") pod \"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed\" (UID: \"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed\") " Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.262773 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnvl4\" (UniqueName: \"kubernetes.io/projected/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-kube-api-access-mnvl4\") pod \"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed\" (UID: \"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed\") " Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.263341 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-utilities" (OuterVolumeSpecName: "utilities") pod "6e981f9f-0caf-490c-a9c2-eb6f0e7609ed" (UID: "6e981f9f-0caf-490c-a9c2-eb6f0e7609ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.316238 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e981f9f-0caf-490c-a9c2-eb6f0e7609ed" (UID: "6e981f9f-0caf-490c-a9c2-eb6f0e7609ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.346510 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-kube-api-access-mnvl4" (OuterVolumeSpecName: "kube-api-access-mnvl4") pod "6e981f9f-0caf-490c-a9c2-eb6f0e7609ed" (UID: "6e981f9f-0caf-490c-a9c2-eb6f0e7609ed"). InnerVolumeSpecName "kube-api-access-mnvl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.375284 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.375324 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnvl4\" (UniqueName: \"kubernetes.io/projected/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-kube-api-access-mnvl4\") on node \"crc\" DevicePath \"\"" Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.375340 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.516691 4997 generic.go:334] "Generic (PLEG): container finished" podID="6e981f9f-0caf-490c-a9c2-eb6f0e7609ed" containerID="0ffc3605a015c9ad6c24e020ab1e1f8ab44c79a8a394bb795e83b562431f3bb9" exitCode=0 Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.516985 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9nlgw" event={"ID":"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed","Type":"ContainerDied","Data":"0ffc3605a015c9ad6c24e020ab1e1f8ab44c79a8a394bb795e83b562431f3bb9"} Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.517799 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9nlgw" event={"ID":"6e981f9f-0caf-490c-a9c2-eb6f0e7609ed","Type":"ContainerDied","Data":"998a30f8b3b33668d69a6e41ea89e3b5834f605b885c35c0f3b1747a0ce41ba5"} Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.517010 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9nlgw" Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.517825 4997 scope.go:117] "RemoveContainer" containerID="0ffc3605a015c9ad6c24e020ab1e1f8ab44c79a8a394bb795e83b562431f3bb9" Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.542804 4997 scope.go:117] "RemoveContainer" containerID="75d855f8293646353533a48c413e4568eb046f6d8e1b0643beb63691dc222ad5" Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.568505 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9nlgw"] Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.574281 4997 scope.go:117] "RemoveContainer" containerID="7ee6e92496bfa9ef055f436a6c168171eeab63623f10da7635d2718877c1ec84" Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.588675 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9nlgw"] Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.631770 4997 scope.go:117] "RemoveContainer" containerID="0ffc3605a015c9ad6c24e020ab1e1f8ab44c79a8a394bb795e83b562431f3bb9" Dec 05 10:13:37 crc kubenswrapper[4997]: E1205 10:13:37.632485 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ffc3605a015c9ad6c24e020ab1e1f8ab44c79a8a394bb795e83b562431f3bb9\": container with ID starting with 0ffc3605a015c9ad6c24e020ab1e1f8ab44c79a8a394bb795e83b562431f3bb9 not found: ID does not exist" containerID="0ffc3605a015c9ad6c24e020ab1e1f8ab44c79a8a394bb795e83b562431f3bb9" Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.632551 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ffc3605a015c9ad6c24e020ab1e1f8ab44c79a8a394bb795e83b562431f3bb9"} err="failed to get container status \"0ffc3605a015c9ad6c24e020ab1e1f8ab44c79a8a394bb795e83b562431f3bb9\": rpc error: code = NotFound desc = could not find container \"0ffc3605a015c9ad6c24e020ab1e1f8ab44c79a8a394bb795e83b562431f3bb9\": container with ID starting with 0ffc3605a015c9ad6c24e020ab1e1f8ab44c79a8a394bb795e83b562431f3bb9 not found: ID does not exist" Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.632582 4997 scope.go:117] "RemoveContainer" containerID="75d855f8293646353533a48c413e4568eb046f6d8e1b0643beb63691dc222ad5" Dec 05 10:13:37 crc kubenswrapper[4997]: E1205 10:13:37.633088 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75d855f8293646353533a48c413e4568eb046f6d8e1b0643beb63691dc222ad5\": container with ID starting with 75d855f8293646353533a48c413e4568eb046f6d8e1b0643beb63691dc222ad5 not found: ID does not exist" containerID="75d855f8293646353533a48c413e4568eb046f6d8e1b0643beb63691dc222ad5" Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.633113 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75d855f8293646353533a48c413e4568eb046f6d8e1b0643beb63691dc222ad5"} err="failed to get container status \"75d855f8293646353533a48c413e4568eb046f6d8e1b0643beb63691dc222ad5\": rpc error: code = NotFound desc = could not find container \"75d855f8293646353533a48c413e4568eb046f6d8e1b0643beb63691dc222ad5\": container with ID starting with 75d855f8293646353533a48c413e4568eb046f6d8e1b0643beb63691dc222ad5 not found: ID does not exist" Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.633127 4997 scope.go:117] "RemoveContainer" containerID="7ee6e92496bfa9ef055f436a6c168171eeab63623f10da7635d2718877c1ec84" Dec 05 10:13:37 crc kubenswrapper[4997]: E1205 10:13:37.633364 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ee6e92496bfa9ef055f436a6c168171eeab63623f10da7635d2718877c1ec84\": container with ID starting with 7ee6e92496bfa9ef055f436a6c168171eeab63623f10da7635d2718877c1ec84 not found: ID does not exist" containerID="7ee6e92496bfa9ef055f436a6c168171eeab63623f10da7635d2718877c1ec84" Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.633390 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ee6e92496bfa9ef055f436a6c168171eeab63623f10da7635d2718877c1ec84"} err="failed to get container status \"7ee6e92496bfa9ef055f436a6c168171eeab63623f10da7635d2718877c1ec84\": rpc error: code = NotFound desc = could not find container \"7ee6e92496bfa9ef055f436a6c168171eeab63623f10da7635d2718877c1ec84\": container with ID starting with 7ee6e92496bfa9ef055f436a6c168171eeab63623f10da7635d2718877c1ec84 not found: ID does not exist" Dec 05 10:13:37 crc kubenswrapper[4997]: I1205 10:13:37.758968 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e981f9f-0caf-490c-a9c2-eb6f0e7609ed" path="/var/lib/kubelet/pods/6e981f9f-0caf-490c-a9c2-eb6f0e7609ed/volumes" Dec 05 10:13:42 crc kubenswrapper[4997]: I1205 10:13:42.749026 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:13:42 crc kubenswrapper[4997]: E1205 10:13:42.749775 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:13:54 crc kubenswrapper[4997]: I1205 10:13:54.750021 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:13:54 crc kubenswrapper[4997]: E1205 10:13:54.750911 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:14:07 crc kubenswrapper[4997]: I1205 10:14:07.749289 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:14:07 crc kubenswrapper[4997]: E1205 10:14:07.750113 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:14:22 crc kubenswrapper[4997]: I1205 10:14:22.749820 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:14:22 crc kubenswrapper[4997]: E1205 10:14:22.750593 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:14:33 crc kubenswrapper[4997]: I1205 10:14:33.339756 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kg44g"] Dec 05 10:14:33 crc kubenswrapper[4997]: E1205 10:14:33.340945 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e981f9f-0caf-490c-a9c2-eb6f0e7609ed" containerName="registry-server" Dec 05 10:14:33 crc kubenswrapper[4997]: I1205 10:14:33.340958 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e981f9f-0caf-490c-a9c2-eb6f0e7609ed" containerName="registry-server" Dec 05 10:14:33 crc kubenswrapper[4997]: E1205 10:14:33.340979 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e981f9f-0caf-490c-a9c2-eb6f0e7609ed" containerName="extract-utilities" Dec 05 10:14:33 crc kubenswrapper[4997]: I1205 10:14:33.340986 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e981f9f-0caf-490c-a9c2-eb6f0e7609ed" containerName="extract-utilities" Dec 05 10:14:33 crc kubenswrapper[4997]: E1205 10:14:33.341009 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e981f9f-0caf-490c-a9c2-eb6f0e7609ed" containerName="extract-content" Dec 05 10:14:33 crc kubenswrapper[4997]: I1205 10:14:33.341015 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e981f9f-0caf-490c-a9c2-eb6f0e7609ed" containerName="extract-content" Dec 05 10:14:33 crc kubenswrapper[4997]: I1205 10:14:33.341229 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e981f9f-0caf-490c-a9c2-eb6f0e7609ed" containerName="registry-server" Dec 05 10:14:33 crc kubenswrapper[4997]: I1205 10:14:33.342869 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kg44g" Dec 05 10:14:33 crc kubenswrapper[4997]: I1205 10:14:33.385827 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kg44g"] Dec 05 10:14:33 crc kubenswrapper[4997]: I1205 10:14:33.476681 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c94733a3-04d7-4b33-a4a3-1d4375071f2f-catalog-content\") pod \"redhat-operators-kg44g\" (UID: \"c94733a3-04d7-4b33-a4a3-1d4375071f2f\") " pod="openshift-marketplace/redhat-operators-kg44g" Dec 05 10:14:33 crc kubenswrapper[4997]: I1205 10:14:33.476757 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzfsc\" (UniqueName: \"kubernetes.io/projected/c94733a3-04d7-4b33-a4a3-1d4375071f2f-kube-api-access-gzfsc\") pod \"redhat-operators-kg44g\" (UID: \"c94733a3-04d7-4b33-a4a3-1d4375071f2f\") " pod="openshift-marketplace/redhat-operators-kg44g" Dec 05 10:14:33 crc kubenswrapper[4997]: I1205 10:14:33.476891 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c94733a3-04d7-4b33-a4a3-1d4375071f2f-utilities\") pod \"redhat-operators-kg44g\" (UID: \"c94733a3-04d7-4b33-a4a3-1d4375071f2f\") " pod="openshift-marketplace/redhat-operators-kg44g" Dec 05 10:14:33 crc kubenswrapper[4997]: I1205 10:14:33.579333 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c94733a3-04d7-4b33-a4a3-1d4375071f2f-catalog-content\") pod \"redhat-operators-kg44g\" (UID: \"c94733a3-04d7-4b33-a4a3-1d4375071f2f\") " pod="openshift-marketplace/redhat-operators-kg44g" Dec 05 10:14:33 crc kubenswrapper[4997]: I1205 10:14:33.579413 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzfsc\" (UniqueName: \"kubernetes.io/projected/c94733a3-04d7-4b33-a4a3-1d4375071f2f-kube-api-access-gzfsc\") pod \"redhat-operators-kg44g\" (UID: \"c94733a3-04d7-4b33-a4a3-1d4375071f2f\") " pod="openshift-marketplace/redhat-operators-kg44g" Dec 05 10:14:33 crc kubenswrapper[4997]: I1205 10:14:33.579447 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c94733a3-04d7-4b33-a4a3-1d4375071f2f-utilities\") pod \"redhat-operators-kg44g\" (UID: \"c94733a3-04d7-4b33-a4a3-1d4375071f2f\") " pod="openshift-marketplace/redhat-operators-kg44g" Dec 05 10:14:33 crc kubenswrapper[4997]: I1205 10:14:33.580246 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c94733a3-04d7-4b33-a4a3-1d4375071f2f-utilities\") pod \"redhat-operators-kg44g\" (UID: \"c94733a3-04d7-4b33-a4a3-1d4375071f2f\") " pod="openshift-marketplace/redhat-operators-kg44g" Dec 05 10:14:33 crc kubenswrapper[4997]: I1205 10:14:33.580564 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c94733a3-04d7-4b33-a4a3-1d4375071f2f-catalog-content\") pod \"redhat-operators-kg44g\" (UID: \"c94733a3-04d7-4b33-a4a3-1d4375071f2f\") " pod="openshift-marketplace/redhat-operators-kg44g" Dec 05 10:14:33 crc kubenswrapper[4997]: I1205 10:14:33.624791 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzfsc\" (UniqueName: \"kubernetes.io/projected/c94733a3-04d7-4b33-a4a3-1d4375071f2f-kube-api-access-gzfsc\") pod \"redhat-operators-kg44g\" (UID: \"c94733a3-04d7-4b33-a4a3-1d4375071f2f\") " pod="openshift-marketplace/redhat-operators-kg44g" Dec 05 10:14:33 crc kubenswrapper[4997]: I1205 10:14:33.686826 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kg44g" Dec 05 10:14:34 crc kubenswrapper[4997]: I1205 10:14:34.311066 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kg44g"] Dec 05 10:14:34 crc kubenswrapper[4997]: W1205 10:14:34.341550 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc94733a3_04d7_4b33_a4a3_1d4375071f2f.slice/crio-2da837d3090b261937a1d5b5958eba76933637babe655e1266e06cc989e9e365 WatchSource:0}: Error finding container 2da837d3090b261937a1d5b5958eba76933637babe655e1266e06cc989e9e365: Status 404 returned error can't find the container with id 2da837d3090b261937a1d5b5958eba76933637babe655e1266e06cc989e9e365 Dec 05 10:14:35 crc kubenswrapper[4997]: I1205 10:14:35.215662 4997 generic.go:334] "Generic (PLEG): container finished" podID="c94733a3-04d7-4b33-a4a3-1d4375071f2f" containerID="bd8929a6cb5166a29898fb1198376c9c946b06878d600bcde1068f1b7703ebcc" exitCode=0 Dec 05 10:14:35 crc kubenswrapper[4997]: I1205 10:14:35.215724 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg44g" event={"ID":"c94733a3-04d7-4b33-a4a3-1d4375071f2f","Type":"ContainerDied","Data":"bd8929a6cb5166a29898fb1198376c9c946b06878d600bcde1068f1b7703ebcc"} Dec 05 10:14:35 crc kubenswrapper[4997]: I1205 10:14:35.215943 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg44g" event={"ID":"c94733a3-04d7-4b33-a4a3-1d4375071f2f","Type":"ContainerStarted","Data":"2da837d3090b261937a1d5b5958eba76933637babe655e1266e06cc989e9e365"} Dec 05 10:14:37 crc kubenswrapper[4997]: I1205 10:14:37.750300 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:14:37 crc kubenswrapper[4997]: E1205 10:14:37.751331 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:14:38 crc kubenswrapper[4997]: I1205 10:14:38.245555 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg44g" event={"ID":"c94733a3-04d7-4b33-a4a3-1d4375071f2f","Type":"ContainerStarted","Data":"5ddfe5f7eb08e58e9fc8b16903d497587c6e7e2aebe8a957c5837d8fbe9f3104"} Dec 05 10:14:41 crc kubenswrapper[4997]: I1205 10:14:41.279808 4997 generic.go:334] "Generic (PLEG): container finished" podID="c94733a3-04d7-4b33-a4a3-1d4375071f2f" containerID="5ddfe5f7eb08e58e9fc8b16903d497587c6e7e2aebe8a957c5837d8fbe9f3104" exitCode=0 Dec 05 10:14:41 crc kubenswrapper[4997]: I1205 10:14:41.279877 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg44g" event={"ID":"c94733a3-04d7-4b33-a4a3-1d4375071f2f","Type":"ContainerDied","Data":"5ddfe5f7eb08e58e9fc8b16903d497587c6e7e2aebe8a957c5837d8fbe9f3104"} Dec 05 10:14:44 crc kubenswrapper[4997]: I1205 10:14:44.323241 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg44g" event={"ID":"c94733a3-04d7-4b33-a4a3-1d4375071f2f","Type":"ContainerStarted","Data":"f973a2e8d3028b52304717ee5965d93f2a6e2abd2db75c9dd77cb727d147a783"} Dec 05 10:14:44 crc kubenswrapper[4997]: I1205 10:14:44.349938 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kg44g" podStartSLOduration=3.400698514 podStartE2EDuration="11.349907682s" podCreationTimestamp="2025-12-05 10:14:33 +0000 UTC" firstStartedPulling="2025-12-05 10:14:35.217189761 +0000 UTC m=+11975.746097022" lastFinishedPulling="2025-12-05 10:14:43.166398929 +0000 UTC m=+11983.695306190" observedRunningTime="2025-12-05 10:14:44.341173577 +0000 UTC m=+11984.870080868" watchObservedRunningTime="2025-12-05 10:14:44.349907682 +0000 UTC m=+11984.878814943" Dec 05 10:14:49 crc kubenswrapper[4997]: I1205 10:14:49.758064 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:14:49 crc kubenswrapper[4997]: E1205 10:14:49.758869 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:14:53 crc kubenswrapper[4997]: I1205 10:14:53.687850 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kg44g" Dec 05 10:14:53 crc kubenswrapper[4997]: I1205 10:14:53.688675 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kg44g" Dec 05 10:14:53 crc kubenswrapper[4997]: I1205 10:14:53.744486 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kg44g" Dec 05 10:14:54 crc kubenswrapper[4997]: I1205 10:14:54.477113 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kg44g" Dec 05 10:14:54 crc kubenswrapper[4997]: I1205 10:14:54.530998 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kg44g"] Dec 05 10:14:56 crc kubenswrapper[4997]: I1205 10:14:56.455417 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kg44g" podUID="c94733a3-04d7-4b33-a4a3-1d4375071f2f" containerName="registry-server" containerID="cri-o://f973a2e8d3028b52304717ee5965d93f2a6e2abd2db75c9dd77cb727d147a783" gracePeriod=2 Dec 05 10:14:57 crc kubenswrapper[4997]: I1205 10:14:57.476507 4997 generic.go:334] "Generic (PLEG): container finished" podID="c94733a3-04d7-4b33-a4a3-1d4375071f2f" containerID="f973a2e8d3028b52304717ee5965d93f2a6e2abd2db75c9dd77cb727d147a783" exitCode=0 Dec 05 10:14:57 crc kubenswrapper[4997]: I1205 10:14:57.476946 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg44g" event={"ID":"c94733a3-04d7-4b33-a4a3-1d4375071f2f","Type":"ContainerDied","Data":"f973a2e8d3028b52304717ee5965d93f2a6e2abd2db75c9dd77cb727d147a783"} Dec 05 10:14:57 crc kubenswrapper[4997]: I1205 10:14:57.765722 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kg44g" Dec 05 10:14:57 crc kubenswrapper[4997]: I1205 10:14:57.907794 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c94733a3-04d7-4b33-a4a3-1d4375071f2f-catalog-content\") pod \"c94733a3-04d7-4b33-a4a3-1d4375071f2f\" (UID: \"c94733a3-04d7-4b33-a4a3-1d4375071f2f\") " Dec 05 10:14:57 crc kubenswrapper[4997]: I1205 10:14:57.908446 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzfsc\" (UniqueName: \"kubernetes.io/projected/c94733a3-04d7-4b33-a4a3-1d4375071f2f-kube-api-access-gzfsc\") pod \"c94733a3-04d7-4b33-a4a3-1d4375071f2f\" (UID: \"c94733a3-04d7-4b33-a4a3-1d4375071f2f\") " Dec 05 10:14:57 crc kubenswrapper[4997]: I1205 10:14:57.908486 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c94733a3-04d7-4b33-a4a3-1d4375071f2f-utilities\") pod \"c94733a3-04d7-4b33-a4a3-1d4375071f2f\" (UID: \"c94733a3-04d7-4b33-a4a3-1d4375071f2f\") " Dec 05 10:14:57 crc kubenswrapper[4997]: I1205 10:14:57.910072 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c94733a3-04d7-4b33-a4a3-1d4375071f2f-utilities" (OuterVolumeSpecName: "utilities") pod "c94733a3-04d7-4b33-a4a3-1d4375071f2f" (UID: "c94733a3-04d7-4b33-a4a3-1d4375071f2f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:14:57 crc kubenswrapper[4997]: I1205 10:14:57.933900 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c94733a3-04d7-4b33-a4a3-1d4375071f2f-kube-api-access-gzfsc" (OuterVolumeSpecName: "kube-api-access-gzfsc") pod "c94733a3-04d7-4b33-a4a3-1d4375071f2f" (UID: "c94733a3-04d7-4b33-a4a3-1d4375071f2f"). InnerVolumeSpecName "kube-api-access-gzfsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:14:58 crc kubenswrapper[4997]: I1205 10:14:58.011557 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzfsc\" (UniqueName: \"kubernetes.io/projected/c94733a3-04d7-4b33-a4a3-1d4375071f2f-kube-api-access-gzfsc\") on node \"crc\" DevicePath \"\"" Dec 05 10:14:58 crc kubenswrapper[4997]: I1205 10:14:58.011596 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c94733a3-04d7-4b33-a4a3-1d4375071f2f-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:14:58 crc kubenswrapper[4997]: I1205 10:14:58.031645 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c94733a3-04d7-4b33-a4a3-1d4375071f2f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c94733a3-04d7-4b33-a4a3-1d4375071f2f" (UID: "c94733a3-04d7-4b33-a4a3-1d4375071f2f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:14:58 crc kubenswrapper[4997]: I1205 10:14:58.113873 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c94733a3-04d7-4b33-a4a3-1d4375071f2f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:14:58 crc kubenswrapper[4997]: I1205 10:14:58.491494 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg44g" event={"ID":"c94733a3-04d7-4b33-a4a3-1d4375071f2f","Type":"ContainerDied","Data":"2da837d3090b261937a1d5b5958eba76933637babe655e1266e06cc989e9e365"} Dec 05 10:14:58 crc kubenswrapper[4997]: I1205 10:14:58.491556 4997 scope.go:117] "RemoveContainer" containerID="f973a2e8d3028b52304717ee5965d93f2a6e2abd2db75c9dd77cb727d147a783" Dec 05 10:14:58 crc kubenswrapper[4997]: I1205 10:14:58.491636 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kg44g" Dec 05 10:14:58 crc kubenswrapper[4997]: I1205 10:14:58.521166 4997 scope.go:117] "RemoveContainer" containerID="5ddfe5f7eb08e58e9fc8b16903d497587c6e7e2aebe8a957c5837d8fbe9f3104" Dec 05 10:14:58 crc kubenswrapper[4997]: I1205 10:14:58.540791 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kg44g"] Dec 05 10:14:58 crc kubenswrapper[4997]: I1205 10:14:58.584417 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kg44g"] Dec 05 10:14:58 crc kubenswrapper[4997]: I1205 10:14:58.596453 4997 scope.go:117] "RemoveContainer" containerID="bd8929a6cb5166a29898fb1198376c9c946b06878d600bcde1068f1b7703ebcc" Dec 05 10:14:59 crc kubenswrapper[4997]: I1205 10:14:59.766387 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c94733a3-04d7-4b33-a4a3-1d4375071f2f" path="/var/lib/kubelet/pods/c94733a3-04d7-4b33-a4a3-1d4375071f2f/volumes" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.179353 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5"] Dec 05 10:15:00 crc kubenswrapper[4997]: E1205 10:15:00.180470 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94733a3-04d7-4b33-a4a3-1d4375071f2f" containerName="extract-content" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.180555 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94733a3-04d7-4b33-a4a3-1d4375071f2f" containerName="extract-content" Dec 05 10:15:00 crc kubenswrapper[4997]: E1205 10:15:00.180656 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94733a3-04d7-4b33-a4a3-1d4375071f2f" containerName="registry-server" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.180791 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94733a3-04d7-4b33-a4a3-1d4375071f2f" containerName="registry-server" Dec 05 10:15:00 crc kubenswrapper[4997]: E1205 10:15:00.180917 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94733a3-04d7-4b33-a4a3-1d4375071f2f" containerName="extract-utilities" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.180973 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94733a3-04d7-4b33-a4a3-1d4375071f2f" containerName="extract-utilities" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.181317 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="c94733a3-04d7-4b33-a4a3-1d4375071f2f" containerName="registry-server" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.182387 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.184676 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.185413 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.206699 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5"] Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.367786 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-secret-volume\") pod \"collect-profiles-29415495-pz9h5\" (UID: \"565e912e-b446-46c4-8f7c-e5aec0fa8cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.368391 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg8ft\" (UniqueName: \"kubernetes.io/projected/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-kube-api-access-qg8ft\") pod \"collect-profiles-29415495-pz9h5\" (UID: \"565e912e-b446-46c4-8f7c-e5aec0fa8cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.368455 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-config-volume\") pod \"collect-profiles-29415495-pz9h5\" (UID: \"565e912e-b446-46c4-8f7c-e5aec0fa8cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.470030 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-config-volume\") pod \"collect-profiles-29415495-pz9h5\" (UID: \"565e912e-b446-46c4-8f7c-e5aec0fa8cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.470197 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-secret-volume\") pod \"collect-profiles-29415495-pz9h5\" (UID: \"565e912e-b446-46c4-8f7c-e5aec0fa8cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.470251 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg8ft\" (UniqueName: \"kubernetes.io/projected/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-kube-api-access-qg8ft\") pod \"collect-profiles-29415495-pz9h5\" (UID: \"565e912e-b446-46c4-8f7c-e5aec0fa8cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.471402 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-config-volume\") pod \"collect-profiles-29415495-pz9h5\" (UID: \"565e912e-b446-46c4-8f7c-e5aec0fa8cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.478135 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-secret-volume\") pod \"collect-profiles-29415495-pz9h5\" (UID: \"565e912e-b446-46c4-8f7c-e5aec0fa8cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.489695 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg8ft\" (UniqueName: \"kubernetes.io/projected/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-kube-api-access-qg8ft\") pod \"collect-profiles-29415495-pz9h5\" (UID: \"565e912e-b446-46c4-8f7c-e5aec0fa8cc6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.518227 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" Dec 05 10:15:00 crc kubenswrapper[4997]: I1205 10:15:00.749467 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:15:00 crc kubenswrapper[4997]: E1205 10:15:00.749987 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:15:01 crc kubenswrapper[4997]: W1205 10:15:01.033558 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod565e912e_b446_46c4_8f7c_e5aec0fa8cc6.slice/crio-cfd0c01a9f76dd8100895cb5fa679cc8a76db3fe4fa33080f0876bf0b5e5edf4 WatchSource:0}: Error finding container cfd0c01a9f76dd8100895cb5fa679cc8a76db3fe4fa33080f0876bf0b5e5edf4: Status 404 returned error can't find the container with id cfd0c01a9f76dd8100895cb5fa679cc8a76db3fe4fa33080f0876bf0b5e5edf4 Dec 05 10:15:01 crc kubenswrapper[4997]: I1205 10:15:01.035973 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5"] Dec 05 10:15:01 crc kubenswrapper[4997]: I1205 10:15:01.534905 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" event={"ID":"565e912e-b446-46c4-8f7c-e5aec0fa8cc6","Type":"ContainerStarted","Data":"15c5a2ecbfa474aa160160f299d2f81b84729cbe0456686665c95e5639288462"} Dec 05 10:15:01 crc kubenswrapper[4997]: I1205 10:15:01.535178 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" event={"ID":"565e912e-b446-46c4-8f7c-e5aec0fa8cc6","Type":"ContainerStarted","Data":"cfd0c01a9f76dd8100895cb5fa679cc8a76db3fe4fa33080f0876bf0b5e5edf4"} Dec 05 10:15:01 crc kubenswrapper[4997]: I1205 10:15:01.570990 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" podStartSLOduration=1.570957468 podStartE2EDuration="1.570957468s" podCreationTimestamp="2025-12-05 10:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:15:01.550369082 +0000 UTC m=+12002.079276363" watchObservedRunningTime="2025-12-05 10:15:01.570957468 +0000 UTC m=+12002.099864739" Dec 05 10:15:02 crc kubenswrapper[4997]: I1205 10:15:02.545441 4997 generic.go:334] "Generic (PLEG): container finished" podID="565e912e-b446-46c4-8f7c-e5aec0fa8cc6" containerID="15c5a2ecbfa474aa160160f299d2f81b84729cbe0456686665c95e5639288462" exitCode=0 Dec 05 10:15:02 crc kubenswrapper[4997]: I1205 10:15:02.546003 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" event={"ID":"565e912e-b446-46c4-8f7c-e5aec0fa8cc6","Type":"ContainerDied","Data":"15c5a2ecbfa474aa160160f299d2f81b84729cbe0456686665c95e5639288462"} Dec 05 10:15:04 crc kubenswrapper[4997]: I1205 10:15:04.553474 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" Dec 05 10:15:04 crc kubenswrapper[4997]: I1205 10:15:04.568527 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" event={"ID":"565e912e-b446-46c4-8f7c-e5aec0fa8cc6","Type":"ContainerDied","Data":"cfd0c01a9f76dd8100895cb5fa679cc8a76db3fe4fa33080f0876bf0b5e5edf4"} Dec 05 10:15:04 crc kubenswrapper[4997]: I1205 10:15:04.568593 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfd0c01a9f76dd8100895cb5fa679cc8a76db3fe4fa33080f0876bf0b5e5edf4" Dec 05 10:15:04 crc kubenswrapper[4997]: I1205 10:15:04.568604 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-pz9h5" Dec 05 10:15:04 crc kubenswrapper[4997]: I1205 10:15:04.664232 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-config-volume\") pod \"565e912e-b446-46c4-8f7c-e5aec0fa8cc6\" (UID: \"565e912e-b446-46c4-8f7c-e5aec0fa8cc6\") " Dec 05 10:15:04 crc kubenswrapper[4997]: I1205 10:15:04.664452 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg8ft\" (UniqueName: \"kubernetes.io/projected/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-kube-api-access-qg8ft\") pod \"565e912e-b446-46c4-8f7c-e5aec0fa8cc6\" (UID: \"565e912e-b446-46c4-8f7c-e5aec0fa8cc6\") " Dec 05 10:15:04 crc kubenswrapper[4997]: I1205 10:15:04.664522 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-secret-volume\") pod \"565e912e-b446-46c4-8f7c-e5aec0fa8cc6\" (UID: \"565e912e-b446-46c4-8f7c-e5aec0fa8cc6\") " Dec 05 10:15:04 crc kubenswrapper[4997]: I1205 10:15:04.665860 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-config-volume" (OuterVolumeSpecName: "config-volume") pod "565e912e-b446-46c4-8f7c-e5aec0fa8cc6" (UID: "565e912e-b446-46c4-8f7c-e5aec0fa8cc6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:15:04 crc kubenswrapper[4997]: I1205 10:15:04.671481 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-kube-api-access-qg8ft" (OuterVolumeSpecName: "kube-api-access-qg8ft") pod "565e912e-b446-46c4-8f7c-e5aec0fa8cc6" (UID: "565e912e-b446-46c4-8f7c-e5aec0fa8cc6"). InnerVolumeSpecName "kube-api-access-qg8ft". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:15:04 crc kubenswrapper[4997]: I1205 10:15:04.672018 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "565e912e-b446-46c4-8f7c-e5aec0fa8cc6" (UID: "565e912e-b446-46c4-8f7c-e5aec0fa8cc6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:15:04 crc kubenswrapper[4997]: I1205 10:15:04.766813 4997 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 10:15:04 crc kubenswrapper[4997]: I1205 10:15:04.766866 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg8ft\" (UniqueName: \"kubernetes.io/projected/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-kube-api-access-qg8ft\") on node \"crc\" DevicePath \"\"" Dec 05 10:15:04 crc kubenswrapper[4997]: I1205 10:15:04.766884 4997 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/565e912e-b446-46c4-8f7c-e5aec0fa8cc6-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 10:15:05 crc kubenswrapper[4997]: I1205 10:15:05.662909 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc"] Dec 05 10:15:05 crc kubenswrapper[4997]: I1205 10:15:05.676323 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415450-9tsfc"] Dec 05 10:15:05 crc kubenswrapper[4997]: I1205 10:15:05.763563 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="115f37ed-c40f-4655-986f-a5313b62cf1e" path="/var/lib/kubelet/pods/115f37ed-c40f-4655-986f-a5313b62cf1e/volumes" Dec 05 10:15:15 crc kubenswrapper[4997]: I1205 10:15:15.750073 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:15:15 crc kubenswrapper[4997]: E1205 10:15:15.751052 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:15:21 crc kubenswrapper[4997]: I1205 10:15:21.759517 4997 scope.go:117] "RemoveContainer" containerID="37a5df49d0f5a50fd25a0baa67fc12f249c5b5565b01ea4439e4afb29b3c1b88" Dec 05 10:15:28 crc kubenswrapper[4997]: I1205 10:15:28.748958 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:15:28 crc kubenswrapper[4997]: E1205 10:15:28.749727 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:15:40 crc kubenswrapper[4997]: I1205 10:15:40.749802 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:15:40 crc kubenswrapper[4997]: E1205 10:15:40.750643 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:15:52 crc kubenswrapper[4997]: I1205 10:15:52.750374 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:15:53 crc kubenswrapper[4997]: I1205 10:15:53.136686 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"8a987de1021e7c3842ed4e7f92ee76fff33696ade660bfaad3dd390f94c1a7e6"} Dec 05 10:18:19 crc kubenswrapper[4997]: I1205 10:18:19.781454 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:18:19 crc kubenswrapper[4997]: I1205 10:18:19.782279 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:18:49 crc kubenswrapper[4997]: I1205 10:18:49.770530 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:18:49 crc kubenswrapper[4997]: I1205 10:18:49.771172 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:19:19 crc kubenswrapper[4997]: I1205 10:19:19.770060 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:19:19 crc kubenswrapper[4997]: I1205 10:19:19.770628 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:19:19 crc kubenswrapper[4997]: I1205 10:19:19.770679 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 10:19:19 crc kubenswrapper[4997]: I1205 10:19:19.771517 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8a987de1021e7c3842ed4e7f92ee76fff33696ade660bfaad3dd390f94c1a7e6"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 10:19:19 crc kubenswrapper[4997]: I1205 10:19:19.771566 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://8a987de1021e7c3842ed4e7f92ee76fff33696ade660bfaad3dd390f94c1a7e6" gracePeriod=600 Dec 05 10:19:20 crc kubenswrapper[4997]: I1205 10:19:20.491634 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="8a987de1021e7c3842ed4e7f92ee76fff33696ade660bfaad3dd390f94c1a7e6" exitCode=0 Dec 05 10:19:20 crc kubenswrapper[4997]: I1205 10:19:20.491728 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"8a987de1021e7c3842ed4e7f92ee76fff33696ade660bfaad3dd390f94c1a7e6"} Dec 05 10:19:20 crc kubenswrapper[4997]: I1205 10:19:20.492184 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc"} Dec 05 10:19:20 crc kubenswrapper[4997]: I1205 10:19:20.492206 4997 scope.go:117] "RemoveContainer" containerID="330e713f457a3a9223d8544b116927a1aa1e0324eeec01e3bd3267787be68d24" Dec 05 10:20:32 crc kubenswrapper[4997]: I1205 10:20:32.261394 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-795cw"] Dec 05 10:20:32 crc kubenswrapper[4997]: E1205 10:20:32.262550 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565e912e-b446-46c4-8f7c-e5aec0fa8cc6" containerName="collect-profiles" Dec 05 10:20:32 crc kubenswrapper[4997]: I1205 10:20:32.262573 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="565e912e-b446-46c4-8f7c-e5aec0fa8cc6" containerName="collect-profiles" Dec 05 10:20:32 crc kubenswrapper[4997]: I1205 10:20:32.262892 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="565e912e-b446-46c4-8f7c-e5aec0fa8cc6" containerName="collect-profiles" Dec 05 10:20:32 crc kubenswrapper[4997]: I1205 10:20:32.265185 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-795cw" Dec 05 10:20:32 crc kubenswrapper[4997]: I1205 10:20:32.273913 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-795cw"] Dec 05 10:20:32 crc kubenswrapper[4997]: I1205 10:20:32.379515 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hml8r\" (UniqueName: \"kubernetes.io/projected/4bbe7e5f-5081-4b98-abbd-259006366891-kube-api-access-hml8r\") pod \"certified-operators-795cw\" (UID: \"4bbe7e5f-5081-4b98-abbd-259006366891\") " pod="openshift-marketplace/certified-operators-795cw" Dec 05 10:20:32 crc kubenswrapper[4997]: I1205 10:20:32.379719 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bbe7e5f-5081-4b98-abbd-259006366891-catalog-content\") pod \"certified-operators-795cw\" (UID: \"4bbe7e5f-5081-4b98-abbd-259006366891\") " pod="openshift-marketplace/certified-operators-795cw" Dec 05 10:20:32 crc kubenswrapper[4997]: I1205 10:20:32.379806 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bbe7e5f-5081-4b98-abbd-259006366891-utilities\") pod \"certified-operators-795cw\" (UID: \"4bbe7e5f-5081-4b98-abbd-259006366891\") " pod="openshift-marketplace/certified-operators-795cw" Dec 05 10:20:32 crc kubenswrapper[4997]: I1205 10:20:32.483768 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hml8r\" (UniqueName: \"kubernetes.io/projected/4bbe7e5f-5081-4b98-abbd-259006366891-kube-api-access-hml8r\") pod \"certified-operators-795cw\" (UID: \"4bbe7e5f-5081-4b98-abbd-259006366891\") " pod="openshift-marketplace/certified-operators-795cw" Dec 05 10:20:32 crc kubenswrapper[4997]: I1205 10:20:32.483973 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bbe7e5f-5081-4b98-abbd-259006366891-catalog-content\") pod \"certified-operators-795cw\" (UID: \"4bbe7e5f-5081-4b98-abbd-259006366891\") " pod="openshift-marketplace/certified-operators-795cw" Dec 05 10:20:32 crc kubenswrapper[4997]: I1205 10:20:32.484121 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bbe7e5f-5081-4b98-abbd-259006366891-utilities\") pod \"certified-operators-795cw\" (UID: \"4bbe7e5f-5081-4b98-abbd-259006366891\") " pod="openshift-marketplace/certified-operators-795cw" Dec 05 10:20:32 crc kubenswrapper[4997]: I1205 10:20:32.484875 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bbe7e5f-5081-4b98-abbd-259006366891-catalog-content\") pod \"certified-operators-795cw\" (UID: \"4bbe7e5f-5081-4b98-abbd-259006366891\") " pod="openshift-marketplace/certified-operators-795cw" Dec 05 10:20:32 crc kubenswrapper[4997]: I1205 10:20:32.484898 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bbe7e5f-5081-4b98-abbd-259006366891-utilities\") pod \"certified-operators-795cw\" (UID: \"4bbe7e5f-5081-4b98-abbd-259006366891\") " pod="openshift-marketplace/certified-operators-795cw" Dec 05 10:20:32 crc kubenswrapper[4997]: I1205 10:20:32.510905 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hml8r\" (UniqueName: \"kubernetes.io/projected/4bbe7e5f-5081-4b98-abbd-259006366891-kube-api-access-hml8r\") pod \"certified-operators-795cw\" (UID: \"4bbe7e5f-5081-4b98-abbd-259006366891\") " pod="openshift-marketplace/certified-operators-795cw" Dec 05 10:20:32 crc kubenswrapper[4997]: I1205 10:20:32.586731 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-795cw" Dec 05 10:20:33 crc kubenswrapper[4997]: I1205 10:20:33.202118 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-795cw"] Dec 05 10:20:33 crc kubenswrapper[4997]: I1205 10:20:33.331379 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-795cw" event={"ID":"4bbe7e5f-5081-4b98-abbd-259006366891","Type":"ContainerStarted","Data":"a9d371ab232869ed9de7a87a950cf4901a3327ad1c46e48ef62d8d3e394aefbf"} Dec 05 10:20:34 crc kubenswrapper[4997]: I1205 10:20:34.343787 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-795cw" event={"ID":"4bbe7e5f-5081-4b98-abbd-259006366891","Type":"ContainerStarted","Data":"0901005ff3f8c8db13fa413adb3fc92ec08b2b930ab83761d3b234343a257572"} Dec 05 10:20:34 crc kubenswrapper[4997]: I1205 10:20:34.345650 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:20:35 crc kubenswrapper[4997]: I1205 10:20:35.360751 4997 generic.go:334] "Generic (PLEG): container finished" podID="4bbe7e5f-5081-4b98-abbd-259006366891" containerID="0901005ff3f8c8db13fa413adb3fc92ec08b2b930ab83761d3b234343a257572" exitCode=0 Dec 05 10:20:35 crc kubenswrapper[4997]: I1205 10:20:35.360823 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-795cw" event={"ID":"4bbe7e5f-5081-4b98-abbd-259006366891","Type":"ContainerDied","Data":"0901005ff3f8c8db13fa413adb3fc92ec08b2b930ab83761d3b234343a257572"} Dec 05 10:20:36 crc kubenswrapper[4997]: I1205 10:20:36.378805 4997 generic.go:334] "Generic (PLEG): container finished" podID="4bbe7e5f-5081-4b98-abbd-259006366891" containerID="52b6182c29f20be7ec2d7f9741aa2b054e3d0bc5417d1b0cbb96d35467578d5b" exitCode=0 Dec 05 10:20:36 crc kubenswrapper[4997]: I1205 10:20:36.379052 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-795cw" event={"ID":"4bbe7e5f-5081-4b98-abbd-259006366891","Type":"ContainerDied","Data":"52b6182c29f20be7ec2d7f9741aa2b054e3d0bc5417d1b0cbb96d35467578d5b"} Dec 05 10:20:38 crc kubenswrapper[4997]: I1205 10:20:38.404658 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-795cw" event={"ID":"4bbe7e5f-5081-4b98-abbd-259006366891","Type":"ContainerStarted","Data":"635821dab0ae02edc0aa2f3a853371c56ec1a7c02751358d5fb555c27729f8a4"} Dec 05 10:20:38 crc kubenswrapper[4997]: I1205 10:20:38.429428 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-795cw" podStartSLOduration=3.07561687 podStartE2EDuration="6.429396837s" podCreationTimestamp="2025-12-05 10:20:32 +0000 UTC" firstStartedPulling="2025-12-05 10:20:34.345431033 +0000 UTC m=+12334.874338294" lastFinishedPulling="2025-12-05 10:20:37.699211 +0000 UTC m=+12338.228118261" observedRunningTime="2025-12-05 10:20:38.42173665 +0000 UTC m=+12338.950643911" watchObservedRunningTime="2025-12-05 10:20:38.429396837 +0000 UTC m=+12338.958304108" Dec 05 10:20:42 crc kubenswrapper[4997]: I1205 10:20:42.586846 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-795cw" Dec 05 10:20:42 crc kubenswrapper[4997]: I1205 10:20:42.587379 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-795cw" Dec 05 10:20:42 crc kubenswrapper[4997]: I1205 10:20:42.640650 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-795cw" Dec 05 10:20:43 crc kubenswrapper[4997]: I1205 10:20:43.508573 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-795cw" Dec 05 10:20:43 crc kubenswrapper[4997]: I1205 10:20:43.572756 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-795cw"] Dec 05 10:20:45 crc kubenswrapper[4997]: I1205 10:20:45.480012 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-795cw" podUID="4bbe7e5f-5081-4b98-abbd-259006366891" containerName="registry-server" containerID="cri-o://635821dab0ae02edc0aa2f3a853371c56ec1a7c02751358d5fb555c27729f8a4" gracePeriod=2 Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.135325 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-795cw" Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.283180 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bbe7e5f-5081-4b98-abbd-259006366891-utilities\") pod \"4bbe7e5f-5081-4b98-abbd-259006366891\" (UID: \"4bbe7e5f-5081-4b98-abbd-259006366891\") " Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.283218 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hml8r\" (UniqueName: \"kubernetes.io/projected/4bbe7e5f-5081-4b98-abbd-259006366891-kube-api-access-hml8r\") pod \"4bbe7e5f-5081-4b98-abbd-259006366891\" (UID: \"4bbe7e5f-5081-4b98-abbd-259006366891\") " Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.283243 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bbe7e5f-5081-4b98-abbd-259006366891-catalog-content\") pod \"4bbe7e5f-5081-4b98-abbd-259006366891\" (UID: \"4bbe7e5f-5081-4b98-abbd-259006366891\") " Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.284291 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bbe7e5f-5081-4b98-abbd-259006366891-utilities" (OuterVolumeSpecName: "utilities") pod "4bbe7e5f-5081-4b98-abbd-259006366891" (UID: "4bbe7e5f-5081-4b98-abbd-259006366891"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.289983 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bbe7e5f-5081-4b98-abbd-259006366891-kube-api-access-hml8r" (OuterVolumeSpecName: "kube-api-access-hml8r") pod "4bbe7e5f-5081-4b98-abbd-259006366891" (UID: "4bbe7e5f-5081-4b98-abbd-259006366891"). InnerVolumeSpecName "kube-api-access-hml8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.386661 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hml8r\" (UniqueName: \"kubernetes.io/projected/4bbe7e5f-5081-4b98-abbd-259006366891-kube-api-access-hml8r\") on node \"crc\" DevicePath \"\"" Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.386718 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bbe7e5f-5081-4b98-abbd-259006366891-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.494456 4997 generic.go:334] "Generic (PLEG): container finished" podID="4bbe7e5f-5081-4b98-abbd-259006366891" containerID="635821dab0ae02edc0aa2f3a853371c56ec1a7c02751358d5fb555c27729f8a4" exitCode=0 Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.494508 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-795cw" event={"ID":"4bbe7e5f-5081-4b98-abbd-259006366891","Type":"ContainerDied","Data":"635821dab0ae02edc0aa2f3a853371c56ec1a7c02751358d5fb555c27729f8a4"} Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.494542 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-795cw" event={"ID":"4bbe7e5f-5081-4b98-abbd-259006366891","Type":"ContainerDied","Data":"a9d371ab232869ed9de7a87a950cf4901a3327ad1c46e48ef62d8d3e394aefbf"} Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.494539 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-795cw" Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.494560 4997 scope.go:117] "RemoveContainer" containerID="635821dab0ae02edc0aa2f3a853371c56ec1a7c02751358d5fb555c27729f8a4" Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.519276 4997 scope.go:117] "RemoveContainer" containerID="52b6182c29f20be7ec2d7f9741aa2b054e3d0bc5417d1b0cbb96d35467578d5b" Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.539907 4997 scope.go:117] "RemoveContainer" containerID="0901005ff3f8c8db13fa413adb3fc92ec08b2b930ab83761d3b234343a257572" Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.585469 4997 scope.go:117] "RemoveContainer" containerID="635821dab0ae02edc0aa2f3a853371c56ec1a7c02751358d5fb555c27729f8a4" Dec 05 10:20:46 crc kubenswrapper[4997]: E1205 10:20:46.586041 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"635821dab0ae02edc0aa2f3a853371c56ec1a7c02751358d5fb555c27729f8a4\": container with ID starting with 635821dab0ae02edc0aa2f3a853371c56ec1a7c02751358d5fb555c27729f8a4 not found: ID does not exist" containerID="635821dab0ae02edc0aa2f3a853371c56ec1a7c02751358d5fb555c27729f8a4" Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.586080 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"635821dab0ae02edc0aa2f3a853371c56ec1a7c02751358d5fb555c27729f8a4"} err="failed to get container status \"635821dab0ae02edc0aa2f3a853371c56ec1a7c02751358d5fb555c27729f8a4\": rpc error: code = NotFound desc = could not find container \"635821dab0ae02edc0aa2f3a853371c56ec1a7c02751358d5fb555c27729f8a4\": container with ID starting with 635821dab0ae02edc0aa2f3a853371c56ec1a7c02751358d5fb555c27729f8a4 not found: ID does not exist" Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.586099 4997 scope.go:117] "RemoveContainer" containerID="52b6182c29f20be7ec2d7f9741aa2b054e3d0bc5417d1b0cbb96d35467578d5b" Dec 05 10:20:46 crc kubenswrapper[4997]: E1205 10:20:46.586452 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52b6182c29f20be7ec2d7f9741aa2b054e3d0bc5417d1b0cbb96d35467578d5b\": container with ID starting with 52b6182c29f20be7ec2d7f9741aa2b054e3d0bc5417d1b0cbb96d35467578d5b not found: ID does not exist" containerID="52b6182c29f20be7ec2d7f9741aa2b054e3d0bc5417d1b0cbb96d35467578d5b" Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.586478 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52b6182c29f20be7ec2d7f9741aa2b054e3d0bc5417d1b0cbb96d35467578d5b"} err="failed to get container status \"52b6182c29f20be7ec2d7f9741aa2b054e3d0bc5417d1b0cbb96d35467578d5b\": rpc error: code = NotFound desc = could not find container \"52b6182c29f20be7ec2d7f9741aa2b054e3d0bc5417d1b0cbb96d35467578d5b\": container with ID starting with 52b6182c29f20be7ec2d7f9741aa2b054e3d0bc5417d1b0cbb96d35467578d5b not found: ID does not exist" Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.586498 4997 scope.go:117] "RemoveContainer" containerID="0901005ff3f8c8db13fa413adb3fc92ec08b2b930ab83761d3b234343a257572" Dec 05 10:20:46 crc kubenswrapper[4997]: E1205 10:20:46.586897 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0901005ff3f8c8db13fa413adb3fc92ec08b2b930ab83761d3b234343a257572\": container with ID starting with 0901005ff3f8c8db13fa413adb3fc92ec08b2b930ab83761d3b234343a257572 not found: ID does not exist" containerID="0901005ff3f8c8db13fa413adb3fc92ec08b2b930ab83761d3b234343a257572" Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.586955 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0901005ff3f8c8db13fa413adb3fc92ec08b2b930ab83761d3b234343a257572"} err="failed to get container status \"0901005ff3f8c8db13fa413adb3fc92ec08b2b930ab83761d3b234343a257572\": rpc error: code = NotFound desc = could not find container \"0901005ff3f8c8db13fa413adb3fc92ec08b2b930ab83761d3b234343a257572\": container with ID starting with 0901005ff3f8c8db13fa413adb3fc92ec08b2b930ab83761d3b234343a257572 not found: ID does not exist" Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.816011 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bbe7e5f-5081-4b98-abbd-259006366891-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4bbe7e5f-5081-4b98-abbd-259006366891" (UID: "4bbe7e5f-5081-4b98-abbd-259006366891"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:20:46 crc kubenswrapper[4997]: I1205 10:20:46.901813 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bbe7e5f-5081-4b98-abbd-259006366891-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:20:47 crc kubenswrapper[4997]: I1205 10:20:47.135237 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-795cw"] Dec 05 10:20:47 crc kubenswrapper[4997]: I1205 10:20:47.148025 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-795cw"] Dec 05 10:20:47 crc kubenswrapper[4997]: I1205 10:20:47.761126 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bbe7e5f-5081-4b98-abbd-259006366891" path="/var/lib/kubelet/pods/4bbe7e5f-5081-4b98-abbd-259006366891/volumes" Dec 05 10:21:49 crc kubenswrapper[4997]: I1205 10:21:49.770132 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:21:49 crc kubenswrapper[4997]: I1205 10:21:49.771928 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:22:19 crc kubenswrapper[4997]: I1205 10:22:19.770098 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:22:19 crc kubenswrapper[4997]: I1205 10:22:19.770779 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:22:49 crc kubenswrapper[4997]: I1205 10:22:49.771008 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:22:49 crc kubenswrapper[4997]: I1205 10:22:49.771827 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:22:49 crc kubenswrapper[4997]: I1205 10:22:49.780755 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 10:22:49 crc kubenswrapper[4997]: I1205 10:22:49.784061 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 10:22:49 crc kubenswrapper[4997]: I1205 10:22:49.784242 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" gracePeriod=600 Dec 05 10:22:49 crc kubenswrapper[4997]: E1205 10:22:49.919755 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:22:50 crc kubenswrapper[4997]: I1205 10:22:50.862545 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" exitCode=0 Dec 05 10:22:50 crc kubenswrapper[4997]: I1205 10:22:50.862634 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc"} Dec 05 10:22:50 crc kubenswrapper[4997]: I1205 10:22:50.862755 4997 scope.go:117] "RemoveContainer" containerID="8a987de1021e7c3842ed4e7f92ee76fff33696ade660bfaad3dd390f94c1a7e6" Dec 05 10:22:50 crc kubenswrapper[4997]: I1205 10:22:50.865378 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:22:50 crc kubenswrapper[4997]: E1205 10:22:50.866248 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:23:04 crc kubenswrapper[4997]: I1205 10:23:04.750027 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:23:04 crc kubenswrapper[4997]: E1205 10:23:04.750741 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.065924 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lmlqr"] Dec 05 10:23:15 crc kubenswrapper[4997]: E1205 10:23:15.067148 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bbe7e5f-5081-4b98-abbd-259006366891" containerName="registry-server" Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.067167 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bbe7e5f-5081-4b98-abbd-259006366891" containerName="registry-server" Dec 05 10:23:15 crc kubenswrapper[4997]: E1205 10:23:15.067190 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bbe7e5f-5081-4b98-abbd-259006366891" containerName="extract-content" Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.067197 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bbe7e5f-5081-4b98-abbd-259006366891" containerName="extract-content" Dec 05 10:23:15 crc kubenswrapper[4997]: E1205 10:23:15.067217 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bbe7e5f-5081-4b98-abbd-259006366891" containerName="extract-utilities" Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.067225 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bbe7e5f-5081-4b98-abbd-259006366891" containerName="extract-utilities" Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.067527 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bbe7e5f-5081-4b98-abbd-259006366891" containerName="registry-server" Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.069692 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lmlqr" Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.104237 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lmlqr"] Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.199299 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f01dfaa-7e04-40ae-957d-6db6849f4890-utilities\") pod \"redhat-marketplace-lmlqr\" (UID: \"8f01dfaa-7e04-40ae-957d-6db6849f4890\") " pod="openshift-marketplace/redhat-marketplace-lmlqr" Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.199409 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f01dfaa-7e04-40ae-957d-6db6849f4890-catalog-content\") pod \"redhat-marketplace-lmlqr\" (UID: \"8f01dfaa-7e04-40ae-957d-6db6849f4890\") " pod="openshift-marketplace/redhat-marketplace-lmlqr" Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.200493 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qw5r\" (UniqueName: \"kubernetes.io/projected/8f01dfaa-7e04-40ae-957d-6db6849f4890-kube-api-access-5qw5r\") pod \"redhat-marketplace-lmlqr\" (UID: \"8f01dfaa-7e04-40ae-957d-6db6849f4890\") " pod="openshift-marketplace/redhat-marketplace-lmlqr" Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.302518 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qw5r\" (UniqueName: \"kubernetes.io/projected/8f01dfaa-7e04-40ae-957d-6db6849f4890-kube-api-access-5qw5r\") pod \"redhat-marketplace-lmlqr\" (UID: \"8f01dfaa-7e04-40ae-957d-6db6849f4890\") " pod="openshift-marketplace/redhat-marketplace-lmlqr" Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.302635 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f01dfaa-7e04-40ae-957d-6db6849f4890-utilities\") pod \"redhat-marketplace-lmlqr\" (UID: \"8f01dfaa-7e04-40ae-957d-6db6849f4890\") " pod="openshift-marketplace/redhat-marketplace-lmlqr" Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.302699 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f01dfaa-7e04-40ae-957d-6db6849f4890-catalog-content\") pod \"redhat-marketplace-lmlqr\" (UID: \"8f01dfaa-7e04-40ae-957d-6db6849f4890\") " pod="openshift-marketplace/redhat-marketplace-lmlqr" Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.303157 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f01dfaa-7e04-40ae-957d-6db6849f4890-catalog-content\") pod \"redhat-marketplace-lmlqr\" (UID: \"8f01dfaa-7e04-40ae-957d-6db6849f4890\") " pod="openshift-marketplace/redhat-marketplace-lmlqr" Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.303678 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f01dfaa-7e04-40ae-957d-6db6849f4890-utilities\") pod \"redhat-marketplace-lmlqr\" (UID: \"8f01dfaa-7e04-40ae-957d-6db6849f4890\") " pod="openshift-marketplace/redhat-marketplace-lmlqr" Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.328479 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qw5r\" (UniqueName: \"kubernetes.io/projected/8f01dfaa-7e04-40ae-957d-6db6849f4890-kube-api-access-5qw5r\") pod \"redhat-marketplace-lmlqr\" (UID: \"8f01dfaa-7e04-40ae-957d-6db6849f4890\") " pod="openshift-marketplace/redhat-marketplace-lmlqr" Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.400332 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lmlqr" Dec 05 10:23:15 crc kubenswrapper[4997]: I1205 10:23:15.935361 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lmlqr"] Dec 05 10:23:16 crc kubenswrapper[4997]: I1205 10:23:16.192104 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lmlqr" event={"ID":"8f01dfaa-7e04-40ae-957d-6db6849f4890","Type":"ContainerStarted","Data":"2483f328d77975175fc60e5e3164de8610fa0e9fb4a8da075b5874fa92946064"} Dec 05 10:23:17 crc kubenswrapper[4997]: I1205 10:23:17.751345 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:23:17 crc kubenswrapper[4997]: E1205 10:23:17.752143 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:23:18 crc kubenswrapper[4997]: I1205 10:23:18.223063 4997 generic.go:334] "Generic (PLEG): container finished" podID="8f01dfaa-7e04-40ae-957d-6db6849f4890" containerID="6d6f292b5a42722375a8b95d69a4a9a90fa1e744190649eecdf546f23eecb224" exitCode=0 Dec 05 10:23:18 crc kubenswrapper[4997]: I1205 10:23:18.223131 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lmlqr" event={"ID":"8f01dfaa-7e04-40ae-957d-6db6849f4890","Type":"ContainerDied","Data":"6d6f292b5a42722375a8b95d69a4a9a90fa1e744190649eecdf546f23eecb224"} Dec 05 10:23:20 crc kubenswrapper[4997]: I1205 10:23:20.245790 4997 generic.go:334] "Generic (PLEG): container finished" podID="8f01dfaa-7e04-40ae-957d-6db6849f4890" containerID="cab4d21814a980c00f12e29b1f9ad9964d147d68f330e467b9f196d29b094e43" exitCode=0 Dec 05 10:23:20 crc kubenswrapper[4997]: I1205 10:23:20.245911 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lmlqr" event={"ID":"8f01dfaa-7e04-40ae-957d-6db6849f4890","Type":"ContainerDied","Data":"cab4d21814a980c00f12e29b1f9ad9964d147d68f330e467b9f196d29b094e43"} Dec 05 10:23:22 crc kubenswrapper[4997]: I1205 10:23:22.274736 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lmlqr" event={"ID":"8f01dfaa-7e04-40ae-957d-6db6849f4890","Type":"ContainerStarted","Data":"97108f6770a4abe53cea9b7bfaf632b5e2dba7f09dcf797ac83430dce44753cf"} Dec 05 10:23:22 crc kubenswrapper[4997]: I1205 10:23:22.300933 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lmlqr" podStartSLOduration=4.275487904 podStartE2EDuration="7.30090811s" podCreationTimestamp="2025-12-05 10:23:15 +0000 UTC" firstStartedPulling="2025-12-05 10:23:18.225291892 +0000 UTC m=+12498.754199153" lastFinishedPulling="2025-12-05 10:23:21.250712098 +0000 UTC m=+12501.779619359" observedRunningTime="2025-12-05 10:23:22.297703654 +0000 UTC m=+12502.826610925" watchObservedRunningTime="2025-12-05 10:23:22.30090811 +0000 UTC m=+12502.829815371" Dec 05 10:23:25 crc kubenswrapper[4997]: I1205 10:23:25.400444 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lmlqr" Dec 05 10:23:25 crc kubenswrapper[4997]: I1205 10:23:25.401746 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lmlqr" Dec 05 10:23:25 crc kubenswrapper[4997]: I1205 10:23:25.447832 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lmlqr" Dec 05 10:23:26 crc kubenswrapper[4997]: I1205 10:23:26.373047 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lmlqr" Dec 05 10:23:26 crc kubenswrapper[4997]: I1205 10:23:26.426181 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lmlqr"] Dec 05 10:23:28 crc kubenswrapper[4997]: I1205 10:23:28.334495 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lmlqr" podUID="8f01dfaa-7e04-40ae-957d-6db6849f4890" containerName="registry-server" containerID="cri-o://97108f6770a4abe53cea9b7bfaf632b5e2dba7f09dcf797ac83430dce44753cf" gracePeriod=2 Dec 05 10:23:28 crc kubenswrapper[4997]: I1205 10:23:28.864590 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lmlqr" Dec 05 10:23:28 crc kubenswrapper[4997]: I1205 10:23:28.917833 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f01dfaa-7e04-40ae-957d-6db6849f4890-utilities\") pod \"8f01dfaa-7e04-40ae-957d-6db6849f4890\" (UID: \"8f01dfaa-7e04-40ae-957d-6db6849f4890\") " Dec 05 10:23:28 crc kubenswrapper[4997]: I1205 10:23:28.917986 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qw5r\" (UniqueName: \"kubernetes.io/projected/8f01dfaa-7e04-40ae-957d-6db6849f4890-kube-api-access-5qw5r\") pod \"8f01dfaa-7e04-40ae-957d-6db6849f4890\" (UID: \"8f01dfaa-7e04-40ae-957d-6db6849f4890\") " Dec 05 10:23:28 crc kubenswrapper[4997]: I1205 10:23:28.918128 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f01dfaa-7e04-40ae-957d-6db6849f4890-catalog-content\") pod \"8f01dfaa-7e04-40ae-957d-6db6849f4890\" (UID: \"8f01dfaa-7e04-40ae-957d-6db6849f4890\") " Dec 05 10:23:28 crc kubenswrapper[4997]: I1205 10:23:28.918954 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f01dfaa-7e04-40ae-957d-6db6849f4890-utilities" (OuterVolumeSpecName: "utilities") pod "8f01dfaa-7e04-40ae-957d-6db6849f4890" (UID: "8f01dfaa-7e04-40ae-957d-6db6849f4890"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:23:28 crc kubenswrapper[4997]: I1205 10:23:28.923578 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f01dfaa-7e04-40ae-957d-6db6849f4890-kube-api-access-5qw5r" (OuterVolumeSpecName: "kube-api-access-5qw5r") pod "8f01dfaa-7e04-40ae-957d-6db6849f4890" (UID: "8f01dfaa-7e04-40ae-957d-6db6849f4890"). InnerVolumeSpecName "kube-api-access-5qw5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:23:28 crc kubenswrapper[4997]: I1205 10:23:28.946648 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f01dfaa-7e04-40ae-957d-6db6849f4890-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f01dfaa-7e04-40ae-957d-6db6849f4890" (UID: "8f01dfaa-7e04-40ae-957d-6db6849f4890"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.021314 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qw5r\" (UniqueName: \"kubernetes.io/projected/8f01dfaa-7e04-40ae-957d-6db6849f4890-kube-api-access-5qw5r\") on node \"crc\" DevicePath \"\"" Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.021452 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f01dfaa-7e04-40ae-957d-6db6849f4890-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.021470 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f01dfaa-7e04-40ae-957d-6db6849f4890-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.363862 4997 generic.go:334] "Generic (PLEG): container finished" podID="8f01dfaa-7e04-40ae-957d-6db6849f4890" containerID="97108f6770a4abe53cea9b7bfaf632b5e2dba7f09dcf797ac83430dce44753cf" exitCode=0 Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.363905 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lmlqr" event={"ID":"8f01dfaa-7e04-40ae-957d-6db6849f4890","Type":"ContainerDied","Data":"97108f6770a4abe53cea9b7bfaf632b5e2dba7f09dcf797ac83430dce44753cf"} Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.363932 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lmlqr" event={"ID":"8f01dfaa-7e04-40ae-957d-6db6849f4890","Type":"ContainerDied","Data":"2483f328d77975175fc60e5e3164de8610fa0e9fb4a8da075b5874fa92946064"} Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.363948 4997 scope.go:117] "RemoveContainer" containerID="97108f6770a4abe53cea9b7bfaf632b5e2dba7f09dcf797ac83430dce44753cf" Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.364288 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lmlqr" Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.394601 4997 scope.go:117] "RemoveContainer" containerID="cab4d21814a980c00f12e29b1f9ad9964d147d68f330e467b9f196d29b094e43" Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.399513 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lmlqr"] Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.412724 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lmlqr"] Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.434994 4997 scope.go:117] "RemoveContainer" containerID="6d6f292b5a42722375a8b95d69a4a9a90fa1e744190649eecdf546f23eecb224" Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.473086 4997 scope.go:117] "RemoveContainer" containerID="97108f6770a4abe53cea9b7bfaf632b5e2dba7f09dcf797ac83430dce44753cf" Dec 05 10:23:29 crc kubenswrapper[4997]: E1205 10:23:29.479172 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97108f6770a4abe53cea9b7bfaf632b5e2dba7f09dcf797ac83430dce44753cf\": container with ID starting with 97108f6770a4abe53cea9b7bfaf632b5e2dba7f09dcf797ac83430dce44753cf not found: ID does not exist" containerID="97108f6770a4abe53cea9b7bfaf632b5e2dba7f09dcf797ac83430dce44753cf" Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.479232 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97108f6770a4abe53cea9b7bfaf632b5e2dba7f09dcf797ac83430dce44753cf"} err="failed to get container status \"97108f6770a4abe53cea9b7bfaf632b5e2dba7f09dcf797ac83430dce44753cf\": rpc error: code = NotFound desc = could not find container \"97108f6770a4abe53cea9b7bfaf632b5e2dba7f09dcf797ac83430dce44753cf\": container with ID starting with 97108f6770a4abe53cea9b7bfaf632b5e2dba7f09dcf797ac83430dce44753cf not found: ID does not exist" Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.479259 4997 scope.go:117] "RemoveContainer" containerID="cab4d21814a980c00f12e29b1f9ad9964d147d68f330e467b9f196d29b094e43" Dec 05 10:23:29 crc kubenswrapper[4997]: E1205 10:23:29.479645 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cab4d21814a980c00f12e29b1f9ad9964d147d68f330e467b9f196d29b094e43\": container with ID starting with cab4d21814a980c00f12e29b1f9ad9964d147d68f330e467b9f196d29b094e43 not found: ID does not exist" containerID="cab4d21814a980c00f12e29b1f9ad9964d147d68f330e467b9f196d29b094e43" Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.479691 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cab4d21814a980c00f12e29b1f9ad9964d147d68f330e467b9f196d29b094e43"} err="failed to get container status \"cab4d21814a980c00f12e29b1f9ad9964d147d68f330e467b9f196d29b094e43\": rpc error: code = NotFound desc = could not find container \"cab4d21814a980c00f12e29b1f9ad9964d147d68f330e467b9f196d29b094e43\": container with ID starting with cab4d21814a980c00f12e29b1f9ad9964d147d68f330e467b9f196d29b094e43 not found: ID does not exist" Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.479722 4997 scope.go:117] "RemoveContainer" containerID="6d6f292b5a42722375a8b95d69a4a9a90fa1e744190649eecdf546f23eecb224" Dec 05 10:23:29 crc kubenswrapper[4997]: E1205 10:23:29.480084 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d6f292b5a42722375a8b95d69a4a9a90fa1e744190649eecdf546f23eecb224\": container with ID starting with 6d6f292b5a42722375a8b95d69a4a9a90fa1e744190649eecdf546f23eecb224 not found: ID does not exist" containerID="6d6f292b5a42722375a8b95d69a4a9a90fa1e744190649eecdf546f23eecb224" Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.480126 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d6f292b5a42722375a8b95d69a4a9a90fa1e744190649eecdf546f23eecb224"} err="failed to get container status \"6d6f292b5a42722375a8b95d69a4a9a90fa1e744190649eecdf546f23eecb224\": rpc error: code = NotFound desc = could not find container \"6d6f292b5a42722375a8b95d69a4a9a90fa1e744190649eecdf546f23eecb224\": container with ID starting with 6d6f292b5a42722375a8b95d69a4a9a90fa1e744190649eecdf546f23eecb224 not found: ID does not exist" Dec 05 10:23:29 crc kubenswrapper[4997]: I1205 10:23:29.760991 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f01dfaa-7e04-40ae-957d-6db6849f4890" path="/var/lib/kubelet/pods/8f01dfaa-7e04-40ae-957d-6db6849f4890/volumes" Dec 05 10:23:32 crc kubenswrapper[4997]: I1205 10:23:32.749322 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:23:32 crc kubenswrapper[4997]: E1205 10:23:32.750339 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:23:44 crc kubenswrapper[4997]: I1205 10:23:44.749986 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:23:44 crc kubenswrapper[4997]: E1205 10:23:44.750872 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:23:52 crc kubenswrapper[4997]: I1205 10:23:52.547993 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-csft2"] Dec 05 10:23:52 crc kubenswrapper[4997]: E1205 10:23:52.548999 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f01dfaa-7e04-40ae-957d-6db6849f4890" containerName="extract-utilities" Dec 05 10:23:52 crc kubenswrapper[4997]: I1205 10:23:52.549013 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f01dfaa-7e04-40ae-957d-6db6849f4890" containerName="extract-utilities" Dec 05 10:23:52 crc kubenswrapper[4997]: E1205 10:23:52.549038 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f01dfaa-7e04-40ae-957d-6db6849f4890" containerName="registry-server" Dec 05 10:23:52 crc kubenswrapper[4997]: I1205 10:23:52.549044 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f01dfaa-7e04-40ae-957d-6db6849f4890" containerName="registry-server" Dec 05 10:23:52 crc kubenswrapper[4997]: E1205 10:23:52.549072 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f01dfaa-7e04-40ae-957d-6db6849f4890" containerName="extract-content" Dec 05 10:23:52 crc kubenswrapper[4997]: I1205 10:23:52.549079 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f01dfaa-7e04-40ae-957d-6db6849f4890" containerName="extract-content" Dec 05 10:23:52 crc kubenswrapper[4997]: I1205 10:23:52.549289 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f01dfaa-7e04-40ae-957d-6db6849f4890" containerName="registry-server" Dec 05 10:23:52 crc kubenswrapper[4997]: I1205 10:23:52.550776 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-csft2" Dec 05 10:23:52 crc kubenswrapper[4997]: I1205 10:23:52.564382 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-csft2"] Dec 05 10:23:52 crc kubenswrapper[4997]: I1205 10:23:52.666956 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554dc3b0-52c2-4a18-af3c-63d371ab90a4-utilities\") pod \"community-operators-csft2\" (UID: \"554dc3b0-52c2-4a18-af3c-63d371ab90a4\") " pod="openshift-marketplace/community-operators-csft2" Dec 05 10:23:52 crc kubenswrapper[4997]: I1205 10:23:52.667164 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29djw\" (UniqueName: \"kubernetes.io/projected/554dc3b0-52c2-4a18-af3c-63d371ab90a4-kube-api-access-29djw\") pod \"community-operators-csft2\" (UID: \"554dc3b0-52c2-4a18-af3c-63d371ab90a4\") " pod="openshift-marketplace/community-operators-csft2" Dec 05 10:23:52 crc kubenswrapper[4997]: I1205 10:23:52.667323 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554dc3b0-52c2-4a18-af3c-63d371ab90a4-catalog-content\") pod \"community-operators-csft2\" (UID: \"554dc3b0-52c2-4a18-af3c-63d371ab90a4\") " pod="openshift-marketplace/community-operators-csft2" Dec 05 10:23:52 crc kubenswrapper[4997]: I1205 10:23:52.769955 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29djw\" (UniqueName: \"kubernetes.io/projected/554dc3b0-52c2-4a18-af3c-63d371ab90a4-kube-api-access-29djw\") pod \"community-operators-csft2\" (UID: \"554dc3b0-52c2-4a18-af3c-63d371ab90a4\") " pod="openshift-marketplace/community-operators-csft2" Dec 05 10:23:52 crc kubenswrapper[4997]: I1205 10:23:52.770182 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554dc3b0-52c2-4a18-af3c-63d371ab90a4-catalog-content\") pod \"community-operators-csft2\" (UID: \"554dc3b0-52c2-4a18-af3c-63d371ab90a4\") " pod="openshift-marketplace/community-operators-csft2" Dec 05 10:23:52 crc kubenswrapper[4997]: I1205 10:23:52.770230 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554dc3b0-52c2-4a18-af3c-63d371ab90a4-utilities\") pod \"community-operators-csft2\" (UID: \"554dc3b0-52c2-4a18-af3c-63d371ab90a4\") " pod="openshift-marketplace/community-operators-csft2" Dec 05 10:23:52 crc kubenswrapper[4997]: I1205 10:23:52.770707 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554dc3b0-52c2-4a18-af3c-63d371ab90a4-catalog-content\") pod \"community-operators-csft2\" (UID: \"554dc3b0-52c2-4a18-af3c-63d371ab90a4\") " pod="openshift-marketplace/community-operators-csft2" Dec 05 10:23:52 crc kubenswrapper[4997]: I1205 10:23:52.770767 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554dc3b0-52c2-4a18-af3c-63d371ab90a4-utilities\") pod \"community-operators-csft2\" (UID: \"554dc3b0-52c2-4a18-af3c-63d371ab90a4\") " pod="openshift-marketplace/community-operators-csft2" Dec 05 10:23:52 crc kubenswrapper[4997]: I1205 10:23:52.789039 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29djw\" (UniqueName: \"kubernetes.io/projected/554dc3b0-52c2-4a18-af3c-63d371ab90a4-kube-api-access-29djw\") pod \"community-operators-csft2\" (UID: \"554dc3b0-52c2-4a18-af3c-63d371ab90a4\") " pod="openshift-marketplace/community-operators-csft2" Dec 05 10:23:52 crc kubenswrapper[4997]: I1205 10:23:52.874600 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-csft2" Dec 05 10:23:53 crc kubenswrapper[4997]: I1205 10:23:53.434858 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-csft2"] Dec 05 10:23:53 crc kubenswrapper[4997]: I1205 10:23:53.611068 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-csft2" event={"ID":"554dc3b0-52c2-4a18-af3c-63d371ab90a4","Type":"ContainerStarted","Data":"80cb64094f200edb58ead8bcacca74ede190927d0c48050839d69ec9c88c6a05"} Dec 05 10:23:54 crc kubenswrapper[4997]: I1205 10:23:54.642851 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-csft2" event={"ID":"554dc3b0-52c2-4a18-af3c-63d371ab90a4","Type":"ContainerStarted","Data":"8fe1939cc6524db31ecbdf0f5b773f1462d749db4d040d57e8d8338044d64cf2"} Dec 05 10:23:55 crc kubenswrapper[4997]: I1205 10:23:55.653634 4997 generic.go:334] "Generic (PLEG): container finished" podID="554dc3b0-52c2-4a18-af3c-63d371ab90a4" containerID="8fe1939cc6524db31ecbdf0f5b773f1462d749db4d040d57e8d8338044d64cf2" exitCode=0 Dec 05 10:23:55 crc kubenswrapper[4997]: I1205 10:23:55.653682 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-csft2" event={"ID":"554dc3b0-52c2-4a18-af3c-63d371ab90a4","Type":"ContainerDied","Data":"8fe1939cc6524db31ecbdf0f5b773f1462d749db4d040d57e8d8338044d64cf2"} Dec 05 10:23:56 crc kubenswrapper[4997]: I1205 10:23:56.750022 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:23:56 crc kubenswrapper[4997]: E1205 10:23:56.750541 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:23:59 crc kubenswrapper[4997]: I1205 10:23:59.695315 4997 generic.go:334] "Generic (PLEG): container finished" podID="554dc3b0-52c2-4a18-af3c-63d371ab90a4" containerID="b18ec892113bdb44c08882928357c8b395f0bb1ada4f7c2dbe00f4642e34f311" exitCode=0 Dec 05 10:23:59 crc kubenswrapper[4997]: I1205 10:23:59.695644 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-csft2" event={"ID":"554dc3b0-52c2-4a18-af3c-63d371ab90a4","Type":"ContainerDied","Data":"b18ec892113bdb44c08882928357c8b395f0bb1ada4f7c2dbe00f4642e34f311"} Dec 05 10:24:01 crc kubenswrapper[4997]: I1205 10:24:01.717285 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-csft2" event={"ID":"554dc3b0-52c2-4a18-af3c-63d371ab90a4","Type":"ContainerStarted","Data":"622540daa01fed305a8fa0f628f8a2174f33e8f94d1356eb68a0e2dda88cdfd2"} Dec 05 10:24:01 crc kubenswrapper[4997]: I1205 10:24:01.741268 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-csft2" podStartSLOduration=4.511967565 podStartE2EDuration="9.741249802s" podCreationTimestamp="2025-12-05 10:23:52 +0000 UTC" firstStartedPulling="2025-12-05 10:23:55.655682341 +0000 UTC m=+12536.184589622" lastFinishedPulling="2025-12-05 10:24:00.884964598 +0000 UTC m=+12541.413871859" observedRunningTime="2025-12-05 10:24:01.732702851 +0000 UTC m=+12542.261610142" watchObservedRunningTime="2025-12-05 10:24:01.741249802 +0000 UTC m=+12542.270157053" Dec 05 10:24:02 crc kubenswrapper[4997]: I1205 10:24:02.875100 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-csft2" Dec 05 10:24:02 crc kubenswrapper[4997]: I1205 10:24:02.875389 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-csft2" Dec 05 10:24:02 crc kubenswrapper[4997]: I1205 10:24:02.924270 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-csft2" Dec 05 10:24:07 crc kubenswrapper[4997]: I1205 10:24:07.749788 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:24:07 crc kubenswrapper[4997]: E1205 10:24:07.750815 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:24:12 crc kubenswrapper[4997]: I1205 10:24:12.931712 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-csft2" Dec 05 10:24:13 crc kubenswrapper[4997]: I1205 10:24:13.012083 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-csft2"] Dec 05 10:24:13 crc kubenswrapper[4997]: I1205 10:24:13.841193 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-csft2" podUID="554dc3b0-52c2-4a18-af3c-63d371ab90a4" containerName="registry-server" containerID="cri-o://622540daa01fed305a8fa0f628f8a2174f33e8f94d1356eb68a0e2dda88cdfd2" gracePeriod=2 Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.346546 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-csft2" Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.430222 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554dc3b0-52c2-4a18-af3c-63d371ab90a4-catalog-content\") pod \"554dc3b0-52c2-4a18-af3c-63d371ab90a4\" (UID: \"554dc3b0-52c2-4a18-af3c-63d371ab90a4\") " Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.430398 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29djw\" (UniqueName: \"kubernetes.io/projected/554dc3b0-52c2-4a18-af3c-63d371ab90a4-kube-api-access-29djw\") pod \"554dc3b0-52c2-4a18-af3c-63d371ab90a4\" (UID: \"554dc3b0-52c2-4a18-af3c-63d371ab90a4\") " Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.430536 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554dc3b0-52c2-4a18-af3c-63d371ab90a4-utilities\") pod \"554dc3b0-52c2-4a18-af3c-63d371ab90a4\" (UID: \"554dc3b0-52c2-4a18-af3c-63d371ab90a4\") " Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.431876 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/554dc3b0-52c2-4a18-af3c-63d371ab90a4-utilities" (OuterVolumeSpecName: "utilities") pod "554dc3b0-52c2-4a18-af3c-63d371ab90a4" (UID: "554dc3b0-52c2-4a18-af3c-63d371ab90a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.436042 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/554dc3b0-52c2-4a18-af3c-63d371ab90a4-kube-api-access-29djw" (OuterVolumeSpecName: "kube-api-access-29djw") pod "554dc3b0-52c2-4a18-af3c-63d371ab90a4" (UID: "554dc3b0-52c2-4a18-af3c-63d371ab90a4"). InnerVolumeSpecName "kube-api-access-29djw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.486557 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/554dc3b0-52c2-4a18-af3c-63d371ab90a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "554dc3b0-52c2-4a18-af3c-63d371ab90a4" (UID: "554dc3b0-52c2-4a18-af3c-63d371ab90a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.533477 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554dc3b0-52c2-4a18-af3c-63d371ab90a4-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.533518 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554dc3b0-52c2-4a18-af3c-63d371ab90a4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.533538 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29djw\" (UniqueName: \"kubernetes.io/projected/554dc3b0-52c2-4a18-af3c-63d371ab90a4-kube-api-access-29djw\") on node \"crc\" DevicePath \"\"" Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.855005 4997 generic.go:334] "Generic (PLEG): container finished" podID="554dc3b0-52c2-4a18-af3c-63d371ab90a4" containerID="622540daa01fed305a8fa0f628f8a2174f33e8f94d1356eb68a0e2dda88cdfd2" exitCode=0 Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.855052 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-csft2" event={"ID":"554dc3b0-52c2-4a18-af3c-63d371ab90a4","Type":"ContainerDied","Data":"622540daa01fed305a8fa0f628f8a2174f33e8f94d1356eb68a0e2dda88cdfd2"} Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.855082 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-csft2" event={"ID":"554dc3b0-52c2-4a18-af3c-63d371ab90a4","Type":"ContainerDied","Data":"80cb64094f200edb58ead8bcacca74ede190927d0c48050839d69ec9c88c6a05"} Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.855104 4997 scope.go:117] "RemoveContainer" containerID="622540daa01fed305a8fa0f628f8a2174f33e8f94d1356eb68a0e2dda88cdfd2" Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.855263 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-csft2" Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.879508 4997 scope.go:117] "RemoveContainer" containerID="b18ec892113bdb44c08882928357c8b395f0bb1ada4f7c2dbe00f4642e34f311" Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.901116 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-csft2"] Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.910764 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-csft2"] Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.920702 4997 scope.go:117] "RemoveContainer" containerID="8fe1939cc6524db31ecbdf0f5b773f1462d749db4d040d57e8d8338044d64cf2" Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.964178 4997 scope.go:117] "RemoveContainer" containerID="622540daa01fed305a8fa0f628f8a2174f33e8f94d1356eb68a0e2dda88cdfd2" Dec 05 10:24:14 crc kubenswrapper[4997]: E1205 10:24:14.964729 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"622540daa01fed305a8fa0f628f8a2174f33e8f94d1356eb68a0e2dda88cdfd2\": container with ID starting with 622540daa01fed305a8fa0f628f8a2174f33e8f94d1356eb68a0e2dda88cdfd2 not found: ID does not exist" containerID="622540daa01fed305a8fa0f628f8a2174f33e8f94d1356eb68a0e2dda88cdfd2" Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.964772 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"622540daa01fed305a8fa0f628f8a2174f33e8f94d1356eb68a0e2dda88cdfd2"} err="failed to get container status \"622540daa01fed305a8fa0f628f8a2174f33e8f94d1356eb68a0e2dda88cdfd2\": rpc error: code = NotFound desc = could not find container \"622540daa01fed305a8fa0f628f8a2174f33e8f94d1356eb68a0e2dda88cdfd2\": container with ID starting with 622540daa01fed305a8fa0f628f8a2174f33e8f94d1356eb68a0e2dda88cdfd2 not found: ID does not exist" Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.964797 4997 scope.go:117] "RemoveContainer" containerID="b18ec892113bdb44c08882928357c8b395f0bb1ada4f7c2dbe00f4642e34f311" Dec 05 10:24:14 crc kubenswrapper[4997]: E1205 10:24:14.965227 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b18ec892113bdb44c08882928357c8b395f0bb1ada4f7c2dbe00f4642e34f311\": container with ID starting with b18ec892113bdb44c08882928357c8b395f0bb1ada4f7c2dbe00f4642e34f311 not found: ID does not exist" containerID="b18ec892113bdb44c08882928357c8b395f0bb1ada4f7c2dbe00f4642e34f311" Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.965290 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b18ec892113bdb44c08882928357c8b395f0bb1ada4f7c2dbe00f4642e34f311"} err="failed to get container status \"b18ec892113bdb44c08882928357c8b395f0bb1ada4f7c2dbe00f4642e34f311\": rpc error: code = NotFound desc = could not find container \"b18ec892113bdb44c08882928357c8b395f0bb1ada4f7c2dbe00f4642e34f311\": container with ID starting with b18ec892113bdb44c08882928357c8b395f0bb1ada4f7c2dbe00f4642e34f311 not found: ID does not exist" Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.965325 4997 scope.go:117] "RemoveContainer" containerID="8fe1939cc6524db31ecbdf0f5b773f1462d749db4d040d57e8d8338044d64cf2" Dec 05 10:24:14 crc kubenswrapper[4997]: E1205 10:24:14.965643 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fe1939cc6524db31ecbdf0f5b773f1462d749db4d040d57e8d8338044d64cf2\": container with ID starting with 8fe1939cc6524db31ecbdf0f5b773f1462d749db4d040d57e8d8338044d64cf2 not found: ID does not exist" containerID="8fe1939cc6524db31ecbdf0f5b773f1462d749db4d040d57e8d8338044d64cf2" Dec 05 10:24:14 crc kubenswrapper[4997]: I1205 10:24:14.965677 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fe1939cc6524db31ecbdf0f5b773f1462d749db4d040d57e8d8338044d64cf2"} err="failed to get container status \"8fe1939cc6524db31ecbdf0f5b773f1462d749db4d040d57e8d8338044d64cf2\": rpc error: code = NotFound desc = could not find container \"8fe1939cc6524db31ecbdf0f5b773f1462d749db4d040d57e8d8338044d64cf2\": container with ID starting with 8fe1939cc6524db31ecbdf0f5b773f1462d749db4d040d57e8d8338044d64cf2 not found: ID does not exist" Dec 05 10:24:15 crc kubenswrapper[4997]: I1205 10:24:15.761548 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="554dc3b0-52c2-4a18-af3c-63d371ab90a4" path="/var/lib/kubelet/pods/554dc3b0-52c2-4a18-af3c-63d371ab90a4/volumes" Dec 05 10:24:18 crc kubenswrapper[4997]: I1205 10:24:18.749992 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:24:18 crc kubenswrapper[4997]: E1205 10:24:18.750703 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:24:32 crc kubenswrapper[4997]: I1205 10:24:32.749080 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:24:32 crc kubenswrapper[4997]: E1205 10:24:32.749902 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:24:35 crc kubenswrapper[4997]: I1205 10:24:35.158794 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xt9w6"] Dec 05 10:24:35 crc kubenswrapper[4997]: E1205 10:24:35.159480 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="554dc3b0-52c2-4a18-af3c-63d371ab90a4" containerName="extract-utilities" Dec 05 10:24:35 crc kubenswrapper[4997]: I1205 10:24:35.159493 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="554dc3b0-52c2-4a18-af3c-63d371ab90a4" containerName="extract-utilities" Dec 05 10:24:35 crc kubenswrapper[4997]: E1205 10:24:35.159510 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="554dc3b0-52c2-4a18-af3c-63d371ab90a4" containerName="registry-server" Dec 05 10:24:35 crc kubenswrapper[4997]: I1205 10:24:35.159515 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="554dc3b0-52c2-4a18-af3c-63d371ab90a4" containerName="registry-server" Dec 05 10:24:35 crc kubenswrapper[4997]: E1205 10:24:35.159531 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="554dc3b0-52c2-4a18-af3c-63d371ab90a4" containerName="extract-content" Dec 05 10:24:35 crc kubenswrapper[4997]: I1205 10:24:35.159537 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="554dc3b0-52c2-4a18-af3c-63d371ab90a4" containerName="extract-content" Dec 05 10:24:35 crc kubenswrapper[4997]: I1205 10:24:35.159774 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="554dc3b0-52c2-4a18-af3c-63d371ab90a4" containerName="registry-server" Dec 05 10:24:35 crc kubenswrapper[4997]: I1205 10:24:35.161239 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xt9w6" Dec 05 10:24:35 crc kubenswrapper[4997]: I1205 10:24:35.309738 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80acf65b-25bd-461b-9169-37a8af196319-utilities\") pod \"redhat-operators-xt9w6\" (UID: \"80acf65b-25bd-461b-9169-37a8af196319\") " pod="openshift-marketplace/redhat-operators-xt9w6" Dec 05 10:24:35 crc kubenswrapper[4997]: I1205 10:24:35.309917 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfflv\" (UniqueName: \"kubernetes.io/projected/80acf65b-25bd-461b-9169-37a8af196319-kube-api-access-jfflv\") pod \"redhat-operators-xt9w6\" (UID: \"80acf65b-25bd-461b-9169-37a8af196319\") " pod="openshift-marketplace/redhat-operators-xt9w6" Dec 05 10:24:35 crc kubenswrapper[4997]: I1205 10:24:35.309969 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80acf65b-25bd-461b-9169-37a8af196319-catalog-content\") pod \"redhat-operators-xt9w6\" (UID: \"80acf65b-25bd-461b-9169-37a8af196319\") " pod="openshift-marketplace/redhat-operators-xt9w6" Dec 05 10:24:35 crc kubenswrapper[4997]: I1205 10:24:35.413488 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfflv\" (UniqueName: \"kubernetes.io/projected/80acf65b-25bd-461b-9169-37a8af196319-kube-api-access-jfflv\") pod \"redhat-operators-xt9w6\" (UID: \"80acf65b-25bd-461b-9169-37a8af196319\") " pod="openshift-marketplace/redhat-operators-xt9w6" Dec 05 10:24:35 crc kubenswrapper[4997]: I1205 10:24:35.413605 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80acf65b-25bd-461b-9169-37a8af196319-catalog-content\") pod \"redhat-operators-xt9w6\" (UID: \"80acf65b-25bd-461b-9169-37a8af196319\") " pod="openshift-marketplace/redhat-operators-xt9w6" Dec 05 10:24:35 crc kubenswrapper[4997]: I1205 10:24:35.413758 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80acf65b-25bd-461b-9169-37a8af196319-utilities\") pod \"redhat-operators-xt9w6\" (UID: \"80acf65b-25bd-461b-9169-37a8af196319\") " pod="openshift-marketplace/redhat-operators-xt9w6" Dec 05 10:24:35 crc kubenswrapper[4997]: I1205 10:24:35.414100 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80acf65b-25bd-461b-9169-37a8af196319-catalog-content\") pod \"redhat-operators-xt9w6\" (UID: \"80acf65b-25bd-461b-9169-37a8af196319\") " pod="openshift-marketplace/redhat-operators-xt9w6" Dec 05 10:24:35 crc kubenswrapper[4997]: I1205 10:24:35.414363 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80acf65b-25bd-461b-9169-37a8af196319-utilities\") pod \"redhat-operators-xt9w6\" (UID: \"80acf65b-25bd-461b-9169-37a8af196319\") " pod="openshift-marketplace/redhat-operators-xt9w6" Dec 05 10:24:35 crc kubenswrapper[4997]: I1205 10:24:35.439207 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfflv\" (UniqueName: \"kubernetes.io/projected/80acf65b-25bd-461b-9169-37a8af196319-kube-api-access-jfflv\") pod \"redhat-operators-xt9w6\" (UID: \"80acf65b-25bd-461b-9169-37a8af196319\") " pod="openshift-marketplace/redhat-operators-xt9w6" Dec 05 10:24:35 crc kubenswrapper[4997]: I1205 10:24:35.482912 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xt9w6" Dec 05 10:24:37 crc kubenswrapper[4997]: I1205 10:24:37.444951 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-2plrj" podUID="24697cfc-2c63-4b44-b4db-88a361476c5f" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 10:24:37 crc kubenswrapper[4997]: I1205 10:24:37.626599 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xt9w6"] Dec 05 10:24:38 crc kubenswrapper[4997]: I1205 10:24:38.178259 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xt9w6"] Dec 05 10:24:39 crc kubenswrapper[4997]: I1205 10:24:39.112131 4997 generic.go:334] "Generic (PLEG): container finished" podID="80acf65b-25bd-461b-9169-37a8af196319" containerID="5e4ef32fcf4f50529caf4dd95c736623bb4910634e5c3cb9269fea95801f3043" exitCode=0 Dec 05 10:24:39 crc kubenswrapper[4997]: I1205 10:24:39.112260 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xt9w6" event={"ID":"80acf65b-25bd-461b-9169-37a8af196319","Type":"ContainerDied","Data":"5e4ef32fcf4f50529caf4dd95c736623bb4910634e5c3cb9269fea95801f3043"} Dec 05 10:24:39 crc kubenswrapper[4997]: I1205 10:24:39.112830 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xt9w6" event={"ID":"80acf65b-25bd-461b-9169-37a8af196319","Type":"ContainerStarted","Data":"47e9d436fe5cf81930ae9621e2235ea440a991e73d0ebdb7fa4bec5ab5b1c95a"} Dec 05 10:24:41 crc kubenswrapper[4997]: I1205 10:24:41.132546 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xt9w6" event={"ID":"80acf65b-25bd-461b-9169-37a8af196319","Type":"ContainerStarted","Data":"5f2644215502e3a4305eabfbc9104027d23a6bd5b8ea1f146568e66d75dc6ef8"} Dec 05 10:24:43 crc kubenswrapper[4997]: I1205 10:24:43.153536 4997 generic.go:334] "Generic (PLEG): container finished" podID="80acf65b-25bd-461b-9169-37a8af196319" containerID="5f2644215502e3a4305eabfbc9104027d23a6bd5b8ea1f146568e66d75dc6ef8" exitCode=0 Dec 05 10:24:43 crc kubenswrapper[4997]: I1205 10:24:43.153633 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xt9w6" event={"ID":"80acf65b-25bd-461b-9169-37a8af196319","Type":"ContainerDied","Data":"5f2644215502e3a4305eabfbc9104027d23a6bd5b8ea1f146568e66d75dc6ef8"} Dec 05 10:24:44 crc kubenswrapper[4997]: I1205 10:24:44.166964 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xt9w6" event={"ID":"80acf65b-25bd-461b-9169-37a8af196319","Type":"ContainerStarted","Data":"35194d95ba630e2ae26c4a6c5ed261f09eb51fdaf64449156824ffd3926b0d3e"} Dec 05 10:24:44 crc kubenswrapper[4997]: I1205 10:24:44.186597 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xt9w6" podStartSLOduration=4.7454254129999995 podStartE2EDuration="9.186574515s" podCreationTimestamp="2025-12-05 10:24:35 +0000 UTC" firstStartedPulling="2025-12-05 10:24:39.114951329 +0000 UTC m=+12579.643858590" lastFinishedPulling="2025-12-05 10:24:43.556100431 +0000 UTC m=+12584.085007692" observedRunningTime="2025-12-05 10:24:44.183139312 +0000 UTC m=+12584.712046573" watchObservedRunningTime="2025-12-05 10:24:44.186574515 +0000 UTC m=+12584.715481776" Dec 05 10:24:45 crc kubenswrapper[4997]: I1205 10:24:45.483808 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xt9w6" Dec 05 10:24:45 crc kubenswrapper[4997]: I1205 10:24:45.485821 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xt9w6" Dec 05 10:24:46 crc kubenswrapper[4997]: I1205 10:24:46.533980 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xt9w6" podUID="80acf65b-25bd-461b-9169-37a8af196319" containerName="registry-server" probeResult="failure" output=< Dec 05 10:24:46 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 10:24:46 crc kubenswrapper[4997]: > Dec 05 10:24:47 crc kubenswrapper[4997]: I1205 10:24:47.749905 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:24:47 crc kubenswrapper[4997]: E1205 10:24:47.750451 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:24:55 crc kubenswrapper[4997]: I1205 10:24:55.529027 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xt9w6" Dec 05 10:24:55 crc kubenswrapper[4997]: I1205 10:24:55.583509 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xt9w6" Dec 05 10:24:55 crc kubenswrapper[4997]: I1205 10:24:55.767964 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xt9w6"] Dec 05 10:24:57 crc kubenswrapper[4997]: I1205 10:24:57.329561 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xt9w6" podUID="80acf65b-25bd-461b-9169-37a8af196319" containerName="registry-server" containerID="cri-o://35194d95ba630e2ae26c4a6c5ed261f09eb51fdaf64449156824ffd3926b0d3e" gracePeriod=2 Dec 05 10:24:57 crc kubenswrapper[4997]: I1205 10:24:57.989942 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xt9w6" Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.101128 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80acf65b-25bd-461b-9169-37a8af196319-utilities\") pod \"80acf65b-25bd-461b-9169-37a8af196319\" (UID: \"80acf65b-25bd-461b-9169-37a8af196319\") " Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.101322 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfflv\" (UniqueName: \"kubernetes.io/projected/80acf65b-25bd-461b-9169-37a8af196319-kube-api-access-jfflv\") pod \"80acf65b-25bd-461b-9169-37a8af196319\" (UID: \"80acf65b-25bd-461b-9169-37a8af196319\") " Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.101357 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80acf65b-25bd-461b-9169-37a8af196319-catalog-content\") pod \"80acf65b-25bd-461b-9169-37a8af196319\" (UID: \"80acf65b-25bd-461b-9169-37a8af196319\") " Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.107970 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80acf65b-25bd-461b-9169-37a8af196319-utilities" (OuterVolumeSpecName: "utilities") pod "80acf65b-25bd-461b-9169-37a8af196319" (UID: "80acf65b-25bd-461b-9169-37a8af196319"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.118982 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80acf65b-25bd-461b-9169-37a8af196319-kube-api-access-jfflv" (OuterVolumeSpecName: "kube-api-access-jfflv") pod "80acf65b-25bd-461b-9169-37a8af196319" (UID: "80acf65b-25bd-461b-9169-37a8af196319"). InnerVolumeSpecName "kube-api-access-jfflv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.205458 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfflv\" (UniqueName: \"kubernetes.io/projected/80acf65b-25bd-461b-9169-37a8af196319-kube-api-access-jfflv\") on node \"crc\" DevicePath \"\"" Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.205509 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80acf65b-25bd-461b-9169-37a8af196319-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.213332 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80acf65b-25bd-461b-9169-37a8af196319-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80acf65b-25bd-461b-9169-37a8af196319" (UID: "80acf65b-25bd-461b-9169-37a8af196319"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.307006 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80acf65b-25bd-461b-9169-37a8af196319-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.343540 4997 generic.go:334] "Generic (PLEG): container finished" podID="80acf65b-25bd-461b-9169-37a8af196319" containerID="35194d95ba630e2ae26c4a6c5ed261f09eb51fdaf64449156824ffd3926b0d3e" exitCode=0 Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.343599 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xt9w6" event={"ID":"80acf65b-25bd-461b-9169-37a8af196319","Type":"ContainerDied","Data":"35194d95ba630e2ae26c4a6c5ed261f09eb51fdaf64449156824ffd3926b0d3e"} Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.343652 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xt9w6" Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.343676 4997 scope.go:117] "RemoveContainer" containerID="35194d95ba630e2ae26c4a6c5ed261f09eb51fdaf64449156824ffd3926b0d3e" Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.343661 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xt9w6" event={"ID":"80acf65b-25bd-461b-9169-37a8af196319","Type":"ContainerDied","Data":"47e9d436fe5cf81930ae9621e2235ea440a991e73d0ebdb7fa4bec5ab5b1c95a"} Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.381415 4997 scope.go:117] "RemoveContainer" containerID="5f2644215502e3a4305eabfbc9104027d23a6bd5b8ea1f146568e66d75dc6ef8" Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.399214 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xt9w6"] Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.418884 4997 scope.go:117] "RemoveContainer" containerID="5e4ef32fcf4f50529caf4dd95c736623bb4910634e5c3cb9269fea95801f3043" Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.423707 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xt9w6"] Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.454976 4997 scope.go:117] "RemoveContainer" containerID="35194d95ba630e2ae26c4a6c5ed261f09eb51fdaf64449156824ffd3926b0d3e" Dec 05 10:24:58 crc kubenswrapper[4997]: E1205 10:24:58.461527 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35194d95ba630e2ae26c4a6c5ed261f09eb51fdaf64449156824ffd3926b0d3e\": container with ID starting with 35194d95ba630e2ae26c4a6c5ed261f09eb51fdaf64449156824ffd3926b0d3e not found: ID does not exist" containerID="35194d95ba630e2ae26c4a6c5ed261f09eb51fdaf64449156824ffd3926b0d3e" Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.461569 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35194d95ba630e2ae26c4a6c5ed261f09eb51fdaf64449156824ffd3926b0d3e"} err="failed to get container status \"35194d95ba630e2ae26c4a6c5ed261f09eb51fdaf64449156824ffd3926b0d3e\": rpc error: code = NotFound desc = could not find container \"35194d95ba630e2ae26c4a6c5ed261f09eb51fdaf64449156824ffd3926b0d3e\": container with ID starting with 35194d95ba630e2ae26c4a6c5ed261f09eb51fdaf64449156824ffd3926b0d3e not found: ID does not exist" Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.461598 4997 scope.go:117] "RemoveContainer" containerID="5f2644215502e3a4305eabfbc9104027d23a6bd5b8ea1f146568e66d75dc6ef8" Dec 05 10:24:58 crc kubenswrapper[4997]: E1205 10:24:58.461998 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f2644215502e3a4305eabfbc9104027d23a6bd5b8ea1f146568e66d75dc6ef8\": container with ID starting with 5f2644215502e3a4305eabfbc9104027d23a6bd5b8ea1f146568e66d75dc6ef8 not found: ID does not exist" containerID="5f2644215502e3a4305eabfbc9104027d23a6bd5b8ea1f146568e66d75dc6ef8" Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.462108 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f2644215502e3a4305eabfbc9104027d23a6bd5b8ea1f146568e66d75dc6ef8"} err="failed to get container status \"5f2644215502e3a4305eabfbc9104027d23a6bd5b8ea1f146568e66d75dc6ef8\": rpc error: code = NotFound desc = could not find container \"5f2644215502e3a4305eabfbc9104027d23a6bd5b8ea1f146568e66d75dc6ef8\": container with ID starting with 5f2644215502e3a4305eabfbc9104027d23a6bd5b8ea1f146568e66d75dc6ef8 not found: ID does not exist" Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.462198 4997 scope.go:117] "RemoveContainer" containerID="5e4ef32fcf4f50529caf4dd95c736623bb4910634e5c3cb9269fea95801f3043" Dec 05 10:24:58 crc kubenswrapper[4997]: E1205 10:24:58.463109 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e4ef32fcf4f50529caf4dd95c736623bb4910634e5c3cb9269fea95801f3043\": container with ID starting with 5e4ef32fcf4f50529caf4dd95c736623bb4910634e5c3cb9269fea95801f3043 not found: ID does not exist" containerID="5e4ef32fcf4f50529caf4dd95c736623bb4910634e5c3cb9269fea95801f3043" Dec 05 10:24:58 crc kubenswrapper[4997]: I1205 10:24:58.463142 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e4ef32fcf4f50529caf4dd95c736623bb4910634e5c3cb9269fea95801f3043"} err="failed to get container status \"5e4ef32fcf4f50529caf4dd95c736623bb4910634e5c3cb9269fea95801f3043\": rpc error: code = NotFound desc = could not find container \"5e4ef32fcf4f50529caf4dd95c736623bb4910634e5c3cb9269fea95801f3043\": container with ID starting with 5e4ef32fcf4f50529caf4dd95c736623bb4910634e5c3cb9269fea95801f3043 not found: ID does not exist" Dec 05 10:24:59 crc kubenswrapper[4997]: I1205 10:24:59.771757 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80acf65b-25bd-461b-9169-37a8af196319" path="/var/lib/kubelet/pods/80acf65b-25bd-461b-9169-37a8af196319/volumes" Dec 05 10:25:02 crc kubenswrapper[4997]: I1205 10:25:02.749563 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:25:02 crc kubenswrapper[4997]: E1205 10:25:02.750364 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:25:15 crc kubenswrapper[4997]: I1205 10:25:15.749353 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:25:15 crc kubenswrapper[4997]: E1205 10:25:15.750267 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:25:26 crc kubenswrapper[4997]: I1205 10:25:26.749718 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:25:26 crc kubenswrapper[4997]: E1205 10:25:26.750641 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:25:41 crc kubenswrapper[4997]: I1205 10:25:41.748596 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:25:41 crc kubenswrapper[4997]: E1205 10:25:41.749329 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:25:54 crc kubenswrapper[4997]: I1205 10:25:54.749877 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:25:54 crc kubenswrapper[4997]: E1205 10:25:54.750826 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:26:06 crc kubenswrapper[4997]: I1205 10:26:06.749683 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:26:06 crc kubenswrapper[4997]: E1205 10:26:06.750858 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:26:18 crc kubenswrapper[4997]: I1205 10:26:18.749701 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:26:18 crc kubenswrapper[4997]: E1205 10:26:18.750433 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:26:30 crc kubenswrapper[4997]: I1205 10:26:30.749414 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:26:30 crc kubenswrapper[4997]: E1205 10:26:30.750279 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:26:45 crc kubenswrapper[4997]: I1205 10:26:45.749177 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:26:45 crc kubenswrapper[4997]: E1205 10:26:45.750066 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:26:56 crc kubenswrapper[4997]: I1205 10:26:56.749847 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:26:56 crc kubenswrapper[4997]: E1205 10:26:56.750629 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:27:08 crc kubenswrapper[4997]: I1205 10:27:08.749513 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:27:08 crc kubenswrapper[4997]: E1205 10:27:08.750327 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:27:23 crc kubenswrapper[4997]: I1205 10:27:23.748976 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:27:23 crc kubenswrapper[4997]: E1205 10:27:23.749675 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:27:25 crc kubenswrapper[4997]: I1205 10:27:25.229786 4997 generic.go:334] "Generic (PLEG): container finished" podID="8688d151-9d21-4107-86cd-f9cee40d08b8" containerID="936834a8577f4301dd95bb05bf37471800043ac7dedb18f11e7f25a128566bad" exitCode=0 Dec 05 10:27:25 crc kubenswrapper[4997]: I1205 10:27:25.229906 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8688d151-9d21-4107-86cd-f9cee40d08b8","Type":"ContainerDied","Data":"936834a8577f4301dd95bb05bf37471800043ac7dedb18f11e7f25a128566bad"} Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.671082 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.755256 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn9zx\" (UniqueName: \"kubernetes.io/projected/8688d151-9d21-4107-86cd-f9cee40d08b8-kube-api-access-zn9zx\") pod \"8688d151-9d21-4107-86cd-f9cee40d08b8\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.755322 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8688d151-9d21-4107-86cd-f9cee40d08b8-openstack-config\") pod \"8688d151-9d21-4107-86cd-f9cee40d08b8\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.755451 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-ssh-key\") pod \"8688d151-9d21-4107-86cd-f9cee40d08b8\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.755507 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8688d151-9d21-4107-86cd-f9cee40d08b8-test-operator-ephemeral-temporary\") pod \"8688d151-9d21-4107-86cd-f9cee40d08b8\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.755544 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8688d151-9d21-4107-86cd-f9cee40d08b8-config-data\") pod \"8688d151-9d21-4107-86cd-f9cee40d08b8\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.755580 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-ca-certs\") pod \"8688d151-9d21-4107-86cd-f9cee40d08b8\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.755641 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"8688d151-9d21-4107-86cd-f9cee40d08b8\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.755670 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8688d151-9d21-4107-86cd-f9cee40d08b8-test-operator-ephemeral-workdir\") pod \"8688d151-9d21-4107-86cd-f9cee40d08b8\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.756601 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-openstack-config-secret\") pod \"8688d151-9d21-4107-86cd-f9cee40d08b8\" (UID: \"8688d151-9d21-4107-86cd-f9cee40d08b8\") " Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.756741 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8688d151-9d21-4107-86cd-f9cee40d08b8-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "8688d151-9d21-4107-86cd-f9cee40d08b8" (UID: "8688d151-9d21-4107-86cd-f9cee40d08b8"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.757170 4997 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8688d151-9d21-4107-86cd-f9cee40d08b8-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.761139 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8688d151-9d21-4107-86cd-f9cee40d08b8-kube-api-access-zn9zx" (OuterVolumeSpecName: "kube-api-access-zn9zx") pod "8688d151-9d21-4107-86cd-f9cee40d08b8" (UID: "8688d151-9d21-4107-86cd-f9cee40d08b8"). InnerVolumeSpecName "kube-api-access-zn9zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.761171 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "test-operator-logs") pod "8688d151-9d21-4107-86cd-f9cee40d08b8" (UID: "8688d151-9d21-4107-86cd-f9cee40d08b8"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.761993 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8688d151-9d21-4107-86cd-f9cee40d08b8-config-data" (OuterVolumeSpecName: "config-data") pod "8688d151-9d21-4107-86cd-f9cee40d08b8" (UID: "8688d151-9d21-4107-86cd-f9cee40d08b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.764040 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8688d151-9d21-4107-86cd-f9cee40d08b8-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "8688d151-9d21-4107-86cd-f9cee40d08b8" (UID: "8688d151-9d21-4107-86cd-f9cee40d08b8"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.809715 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8688d151-9d21-4107-86cd-f9cee40d08b8" (UID: "8688d151-9d21-4107-86cd-f9cee40d08b8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.809757 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "8688d151-9d21-4107-86cd-f9cee40d08b8" (UID: "8688d151-9d21-4107-86cd-f9cee40d08b8"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.809892 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "8688d151-9d21-4107-86cd-f9cee40d08b8" (UID: "8688d151-9d21-4107-86cd-f9cee40d08b8"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.811303 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8688d151-9d21-4107-86cd-f9cee40d08b8-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "8688d151-9d21-4107-86cd-f9cee40d08b8" (UID: "8688d151-9d21-4107-86cd-f9cee40d08b8"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.859761 4997 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.860692 4997 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8688d151-9d21-4107-86cd-f9cee40d08b8-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.860709 4997 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.860750 4997 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.860770 4997 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8688d151-9d21-4107-86cd-f9cee40d08b8-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.860789 4997 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8688d151-9d21-4107-86cd-f9cee40d08b8-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.860807 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn9zx\" (UniqueName: \"kubernetes.io/projected/8688d151-9d21-4107-86cd-f9cee40d08b8-kube-api-access-zn9zx\") on node \"crc\" DevicePath \"\"" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.860822 4997 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8688d151-9d21-4107-86cd-f9cee40d08b8-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.882887 4997 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 05 10:27:26 crc kubenswrapper[4997]: I1205 10:27:26.962709 4997 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 05 10:27:27 crc kubenswrapper[4997]: I1205 10:27:27.255812 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8688d151-9d21-4107-86cd-f9cee40d08b8","Type":"ContainerDied","Data":"9617e0c85ed24cd7886c6da10dd3b69b803d3f3cd30ecd4a3085d87a5323fb87"} Dec 05 10:27:27 crc kubenswrapper[4997]: I1205 10:27:27.255852 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9617e0c85ed24cd7886c6da10dd3b69b803d3f3cd30ecd4a3085d87a5323fb87" Dec 05 10:27:27 crc kubenswrapper[4997]: I1205 10:27:27.255938 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 05 10:27:29 crc kubenswrapper[4997]: I1205 10:27:29.721146 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 05 10:27:29 crc kubenswrapper[4997]: E1205 10:27:29.722120 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80acf65b-25bd-461b-9169-37a8af196319" containerName="registry-server" Dec 05 10:27:29 crc kubenswrapper[4997]: I1205 10:27:29.722157 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="80acf65b-25bd-461b-9169-37a8af196319" containerName="registry-server" Dec 05 10:27:29 crc kubenswrapper[4997]: E1205 10:27:29.722210 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80acf65b-25bd-461b-9169-37a8af196319" containerName="extract-content" Dec 05 10:27:29 crc kubenswrapper[4997]: I1205 10:27:29.722221 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="80acf65b-25bd-461b-9169-37a8af196319" containerName="extract-content" Dec 05 10:27:29 crc kubenswrapper[4997]: E1205 10:27:29.722233 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8688d151-9d21-4107-86cd-f9cee40d08b8" containerName="tempest-tests-tempest-tests-runner" Dec 05 10:27:29 crc kubenswrapper[4997]: I1205 10:27:29.722241 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8688d151-9d21-4107-86cd-f9cee40d08b8" containerName="tempest-tests-tempest-tests-runner" Dec 05 10:27:29 crc kubenswrapper[4997]: E1205 10:27:29.722262 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80acf65b-25bd-461b-9169-37a8af196319" containerName="extract-utilities" Dec 05 10:27:29 crc kubenswrapper[4997]: I1205 10:27:29.722281 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="80acf65b-25bd-461b-9169-37a8af196319" containerName="extract-utilities" Dec 05 10:27:29 crc kubenswrapper[4997]: I1205 10:27:29.722504 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="80acf65b-25bd-461b-9169-37a8af196319" containerName="registry-server" Dec 05 10:27:29 crc kubenswrapper[4997]: I1205 10:27:29.722521 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="8688d151-9d21-4107-86cd-f9cee40d08b8" containerName="tempest-tests-tempest-tests-runner" Dec 05 10:27:29 crc kubenswrapper[4997]: I1205 10:27:29.723308 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:27:29 crc kubenswrapper[4997]: I1205 10:27:29.725582 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-qsjdm" Dec 05 10:27:29 crc kubenswrapper[4997]: I1205 10:27:29.734060 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 05 10:27:29 crc kubenswrapper[4997]: I1205 10:27:29.824328 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d45c99f0-7cf6-4abc-85d3-529596e9f01a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:27:29 crc kubenswrapper[4997]: I1205 10:27:29.824422 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4zm9\" (UniqueName: \"kubernetes.io/projected/d45c99f0-7cf6-4abc-85d3-529596e9f01a-kube-api-access-f4zm9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d45c99f0-7cf6-4abc-85d3-529596e9f01a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:27:29 crc kubenswrapper[4997]: I1205 10:27:29.927430 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d45c99f0-7cf6-4abc-85d3-529596e9f01a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:27:29 crc kubenswrapper[4997]: I1205 10:27:29.927527 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4zm9\" (UniqueName: \"kubernetes.io/projected/d45c99f0-7cf6-4abc-85d3-529596e9f01a-kube-api-access-f4zm9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d45c99f0-7cf6-4abc-85d3-529596e9f01a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:27:29 crc kubenswrapper[4997]: I1205 10:27:29.927890 4997 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d45c99f0-7cf6-4abc-85d3-529596e9f01a\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:27:29 crc kubenswrapper[4997]: I1205 10:27:29.952129 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4zm9\" (UniqueName: \"kubernetes.io/projected/d45c99f0-7cf6-4abc-85d3-529596e9f01a-kube-api-access-f4zm9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d45c99f0-7cf6-4abc-85d3-529596e9f01a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:27:29 crc kubenswrapper[4997]: I1205 10:27:29.979174 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d45c99f0-7cf6-4abc-85d3-529596e9f01a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:27:30 crc kubenswrapper[4997]: I1205 10:27:30.049707 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:27:30 crc kubenswrapper[4997]: I1205 10:27:30.503757 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 05 10:27:30 crc kubenswrapper[4997]: I1205 10:27:30.514114 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:27:31 crc kubenswrapper[4997]: I1205 10:27:31.306448 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"d45c99f0-7cf6-4abc-85d3-529596e9f01a","Type":"ContainerStarted","Data":"bb83220130b273bffe73213b36d759429b60b847d6072226a0f2d97213f72488"} Dec 05 10:27:32 crc kubenswrapper[4997]: I1205 10:27:32.335643 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"d45c99f0-7cf6-4abc-85d3-529596e9f01a","Type":"ContainerStarted","Data":"21a7292f918ad75a2ecc7276c2bc0051c35c3012b0d8a3ff204b59dbcc12bc59"} Dec 05 10:27:32 crc kubenswrapper[4997]: I1205 10:27:32.390055 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.422594227 podStartE2EDuration="3.390033411s" podCreationTimestamp="2025-12-05 10:27:29 +0000 UTC" firstStartedPulling="2025-12-05 10:27:30.513812548 +0000 UTC m=+12751.042719809" lastFinishedPulling="2025-12-05 10:27:31.481251692 +0000 UTC m=+12752.010158993" observedRunningTime="2025-12-05 10:27:32.36672209 +0000 UTC m=+12752.895629381" watchObservedRunningTime="2025-12-05 10:27:32.390033411 +0000 UTC m=+12752.918940672" Dec 05 10:27:38 crc kubenswrapper[4997]: I1205 10:27:38.750015 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:27:38 crc kubenswrapper[4997]: E1205 10:27:38.750847 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:27:51 crc kubenswrapper[4997]: I1205 10:27:51.749028 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:27:52 crc kubenswrapper[4997]: I1205 10:27:52.560736 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"570182e001027ed044c074b2d6343f7b4f00f9a0f33006cc726adf1d3379a622"} Dec 05 10:28:55 crc kubenswrapper[4997]: I1205 10:28:55.124363 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-w7q74/must-gather-mnk5f"] Dec 05 10:28:55 crc kubenswrapper[4997]: I1205 10:28:55.127290 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w7q74/must-gather-mnk5f" Dec 05 10:28:55 crc kubenswrapper[4997]: I1205 10:28:55.129282 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-w7q74"/"kube-root-ca.crt" Dec 05 10:28:55 crc kubenswrapper[4997]: I1205 10:28:55.129566 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-w7q74"/"default-dockercfg-vhvm5" Dec 05 10:28:55 crc kubenswrapper[4997]: I1205 10:28:55.129790 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-w7q74"/"openshift-service-ca.crt" Dec 05 10:28:55 crc kubenswrapper[4997]: I1205 10:28:55.142329 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-w7q74/must-gather-mnk5f"] Dec 05 10:28:55 crc kubenswrapper[4997]: I1205 10:28:55.289917 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b7478a1e-4ad6-421a-ab0f-9795d95be179-must-gather-output\") pod \"must-gather-mnk5f\" (UID: \"b7478a1e-4ad6-421a-ab0f-9795d95be179\") " pod="openshift-must-gather-w7q74/must-gather-mnk5f" Dec 05 10:28:55 crc kubenswrapper[4997]: I1205 10:28:55.290101 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjc4d\" (UniqueName: \"kubernetes.io/projected/b7478a1e-4ad6-421a-ab0f-9795d95be179-kube-api-access-kjc4d\") pod \"must-gather-mnk5f\" (UID: \"b7478a1e-4ad6-421a-ab0f-9795d95be179\") " pod="openshift-must-gather-w7q74/must-gather-mnk5f" Dec 05 10:28:55 crc kubenswrapper[4997]: I1205 10:28:55.392203 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjc4d\" (UniqueName: \"kubernetes.io/projected/b7478a1e-4ad6-421a-ab0f-9795d95be179-kube-api-access-kjc4d\") pod \"must-gather-mnk5f\" (UID: \"b7478a1e-4ad6-421a-ab0f-9795d95be179\") " pod="openshift-must-gather-w7q74/must-gather-mnk5f" Dec 05 10:28:55 crc kubenswrapper[4997]: I1205 10:28:55.392335 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b7478a1e-4ad6-421a-ab0f-9795d95be179-must-gather-output\") pod \"must-gather-mnk5f\" (UID: \"b7478a1e-4ad6-421a-ab0f-9795d95be179\") " pod="openshift-must-gather-w7q74/must-gather-mnk5f" Dec 05 10:28:55 crc kubenswrapper[4997]: I1205 10:28:55.392851 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b7478a1e-4ad6-421a-ab0f-9795d95be179-must-gather-output\") pod \"must-gather-mnk5f\" (UID: \"b7478a1e-4ad6-421a-ab0f-9795d95be179\") " pod="openshift-must-gather-w7q74/must-gather-mnk5f" Dec 05 10:28:55 crc kubenswrapper[4997]: I1205 10:28:55.410836 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjc4d\" (UniqueName: \"kubernetes.io/projected/b7478a1e-4ad6-421a-ab0f-9795d95be179-kube-api-access-kjc4d\") pod \"must-gather-mnk5f\" (UID: \"b7478a1e-4ad6-421a-ab0f-9795d95be179\") " pod="openshift-must-gather-w7q74/must-gather-mnk5f" Dec 05 10:28:55 crc kubenswrapper[4997]: I1205 10:28:55.451067 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w7q74/must-gather-mnk5f" Dec 05 10:28:56 crc kubenswrapper[4997]: I1205 10:28:56.020239 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-w7q74/must-gather-mnk5f"] Dec 05 10:28:56 crc kubenswrapper[4997]: I1205 10:28:56.275155 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w7q74/must-gather-mnk5f" event={"ID":"b7478a1e-4ad6-421a-ab0f-9795d95be179","Type":"ContainerStarted","Data":"5515eee8eb0fd733eb755f9bc8eb9d6e28382fa8d64a5dc21322b38c38c28bcf"} Dec 05 10:29:00 crc kubenswrapper[4997]: I1205 10:29:00.330262 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w7q74/must-gather-mnk5f" event={"ID":"b7478a1e-4ad6-421a-ab0f-9795d95be179","Type":"ContainerStarted","Data":"2050dfdcf13f3aec023eff86ac4567094633839815a0aae97102a7c2f8a21b52"} Dec 05 10:29:01 crc kubenswrapper[4997]: I1205 10:29:01.357563 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w7q74/must-gather-mnk5f" event={"ID":"b7478a1e-4ad6-421a-ab0f-9795d95be179","Type":"ContainerStarted","Data":"6513b33bd9eca34b0e643050c3d9e2cea0cca4ce48f4f79879ec30f2ab93f68f"} Dec 05 10:29:01 crc kubenswrapper[4997]: I1205 10:29:01.378170 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-w7q74/must-gather-mnk5f" podStartSLOduration=2.785419331 podStartE2EDuration="6.378150639s" podCreationTimestamp="2025-12-05 10:28:55 +0000 UTC" firstStartedPulling="2025-12-05 10:28:56.017589598 +0000 UTC m=+12836.546496859" lastFinishedPulling="2025-12-05 10:28:59.610320896 +0000 UTC m=+12840.139228167" observedRunningTime="2025-12-05 10:29:01.374241993 +0000 UTC m=+12841.903149264" watchObservedRunningTime="2025-12-05 10:29:01.378150639 +0000 UTC m=+12841.907057900" Dec 05 10:29:05 crc kubenswrapper[4997]: I1205 10:29:05.877905 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-w7q74/crc-debug-p2v4t"] Dec 05 10:29:05 crc kubenswrapper[4997]: I1205 10:29:05.880042 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w7q74/crc-debug-p2v4t" Dec 05 10:29:06 crc kubenswrapper[4997]: I1205 10:29:06.040387 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bffd6b65-c6f6-46ee-8179-acb6082c54f4-host\") pod \"crc-debug-p2v4t\" (UID: \"bffd6b65-c6f6-46ee-8179-acb6082c54f4\") " pod="openshift-must-gather-w7q74/crc-debug-p2v4t" Dec 05 10:29:06 crc kubenswrapper[4997]: I1205 10:29:06.040925 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4xnh\" (UniqueName: \"kubernetes.io/projected/bffd6b65-c6f6-46ee-8179-acb6082c54f4-kube-api-access-t4xnh\") pod \"crc-debug-p2v4t\" (UID: \"bffd6b65-c6f6-46ee-8179-acb6082c54f4\") " pod="openshift-must-gather-w7q74/crc-debug-p2v4t" Dec 05 10:29:06 crc kubenswrapper[4997]: I1205 10:29:06.143378 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4xnh\" (UniqueName: \"kubernetes.io/projected/bffd6b65-c6f6-46ee-8179-acb6082c54f4-kube-api-access-t4xnh\") pod \"crc-debug-p2v4t\" (UID: \"bffd6b65-c6f6-46ee-8179-acb6082c54f4\") " pod="openshift-must-gather-w7q74/crc-debug-p2v4t" Dec 05 10:29:06 crc kubenswrapper[4997]: I1205 10:29:06.143495 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bffd6b65-c6f6-46ee-8179-acb6082c54f4-host\") pod \"crc-debug-p2v4t\" (UID: \"bffd6b65-c6f6-46ee-8179-acb6082c54f4\") " pod="openshift-must-gather-w7q74/crc-debug-p2v4t" Dec 05 10:29:06 crc kubenswrapper[4997]: I1205 10:29:06.143760 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bffd6b65-c6f6-46ee-8179-acb6082c54f4-host\") pod \"crc-debug-p2v4t\" (UID: \"bffd6b65-c6f6-46ee-8179-acb6082c54f4\") " pod="openshift-must-gather-w7q74/crc-debug-p2v4t" Dec 05 10:29:06 crc kubenswrapper[4997]: I1205 10:29:06.174474 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4xnh\" (UniqueName: \"kubernetes.io/projected/bffd6b65-c6f6-46ee-8179-acb6082c54f4-kube-api-access-t4xnh\") pod \"crc-debug-p2v4t\" (UID: \"bffd6b65-c6f6-46ee-8179-acb6082c54f4\") " pod="openshift-must-gather-w7q74/crc-debug-p2v4t" Dec 05 10:29:06 crc kubenswrapper[4997]: I1205 10:29:06.200163 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w7q74/crc-debug-p2v4t" Dec 05 10:29:06 crc kubenswrapper[4997]: I1205 10:29:06.407332 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w7q74/crc-debug-p2v4t" event={"ID":"bffd6b65-c6f6-46ee-8179-acb6082c54f4","Type":"ContainerStarted","Data":"f74f71b2f0b408b6f956172312f15b07744d3e63687d22f4713290c34ead0340"} Dec 05 10:29:20 crc kubenswrapper[4997]: I1205 10:29:20.581176 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w7q74/crc-debug-p2v4t" event={"ID":"bffd6b65-c6f6-46ee-8179-acb6082c54f4","Type":"ContainerStarted","Data":"00e3b8aabdd0935bcf20f2590b78be9b23872763fa0a167112e159c91ef55cde"} Dec 05 10:29:20 crc kubenswrapper[4997]: I1205 10:29:20.593353 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-w7q74/crc-debug-p2v4t" podStartSLOduration=2.131573763 podStartE2EDuration="15.593332337s" podCreationTimestamp="2025-12-05 10:29:05 +0000 UTC" firstStartedPulling="2025-12-05 10:29:06.247255714 +0000 UTC m=+12846.776162976" lastFinishedPulling="2025-12-05 10:29:19.709014289 +0000 UTC m=+12860.237921550" observedRunningTime="2025-12-05 10:29:20.591684482 +0000 UTC m=+12861.120591753" watchObservedRunningTime="2025-12-05 10:29:20.593332337 +0000 UTC m=+12861.122239598" Dec 05 10:30:00 crc kubenswrapper[4997]: I1205 10:30:00.156945 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh"] Dec 05 10:30:00 crc kubenswrapper[4997]: I1205 10:30:00.159170 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh" Dec 05 10:30:00 crc kubenswrapper[4997]: I1205 10:30:00.161753 4997 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 10:30:00 crc kubenswrapper[4997]: I1205 10:30:00.164333 4997 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 10:30:00 crc kubenswrapper[4997]: I1205 10:30:00.169164 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh"] Dec 05 10:30:00 crc kubenswrapper[4997]: I1205 10:30:00.249793 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gtw9\" (UniqueName: \"kubernetes.io/projected/4fd12091-e937-46df-a4e1-5afc4511a744-kube-api-access-6gtw9\") pod \"collect-profiles-29415510-98svh\" (UID: \"4fd12091-e937-46df-a4e1-5afc4511a744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh" Dec 05 10:30:00 crc kubenswrapper[4997]: I1205 10:30:00.249874 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4fd12091-e937-46df-a4e1-5afc4511a744-secret-volume\") pod \"collect-profiles-29415510-98svh\" (UID: \"4fd12091-e937-46df-a4e1-5afc4511a744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh" Dec 05 10:30:00 crc kubenswrapper[4997]: I1205 10:30:00.250368 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4fd12091-e937-46df-a4e1-5afc4511a744-config-volume\") pod \"collect-profiles-29415510-98svh\" (UID: \"4fd12091-e937-46df-a4e1-5afc4511a744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh" Dec 05 10:30:00 crc kubenswrapper[4997]: I1205 10:30:00.352974 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4fd12091-e937-46df-a4e1-5afc4511a744-config-volume\") pod \"collect-profiles-29415510-98svh\" (UID: \"4fd12091-e937-46df-a4e1-5afc4511a744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh" Dec 05 10:30:00 crc kubenswrapper[4997]: I1205 10:30:00.353080 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gtw9\" (UniqueName: \"kubernetes.io/projected/4fd12091-e937-46df-a4e1-5afc4511a744-kube-api-access-6gtw9\") pod \"collect-profiles-29415510-98svh\" (UID: \"4fd12091-e937-46df-a4e1-5afc4511a744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh" Dec 05 10:30:00 crc kubenswrapper[4997]: I1205 10:30:00.353145 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4fd12091-e937-46df-a4e1-5afc4511a744-secret-volume\") pod \"collect-profiles-29415510-98svh\" (UID: \"4fd12091-e937-46df-a4e1-5afc4511a744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh" Dec 05 10:30:00 crc kubenswrapper[4997]: I1205 10:30:00.354005 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4fd12091-e937-46df-a4e1-5afc4511a744-config-volume\") pod \"collect-profiles-29415510-98svh\" (UID: \"4fd12091-e937-46df-a4e1-5afc4511a744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh" Dec 05 10:30:00 crc kubenswrapper[4997]: I1205 10:30:00.366156 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4fd12091-e937-46df-a4e1-5afc4511a744-secret-volume\") pod \"collect-profiles-29415510-98svh\" (UID: \"4fd12091-e937-46df-a4e1-5afc4511a744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh" Dec 05 10:30:00 crc kubenswrapper[4997]: I1205 10:30:00.379425 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gtw9\" (UniqueName: \"kubernetes.io/projected/4fd12091-e937-46df-a4e1-5afc4511a744-kube-api-access-6gtw9\") pod \"collect-profiles-29415510-98svh\" (UID: \"4fd12091-e937-46df-a4e1-5afc4511a744\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh" Dec 05 10:30:00 crc kubenswrapper[4997]: I1205 10:30:00.488079 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh" Dec 05 10:30:01 crc kubenswrapper[4997]: I1205 10:30:01.181555 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh"] Dec 05 10:30:02 crc kubenswrapper[4997]: I1205 10:30:02.043282 4997 generic.go:334] "Generic (PLEG): container finished" podID="4fd12091-e937-46df-a4e1-5afc4511a744" containerID="50466c9f7b0babd053fbd4b45408f9b4e2cd2f1e0747cf0c8d3c1e932a869a9e" exitCode=0 Dec 05 10:30:02 crc kubenswrapper[4997]: I1205 10:30:02.043714 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh" event={"ID":"4fd12091-e937-46df-a4e1-5afc4511a744","Type":"ContainerDied","Data":"50466c9f7b0babd053fbd4b45408f9b4e2cd2f1e0747cf0c8d3c1e932a869a9e"} Dec 05 10:30:02 crc kubenswrapper[4997]: I1205 10:30:02.043782 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh" event={"ID":"4fd12091-e937-46df-a4e1-5afc4511a744","Type":"ContainerStarted","Data":"145776b18d744d7945d4d40664519c27b0bdb92bc508d3425b7d0c5edf2975ba"} Dec 05 10:30:03 crc kubenswrapper[4997]: I1205 10:30:03.465249 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh" Dec 05 10:30:03 crc kubenswrapper[4997]: I1205 10:30:03.533886 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4fd12091-e937-46df-a4e1-5afc4511a744-secret-volume\") pod \"4fd12091-e937-46df-a4e1-5afc4511a744\" (UID: \"4fd12091-e937-46df-a4e1-5afc4511a744\") " Dec 05 10:30:03 crc kubenswrapper[4997]: I1205 10:30:03.534010 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gtw9\" (UniqueName: \"kubernetes.io/projected/4fd12091-e937-46df-a4e1-5afc4511a744-kube-api-access-6gtw9\") pod \"4fd12091-e937-46df-a4e1-5afc4511a744\" (UID: \"4fd12091-e937-46df-a4e1-5afc4511a744\") " Dec 05 10:30:03 crc kubenswrapper[4997]: I1205 10:30:03.534067 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4fd12091-e937-46df-a4e1-5afc4511a744-config-volume\") pod \"4fd12091-e937-46df-a4e1-5afc4511a744\" (UID: \"4fd12091-e937-46df-a4e1-5afc4511a744\") " Dec 05 10:30:03 crc kubenswrapper[4997]: I1205 10:30:03.534898 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fd12091-e937-46df-a4e1-5afc4511a744-config-volume" (OuterVolumeSpecName: "config-volume") pod "4fd12091-e937-46df-a4e1-5afc4511a744" (UID: "4fd12091-e937-46df-a4e1-5afc4511a744"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:30:03 crc kubenswrapper[4997]: I1205 10:30:03.535681 4997 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4fd12091-e937-46df-a4e1-5afc4511a744-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:03 crc kubenswrapper[4997]: I1205 10:30:03.539815 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fd12091-e937-46df-a4e1-5afc4511a744-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4fd12091-e937-46df-a4e1-5afc4511a744" (UID: "4fd12091-e937-46df-a4e1-5afc4511a744"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:30:03 crc kubenswrapper[4997]: I1205 10:30:03.544751 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fd12091-e937-46df-a4e1-5afc4511a744-kube-api-access-6gtw9" (OuterVolumeSpecName: "kube-api-access-6gtw9") pod "4fd12091-e937-46df-a4e1-5afc4511a744" (UID: "4fd12091-e937-46df-a4e1-5afc4511a744"). InnerVolumeSpecName "kube-api-access-6gtw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:30:03 crc kubenswrapper[4997]: I1205 10:30:03.637696 4997 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4fd12091-e937-46df-a4e1-5afc4511a744-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:03 crc kubenswrapper[4997]: I1205 10:30:03.637750 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gtw9\" (UniqueName: \"kubernetes.io/projected/4fd12091-e937-46df-a4e1-5afc4511a744-kube-api-access-6gtw9\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:04 crc kubenswrapper[4997]: I1205 10:30:04.068456 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh" event={"ID":"4fd12091-e937-46df-a4e1-5afc4511a744","Type":"ContainerDied","Data":"145776b18d744d7945d4d40664519c27b0bdb92bc508d3425b7d0c5edf2975ba"} Dec 05 10:30:04 crc kubenswrapper[4997]: I1205 10:30:04.068501 4997 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="145776b18d744d7945d4d40664519c27b0bdb92bc508d3425b7d0c5edf2975ba" Dec 05 10:30:04 crc kubenswrapper[4997]: I1205 10:30:04.068570 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-98svh" Dec 05 10:30:04 crc kubenswrapper[4997]: I1205 10:30:04.549267 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4"] Dec 05 10:30:04 crc kubenswrapper[4997]: I1205 10:30:04.563472 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415465-w57s4"] Dec 05 10:30:05 crc kubenswrapper[4997]: I1205 10:30:05.763850 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a04c950b-c9b4-4052-bad7-53a68c7568c5" path="/var/lib/kubelet/pods/a04c950b-c9b4-4052-bad7-53a68c7568c5/volumes" Dec 05 10:30:18 crc kubenswrapper[4997]: I1205 10:30:18.225908 4997 generic.go:334] "Generic (PLEG): container finished" podID="bffd6b65-c6f6-46ee-8179-acb6082c54f4" containerID="00e3b8aabdd0935bcf20f2590b78be9b23872763fa0a167112e159c91ef55cde" exitCode=0 Dec 05 10:30:18 crc kubenswrapper[4997]: I1205 10:30:18.225978 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w7q74/crc-debug-p2v4t" event={"ID":"bffd6b65-c6f6-46ee-8179-acb6082c54f4","Type":"ContainerDied","Data":"00e3b8aabdd0935bcf20f2590b78be9b23872763fa0a167112e159c91ef55cde"} Dec 05 10:30:19 crc kubenswrapper[4997]: I1205 10:30:19.340837 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w7q74/crc-debug-p2v4t" Dec 05 10:30:19 crc kubenswrapper[4997]: I1205 10:30:19.391230 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-w7q74/crc-debug-p2v4t"] Dec 05 10:30:19 crc kubenswrapper[4997]: I1205 10:30:19.400847 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-w7q74/crc-debug-p2v4t"] Dec 05 10:30:19 crc kubenswrapper[4997]: I1205 10:30:19.459844 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4xnh\" (UniqueName: \"kubernetes.io/projected/bffd6b65-c6f6-46ee-8179-acb6082c54f4-kube-api-access-t4xnh\") pod \"bffd6b65-c6f6-46ee-8179-acb6082c54f4\" (UID: \"bffd6b65-c6f6-46ee-8179-acb6082c54f4\") " Dec 05 10:30:19 crc kubenswrapper[4997]: I1205 10:30:19.460118 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bffd6b65-c6f6-46ee-8179-acb6082c54f4-host\") pod \"bffd6b65-c6f6-46ee-8179-acb6082c54f4\" (UID: \"bffd6b65-c6f6-46ee-8179-acb6082c54f4\") " Dec 05 10:30:19 crc kubenswrapper[4997]: I1205 10:30:19.460244 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bffd6b65-c6f6-46ee-8179-acb6082c54f4-host" (OuterVolumeSpecName: "host") pod "bffd6b65-c6f6-46ee-8179-acb6082c54f4" (UID: "bffd6b65-c6f6-46ee-8179-acb6082c54f4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:30:19 crc kubenswrapper[4997]: I1205 10:30:19.461504 4997 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bffd6b65-c6f6-46ee-8179-acb6082c54f4-host\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:19 crc kubenswrapper[4997]: I1205 10:30:19.468998 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bffd6b65-c6f6-46ee-8179-acb6082c54f4-kube-api-access-t4xnh" (OuterVolumeSpecName: "kube-api-access-t4xnh") pod "bffd6b65-c6f6-46ee-8179-acb6082c54f4" (UID: "bffd6b65-c6f6-46ee-8179-acb6082c54f4"). InnerVolumeSpecName "kube-api-access-t4xnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:30:19 crc kubenswrapper[4997]: I1205 10:30:19.563659 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4xnh\" (UniqueName: \"kubernetes.io/projected/bffd6b65-c6f6-46ee-8179-acb6082c54f4-kube-api-access-t4xnh\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:19 crc kubenswrapper[4997]: I1205 10:30:19.761675 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bffd6b65-c6f6-46ee-8179-acb6082c54f4" path="/var/lib/kubelet/pods/bffd6b65-c6f6-46ee-8179-acb6082c54f4/volumes" Dec 05 10:30:19 crc kubenswrapper[4997]: I1205 10:30:19.769945 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:30:19 crc kubenswrapper[4997]: I1205 10:30:19.770003 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:30:20 crc kubenswrapper[4997]: I1205 10:30:20.246226 4997 scope.go:117] "RemoveContainer" containerID="00e3b8aabdd0935bcf20f2590b78be9b23872763fa0a167112e159c91ef55cde" Dec 05 10:30:20 crc kubenswrapper[4997]: I1205 10:30:20.246263 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w7q74/crc-debug-p2v4t" Dec 05 10:30:20 crc kubenswrapper[4997]: I1205 10:30:20.550407 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-w7q74/crc-debug-wzsp8"] Dec 05 10:30:20 crc kubenswrapper[4997]: E1205 10:30:20.550944 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fd12091-e937-46df-a4e1-5afc4511a744" containerName="collect-profiles" Dec 05 10:30:20 crc kubenswrapper[4997]: I1205 10:30:20.550959 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fd12091-e937-46df-a4e1-5afc4511a744" containerName="collect-profiles" Dec 05 10:30:20 crc kubenswrapper[4997]: E1205 10:30:20.550996 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bffd6b65-c6f6-46ee-8179-acb6082c54f4" containerName="container-00" Dec 05 10:30:20 crc kubenswrapper[4997]: I1205 10:30:20.551005 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="bffd6b65-c6f6-46ee-8179-acb6082c54f4" containerName="container-00" Dec 05 10:30:20 crc kubenswrapper[4997]: I1205 10:30:20.551248 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fd12091-e937-46df-a4e1-5afc4511a744" containerName="collect-profiles" Dec 05 10:30:20 crc kubenswrapper[4997]: I1205 10:30:20.551283 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="bffd6b65-c6f6-46ee-8179-acb6082c54f4" containerName="container-00" Dec 05 10:30:20 crc kubenswrapper[4997]: I1205 10:30:20.552192 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w7q74/crc-debug-wzsp8" Dec 05 10:30:20 crc kubenswrapper[4997]: I1205 10:30:20.583700 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9da9f229-7923-4e41-ba0a-a135f19f72fa-host\") pod \"crc-debug-wzsp8\" (UID: \"9da9f229-7923-4e41-ba0a-a135f19f72fa\") " pod="openshift-must-gather-w7q74/crc-debug-wzsp8" Dec 05 10:30:20 crc kubenswrapper[4997]: I1205 10:30:20.583757 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj24m\" (UniqueName: \"kubernetes.io/projected/9da9f229-7923-4e41-ba0a-a135f19f72fa-kube-api-access-zj24m\") pod \"crc-debug-wzsp8\" (UID: \"9da9f229-7923-4e41-ba0a-a135f19f72fa\") " pod="openshift-must-gather-w7q74/crc-debug-wzsp8" Dec 05 10:30:20 crc kubenswrapper[4997]: I1205 10:30:20.685536 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9da9f229-7923-4e41-ba0a-a135f19f72fa-host\") pod \"crc-debug-wzsp8\" (UID: \"9da9f229-7923-4e41-ba0a-a135f19f72fa\") " pod="openshift-must-gather-w7q74/crc-debug-wzsp8" Dec 05 10:30:20 crc kubenswrapper[4997]: I1205 10:30:20.685603 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj24m\" (UniqueName: \"kubernetes.io/projected/9da9f229-7923-4e41-ba0a-a135f19f72fa-kube-api-access-zj24m\") pod \"crc-debug-wzsp8\" (UID: \"9da9f229-7923-4e41-ba0a-a135f19f72fa\") " pod="openshift-must-gather-w7q74/crc-debug-wzsp8" Dec 05 10:30:20 crc kubenswrapper[4997]: I1205 10:30:20.685716 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9da9f229-7923-4e41-ba0a-a135f19f72fa-host\") pod \"crc-debug-wzsp8\" (UID: \"9da9f229-7923-4e41-ba0a-a135f19f72fa\") " pod="openshift-must-gather-w7q74/crc-debug-wzsp8" Dec 05 10:30:20 crc kubenswrapper[4997]: I1205 10:30:20.705534 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj24m\" (UniqueName: \"kubernetes.io/projected/9da9f229-7923-4e41-ba0a-a135f19f72fa-kube-api-access-zj24m\") pod \"crc-debug-wzsp8\" (UID: \"9da9f229-7923-4e41-ba0a-a135f19f72fa\") " pod="openshift-must-gather-w7q74/crc-debug-wzsp8" Dec 05 10:30:20 crc kubenswrapper[4997]: I1205 10:30:20.872885 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w7q74/crc-debug-wzsp8" Dec 05 10:30:21 crc kubenswrapper[4997]: I1205 10:30:21.256703 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w7q74/crc-debug-wzsp8" event={"ID":"9da9f229-7923-4e41-ba0a-a135f19f72fa","Type":"ContainerStarted","Data":"b700acd7f23b2dc921cc0ee7554efd2253c389bf9257c28b8f9bc827de3028fe"} Dec 05 10:30:21 crc kubenswrapper[4997]: I1205 10:30:21.256928 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w7q74/crc-debug-wzsp8" event={"ID":"9da9f229-7923-4e41-ba0a-a135f19f72fa","Type":"ContainerStarted","Data":"4ef68538e13419d889af75023b66e2fc8236c786648d3a53427778849f8c1e11"} Dec 05 10:30:22 crc kubenswrapper[4997]: I1205 10:30:22.272061 4997 generic.go:334] "Generic (PLEG): container finished" podID="9da9f229-7923-4e41-ba0a-a135f19f72fa" containerID="b700acd7f23b2dc921cc0ee7554efd2253c389bf9257c28b8f9bc827de3028fe" exitCode=0 Dec 05 10:30:22 crc kubenswrapper[4997]: I1205 10:30:22.272157 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w7q74/crc-debug-wzsp8" event={"ID":"9da9f229-7923-4e41-ba0a-a135f19f72fa","Type":"ContainerDied","Data":"b700acd7f23b2dc921cc0ee7554efd2253c389bf9257c28b8f9bc827de3028fe"} Dec 05 10:30:22 crc kubenswrapper[4997]: I1205 10:30:22.332566 4997 scope.go:117] "RemoveContainer" containerID="a9cfe68a24f23da0b5b08d562f31949505a689769f05eec7b642770c3df12856" Dec 05 10:30:23 crc kubenswrapper[4997]: I1205 10:30:23.098938 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-w7q74/crc-debug-wzsp8"] Dec 05 10:30:23 crc kubenswrapper[4997]: I1205 10:30:23.113443 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-w7q74/crc-debug-wzsp8"] Dec 05 10:30:23 crc kubenswrapper[4997]: I1205 10:30:23.393757 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w7q74/crc-debug-wzsp8" Dec 05 10:30:23 crc kubenswrapper[4997]: I1205 10:30:23.441188 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zj24m\" (UniqueName: \"kubernetes.io/projected/9da9f229-7923-4e41-ba0a-a135f19f72fa-kube-api-access-zj24m\") pod \"9da9f229-7923-4e41-ba0a-a135f19f72fa\" (UID: \"9da9f229-7923-4e41-ba0a-a135f19f72fa\") " Dec 05 10:30:23 crc kubenswrapper[4997]: I1205 10:30:23.441255 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9da9f229-7923-4e41-ba0a-a135f19f72fa-host\") pod \"9da9f229-7923-4e41-ba0a-a135f19f72fa\" (UID: \"9da9f229-7923-4e41-ba0a-a135f19f72fa\") " Dec 05 10:30:23 crc kubenswrapper[4997]: I1205 10:30:23.441677 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9da9f229-7923-4e41-ba0a-a135f19f72fa-host" (OuterVolumeSpecName: "host") pod "9da9f229-7923-4e41-ba0a-a135f19f72fa" (UID: "9da9f229-7923-4e41-ba0a-a135f19f72fa"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:30:23 crc kubenswrapper[4997]: I1205 10:30:23.442024 4997 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9da9f229-7923-4e41-ba0a-a135f19f72fa-host\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:23 crc kubenswrapper[4997]: I1205 10:30:23.448903 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9da9f229-7923-4e41-ba0a-a135f19f72fa-kube-api-access-zj24m" (OuterVolumeSpecName: "kube-api-access-zj24m") pod "9da9f229-7923-4e41-ba0a-a135f19f72fa" (UID: "9da9f229-7923-4e41-ba0a-a135f19f72fa"). InnerVolumeSpecName "kube-api-access-zj24m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:30:23 crc kubenswrapper[4997]: I1205 10:30:23.544361 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zj24m\" (UniqueName: \"kubernetes.io/projected/9da9f229-7923-4e41-ba0a-a135f19f72fa-kube-api-access-zj24m\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:23 crc kubenswrapper[4997]: I1205 10:30:23.760144 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9da9f229-7923-4e41-ba0a-a135f19f72fa" path="/var/lib/kubelet/pods/9da9f229-7923-4e41-ba0a-a135f19f72fa/volumes" Dec 05 10:30:24 crc kubenswrapper[4997]: I1205 10:30:24.304813 4997 scope.go:117] "RemoveContainer" containerID="b700acd7f23b2dc921cc0ee7554efd2253c389bf9257c28b8f9bc827de3028fe" Dec 05 10:30:24 crc kubenswrapper[4997]: I1205 10:30:24.305040 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w7q74/crc-debug-wzsp8" Dec 05 10:30:24 crc kubenswrapper[4997]: I1205 10:30:24.508948 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-w7q74/crc-debug-85pfx"] Dec 05 10:30:24 crc kubenswrapper[4997]: E1205 10:30:24.509369 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9da9f229-7923-4e41-ba0a-a135f19f72fa" containerName="container-00" Dec 05 10:30:24 crc kubenswrapper[4997]: I1205 10:30:24.509382 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="9da9f229-7923-4e41-ba0a-a135f19f72fa" containerName="container-00" Dec 05 10:30:24 crc kubenswrapper[4997]: I1205 10:30:24.509623 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="9da9f229-7923-4e41-ba0a-a135f19f72fa" containerName="container-00" Dec 05 10:30:24 crc kubenswrapper[4997]: I1205 10:30:24.510330 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w7q74/crc-debug-85pfx" Dec 05 10:30:24 crc kubenswrapper[4997]: I1205 10:30:24.564304 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk2jh\" (UniqueName: \"kubernetes.io/projected/8e2910cc-21c3-4642-b5dc-9bc583a595b7-kube-api-access-mk2jh\") pod \"crc-debug-85pfx\" (UID: \"8e2910cc-21c3-4642-b5dc-9bc583a595b7\") " pod="openshift-must-gather-w7q74/crc-debug-85pfx" Dec 05 10:30:24 crc kubenswrapper[4997]: I1205 10:30:24.564411 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8e2910cc-21c3-4642-b5dc-9bc583a595b7-host\") pod \"crc-debug-85pfx\" (UID: \"8e2910cc-21c3-4642-b5dc-9bc583a595b7\") " pod="openshift-must-gather-w7q74/crc-debug-85pfx" Dec 05 10:30:24 crc kubenswrapper[4997]: I1205 10:30:24.666720 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk2jh\" (UniqueName: \"kubernetes.io/projected/8e2910cc-21c3-4642-b5dc-9bc583a595b7-kube-api-access-mk2jh\") pod \"crc-debug-85pfx\" (UID: \"8e2910cc-21c3-4642-b5dc-9bc583a595b7\") " pod="openshift-must-gather-w7q74/crc-debug-85pfx" Dec 05 10:30:24 crc kubenswrapper[4997]: I1205 10:30:24.666804 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8e2910cc-21c3-4642-b5dc-9bc583a595b7-host\") pod \"crc-debug-85pfx\" (UID: \"8e2910cc-21c3-4642-b5dc-9bc583a595b7\") " pod="openshift-must-gather-w7q74/crc-debug-85pfx" Dec 05 10:30:24 crc kubenswrapper[4997]: I1205 10:30:24.666967 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8e2910cc-21c3-4642-b5dc-9bc583a595b7-host\") pod \"crc-debug-85pfx\" (UID: \"8e2910cc-21c3-4642-b5dc-9bc583a595b7\") " pod="openshift-must-gather-w7q74/crc-debug-85pfx" Dec 05 10:30:24 crc kubenswrapper[4997]: I1205 10:30:24.684800 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk2jh\" (UniqueName: \"kubernetes.io/projected/8e2910cc-21c3-4642-b5dc-9bc583a595b7-kube-api-access-mk2jh\") pod \"crc-debug-85pfx\" (UID: \"8e2910cc-21c3-4642-b5dc-9bc583a595b7\") " pod="openshift-must-gather-w7q74/crc-debug-85pfx" Dec 05 10:30:24 crc kubenswrapper[4997]: I1205 10:30:24.826454 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w7q74/crc-debug-85pfx" Dec 05 10:30:24 crc kubenswrapper[4997]: W1205 10:30:24.856460 4997 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e2910cc_21c3_4642_b5dc_9bc583a595b7.slice/crio-540ea2ddf7bfccd29e48dadbca5a659d63d7b004e44f4ef792c079bd179367fa WatchSource:0}: Error finding container 540ea2ddf7bfccd29e48dadbca5a659d63d7b004e44f4ef792c079bd179367fa: Status 404 returned error can't find the container with id 540ea2ddf7bfccd29e48dadbca5a659d63d7b004e44f4ef792c079bd179367fa Dec 05 10:30:25 crc kubenswrapper[4997]: I1205 10:30:25.319057 4997 generic.go:334] "Generic (PLEG): container finished" podID="8e2910cc-21c3-4642-b5dc-9bc583a595b7" containerID="1334a9f65978815b2f987303bf75de1612ce5034aed83255c41bc0f461a19175" exitCode=0 Dec 05 10:30:25 crc kubenswrapper[4997]: I1205 10:30:25.319280 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w7q74/crc-debug-85pfx" event={"ID":"8e2910cc-21c3-4642-b5dc-9bc583a595b7","Type":"ContainerDied","Data":"1334a9f65978815b2f987303bf75de1612ce5034aed83255c41bc0f461a19175"} Dec 05 10:30:25 crc kubenswrapper[4997]: I1205 10:30:25.319470 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w7q74/crc-debug-85pfx" event={"ID":"8e2910cc-21c3-4642-b5dc-9bc583a595b7","Type":"ContainerStarted","Data":"540ea2ddf7bfccd29e48dadbca5a659d63d7b004e44f4ef792c079bd179367fa"} Dec 05 10:30:25 crc kubenswrapper[4997]: I1205 10:30:25.358991 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-w7q74/crc-debug-85pfx"] Dec 05 10:30:25 crc kubenswrapper[4997]: I1205 10:30:25.371301 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-w7q74/crc-debug-85pfx"] Dec 05 10:30:26 crc kubenswrapper[4997]: I1205 10:30:26.443955 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w7q74/crc-debug-85pfx" Dec 05 10:30:26 crc kubenswrapper[4997]: I1205 10:30:26.510045 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8e2910cc-21c3-4642-b5dc-9bc583a595b7-host\") pod \"8e2910cc-21c3-4642-b5dc-9bc583a595b7\" (UID: \"8e2910cc-21c3-4642-b5dc-9bc583a595b7\") " Dec 05 10:30:26 crc kubenswrapper[4997]: I1205 10:30:26.510201 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mk2jh\" (UniqueName: \"kubernetes.io/projected/8e2910cc-21c3-4642-b5dc-9bc583a595b7-kube-api-access-mk2jh\") pod \"8e2910cc-21c3-4642-b5dc-9bc583a595b7\" (UID: \"8e2910cc-21c3-4642-b5dc-9bc583a595b7\") " Dec 05 10:30:26 crc kubenswrapper[4997]: I1205 10:30:26.510274 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e2910cc-21c3-4642-b5dc-9bc583a595b7-host" (OuterVolumeSpecName: "host") pod "8e2910cc-21c3-4642-b5dc-9bc583a595b7" (UID: "8e2910cc-21c3-4642-b5dc-9bc583a595b7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:30:26 crc kubenswrapper[4997]: I1205 10:30:26.510756 4997 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8e2910cc-21c3-4642-b5dc-9bc583a595b7-host\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:26 crc kubenswrapper[4997]: I1205 10:30:26.515135 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e2910cc-21c3-4642-b5dc-9bc583a595b7-kube-api-access-mk2jh" (OuterVolumeSpecName: "kube-api-access-mk2jh") pod "8e2910cc-21c3-4642-b5dc-9bc583a595b7" (UID: "8e2910cc-21c3-4642-b5dc-9bc583a595b7"). InnerVolumeSpecName "kube-api-access-mk2jh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:30:26 crc kubenswrapper[4997]: I1205 10:30:26.612884 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mk2jh\" (UniqueName: \"kubernetes.io/projected/8e2910cc-21c3-4642-b5dc-9bc583a595b7-kube-api-access-mk2jh\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:27 crc kubenswrapper[4997]: I1205 10:30:27.342314 4997 scope.go:117] "RemoveContainer" containerID="1334a9f65978815b2f987303bf75de1612ce5034aed83255c41bc0f461a19175" Dec 05 10:30:27 crc kubenswrapper[4997]: I1205 10:30:27.342676 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w7q74/crc-debug-85pfx" Dec 05 10:30:27 crc kubenswrapper[4997]: I1205 10:30:27.761896 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e2910cc-21c3-4642-b5dc-9bc583a595b7" path="/var/lib/kubelet/pods/8e2910cc-21c3-4642-b5dc-9bc583a595b7/volumes" Dec 05 10:30:42 crc kubenswrapper[4997]: I1205 10:30:42.859894 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bxms4"] Dec 05 10:30:42 crc kubenswrapper[4997]: E1205 10:30:42.862487 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e2910cc-21c3-4642-b5dc-9bc583a595b7" containerName="container-00" Dec 05 10:30:42 crc kubenswrapper[4997]: I1205 10:30:42.862528 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e2910cc-21c3-4642-b5dc-9bc583a595b7" containerName="container-00" Dec 05 10:30:42 crc kubenswrapper[4997]: I1205 10:30:42.863036 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e2910cc-21c3-4642-b5dc-9bc583a595b7" containerName="container-00" Dec 05 10:30:42 crc kubenswrapper[4997]: I1205 10:30:42.865235 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxms4" Dec 05 10:30:42 crc kubenswrapper[4997]: I1205 10:30:42.888835 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bxms4"] Dec 05 10:30:42 crc kubenswrapper[4997]: I1205 10:30:42.979170 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01fae4ef-8762-4aac-8a32-db29b9f76c0d-utilities\") pod \"certified-operators-bxms4\" (UID: \"01fae4ef-8762-4aac-8a32-db29b9f76c0d\") " pod="openshift-marketplace/certified-operators-bxms4" Dec 05 10:30:42 crc kubenswrapper[4997]: I1205 10:30:42.979337 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2lds\" (UniqueName: \"kubernetes.io/projected/01fae4ef-8762-4aac-8a32-db29b9f76c0d-kube-api-access-b2lds\") pod \"certified-operators-bxms4\" (UID: \"01fae4ef-8762-4aac-8a32-db29b9f76c0d\") " pod="openshift-marketplace/certified-operators-bxms4" Dec 05 10:30:42 crc kubenswrapper[4997]: I1205 10:30:42.979378 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01fae4ef-8762-4aac-8a32-db29b9f76c0d-catalog-content\") pod \"certified-operators-bxms4\" (UID: \"01fae4ef-8762-4aac-8a32-db29b9f76c0d\") " pod="openshift-marketplace/certified-operators-bxms4" Dec 05 10:30:43 crc kubenswrapper[4997]: I1205 10:30:43.080890 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2lds\" (UniqueName: \"kubernetes.io/projected/01fae4ef-8762-4aac-8a32-db29b9f76c0d-kube-api-access-b2lds\") pod \"certified-operators-bxms4\" (UID: \"01fae4ef-8762-4aac-8a32-db29b9f76c0d\") " pod="openshift-marketplace/certified-operators-bxms4" Dec 05 10:30:43 crc kubenswrapper[4997]: I1205 10:30:43.081261 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01fae4ef-8762-4aac-8a32-db29b9f76c0d-catalog-content\") pod \"certified-operators-bxms4\" (UID: \"01fae4ef-8762-4aac-8a32-db29b9f76c0d\") " pod="openshift-marketplace/certified-operators-bxms4" Dec 05 10:30:43 crc kubenswrapper[4997]: I1205 10:30:43.081437 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01fae4ef-8762-4aac-8a32-db29b9f76c0d-utilities\") pod \"certified-operators-bxms4\" (UID: \"01fae4ef-8762-4aac-8a32-db29b9f76c0d\") " pod="openshift-marketplace/certified-operators-bxms4" Dec 05 10:30:43 crc kubenswrapper[4997]: I1205 10:30:43.081761 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01fae4ef-8762-4aac-8a32-db29b9f76c0d-catalog-content\") pod \"certified-operators-bxms4\" (UID: \"01fae4ef-8762-4aac-8a32-db29b9f76c0d\") " pod="openshift-marketplace/certified-operators-bxms4" Dec 05 10:30:43 crc kubenswrapper[4997]: I1205 10:30:43.082035 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01fae4ef-8762-4aac-8a32-db29b9f76c0d-utilities\") pod \"certified-operators-bxms4\" (UID: \"01fae4ef-8762-4aac-8a32-db29b9f76c0d\") " pod="openshift-marketplace/certified-operators-bxms4" Dec 05 10:30:43 crc kubenswrapper[4997]: I1205 10:30:43.099160 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2lds\" (UniqueName: \"kubernetes.io/projected/01fae4ef-8762-4aac-8a32-db29b9f76c0d-kube-api-access-b2lds\") pod \"certified-operators-bxms4\" (UID: \"01fae4ef-8762-4aac-8a32-db29b9f76c0d\") " pod="openshift-marketplace/certified-operators-bxms4" Dec 05 10:30:43 crc kubenswrapper[4997]: I1205 10:30:43.210669 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxms4" Dec 05 10:30:43 crc kubenswrapper[4997]: I1205 10:30:43.560242 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bxms4"] Dec 05 10:30:43 crc kubenswrapper[4997]: I1205 10:30:43.572571 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxms4" event={"ID":"01fae4ef-8762-4aac-8a32-db29b9f76c0d","Type":"ContainerStarted","Data":"70eb461cbaf859cfb0177bad72788270b2e78a3063402b92bafdca05d9f10bf5"} Dec 05 10:30:44 crc kubenswrapper[4997]: I1205 10:30:44.587743 4997 generic.go:334] "Generic (PLEG): container finished" podID="01fae4ef-8762-4aac-8a32-db29b9f76c0d" containerID="dd7f3c25de06a420788ca688c08067d762040486fe600821caddd6828452bb4e" exitCode=0 Dec 05 10:30:44 crc kubenswrapper[4997]: I1205 10:30:44.587819 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxms4" event={"ID":"01fae4ef-8762-4aac-8a32-db29b9f76c0d","Type":"ContainerDied","Data":"dd7f3c25de06a420788ca688c08067d762040486fe600821caddd6828452bb4e"} Dec 05 10:30:45 crc kubenswrapper[4997]: I1205 10:30:45.600287 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxms4" event={"ID":"01fae4ef-8762-4aac-8a32-db29b9f76c0d","Type":"ContainerStarted","Data":"73e1cfb909f64c6e9fe727944533f8cbf7e0281f6568a559b5d14c0f6fe1528d"} Dec 05 10:30:46 crc kubenswrapper[4997]: I1205 10:30:46.612553 4997 generic.go:334] "Generic (PLEG): container finished" podID="01fae4ef-8762-4aac-8a32-db29b9f76c0d" containerID="73e1cfb909f64c6e9fe727944533f8cbf7e0281f6568a559b5d14c0f6fe1528d" exitCode=0 Dec 05 10:30:46 crc kubenswrapper[4997]: I1205 10:30:46.612627 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxms4" event={"ID":"01fae4ef-8762-4aac-8a32-db29b9f76c0d","Type":"ContainerDied","Data":"73e1cfb909f64c6e9fe727944533f8cbf7e0281f6568a559b5d14c0f6fe1528d"} Dec 05 10:30:47 crc kubenswrapper[4997]: I1205 10:30:47.625736 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxms4" event={"ID":"01fae4ef-8762-4aac-8a32-db29b9f76c0d","Type":"ContainerStarted","Data":"a9b4d3922e0825904afa982033a30604f00d467e331be401f3d3b912fc3289d7"} Dec 05 10:30:47 crc kubenswrapper[4997]: I1205 10:30:47.642188 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bxms4" podStartSLOduration=3.014886546 podStartE2EDuration="5.642173903s" podCreationTimestamp="2025-12-05 10:30:42 +0000 UTC" firstStartedPulling="2025-12-05 10:30:44.589376673 +0000 UTC m=+12945.118283934" lastFinishedPulling="2025-12-05 10:30:47.21666403 +0000 UTC m=+12947.745571291" observedRunningTime="2025-12-05 10:30:47.641054833 +0000 UTC m=+12948.169962094" watchObservedRunningTime="2025-12-05 10:30:47.642173903 +0000 UTC m=+12948.171081164" Dec 05 10:30:49 crc kubenswrapper[4997]: I1205 10:30:49.770400 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:30:49 crc kubenswrapper[4997]: I1205 10:30:49.770987 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:30:53 crc kubenswrapper[4997]: I1205 10:30:53.211189 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bxms4" Dec 05 10:30:53 crc kubenswrapper[4997]: I1205 10:30:53.211837 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bxms4" Dec 05 10:30:53 crc kubenswrapper[4997]: I1205 10:30:53.267681 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bxms4" Dec 05 10:30:53 crc kubenswrapper[4997]: I1205 10:30:53.770020 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bxms4" Dec 05 10:30:54 crc kubenswrapper[4997]: I1205 10:30:54.242215 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bxms4"] Dec 05 10:30:55 crc kubenswrapper[4997]: I1205 10:30:55.736166 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bxms4" podUID="01fae4ef-8762-4aac-8a32-db29b9f76c0d" containerName="registry-server" containerID="cri-o://a9b4d3922e0825904afa982033a30604f00d467e331be401f3d3b912fc3289d7" gracePeriod=2 Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.360351 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxms4" Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.522926 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01fae4ef-8762-4aac-8a32-db29b9f76c0d-catalog-content\") pod \"01fae4ef-8762-4aac-8a32-db29b9f76c0d\" (UID: \"01fae4ef-8762-4aac-8a32-db29b9f76c0d\") " Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.523018 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2lds\" (UniqueName: \"kubernetes.io/projected/01fae4ef-8762-4aac-8a32-db29b9f76c0d-kube-api-access-b2lds\") pod \"01fae4ef-8762-4aac-8a32-db29b9f76c0d\" (UID: \"01fae4ef-8762-4aac-8a32-db29b9f76c0d\") " Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.523108 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01fae4ef-8762-4aac-8a32-db29b9f76c0d-utilities\") pod \"01fae4ef-8762-4aac-8a32-db29b9f76c0d\" (UID: \"01fae4ef-8762-4aac-8a32-db29b9f76c0d\") " Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.524843 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01fae4ef-8762-4aac-8a32-db29b9f76c0d-utilities" (OuterVolumeSpecName: "utilities") pod "01fae4ef-8762-4aac-8a32-db29b9f76c0d" (UID: "01fae4ef-8762-4aac-8a32-db29b9f76c0d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.530499 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01fae4ef-8762-4aac-8a32-db29b9f76c0d-kube-api-access-b2lds" (OuterVolumeSpecName: "kube-api-access-b2lds") pod "01fae4ef-8762-4aac-8a32-db29b9f76c0d" (UID: "01fae4ef-8762-4aac-8a32-db29b9f76c0d"). InnerVolumeSpecName "kube-api-access-b2lds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.568551 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01fae4ef-8762-4aac-8a32-db29b9f76c0d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "01fae4ef-8762-4aac-8a32-db29b9f76c0d" (UID: "01fae4ef-8762-4aac-8a32-db29b9f76c0d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.626018 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01fae4ef-8762-4aac-8a32-db29b9f76c0d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.626266 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2lds\" (UniqueName: \"kubernetes.io/projected/01fae4ef-8762-4aac-8a32-db29b9f76c0d-kube-api-access-b2lds\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.626281 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01fae4ef-8762-4aac-8a32-db29b9f76c0d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.749271 4997 generic.go:334] "Generic (PLEG): container finished" podID="01fae4ef-8762-4aac-8a32-db29b9f76c0d" containerID="a9b4d3922e0825904afa982033a30604f00d467e331be401f3d3b912fc3289d7" exitCode=0 Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.749318 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxms4" Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.749326 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxms4" event={"ID":"01fae4ef-8762-4aac-8a32-db29b9f76c0d","Type":"ContainerDied","Data":"a9b4d3922e0825904afa982033a30604f00d467e331be401f3d3b912fc3289d7"} Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.749369 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxms4" event={"ID":"01fae4ef-8762-4aac-8a32-db29b9f76c0d","Type":"ContainerDied","Data":"70eb461cbaf859cfb0177bad72788270b2e78a3063402b92bafdca05d9f10bf5"} Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.749392 4997 scope.go:117] "RemoveContainer" containerID="a9b4d3922e0825904afa982033a30604f00d467e331be401f3d3b912fc3289d7" Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.787988 4997 scope.go:117] "RemoveContainer" containerID="73e1cfb909f64c6e9fe727944533f8cbf7e0281f6568a559b5d14c0f6fe1528d" Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.808852 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bxms4"] Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.813486 4997 scope.go:117] "RemoveContainer" containerID="dd7f3c25de06a420788ca688c08067d762040486fe600821caddd6828452bb4e" Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.822813 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bxms4"] Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.902431 4997 scope.go:117] "RemoveContainer" containerID="a9b4d3922e0825904afa982033a30604f00d467e331be401f3d3b912fc3289d7" Dec 05 10:30:56 crc kubenswrapper[4997]: E1205 10:30:56.902912 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9b4d3922e0825904afa982033a30604f00d467e331be401f3d3b912fc3289d7\": container with ID starting with a9b4d3922e0825904afa982033a30604f00d467e331be401f3d3b912fc3289d7 not found: ID does not exist" containerID="a9b4d3922e0825904afa982033a30604f00d467e331be401f3d3b912fc3289d7" Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.902979 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9b4d3922e0825904afa982033a30604f00d467e331be401f3d3b912fc3289d7"} err="failed to get container status \"a9b4d3922e0825904afa982033a30604f00d467e331be401f3d3b912fc3289d7\": rpc error: code = NotFound desc = could not find container \"a9b4d3922e0825904afa982033a30604f00d467e331be401f3d3b912fc3289d7\": container with ID starting with a9b4d3922e0825904afa982033a30604f00d467e331be401f3d3b912fc3289d7 not found: ID does not exist" Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.903021 4997 scope.go:117] "RemoveContainer" containerID="73e1cfb909f64c6e9fe727944533f8cbf7e0281f6568a559b5d14c0f6fe1528d" Dec 05 10:30:56 crc kubenswrapper[4997]: E1205 10:30:56.903382 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73e1cfb909f64c6e9fe727944533f8cbf7e0281f6568a559b5d14c0f6fe1528d\": container with ID starting with 73e1cfb909f64c6e9fe727944533f8cbf7e0281f6568a559b5d14c0f6fe1528d not found: ID does not exist" containerID="73e1cfb909f64c6e9fe727944533f8cbf7e0281f6568a559b5d14c0f6fe1528d" Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.903444 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73e1cfb909f64c6e9fe727944533f8cbf7e0281f6568a559b5d14c0f6fe1528d"} err="failed to get container status \"73e1cfb909f64c6e9fe727944533f8cbf7e0281f6568a559b5d14c0f6fe1528d\": rpc error: code = NotFound desc = could not find container \"73e1cfb909f64c6e9fe727944533f8cbf7e0281f6568a559b5d14c0f6fe1528d\": container with ID starting with 73e1cfb909f64c6e9fe727944533f8cbf7e0281f6568a559b5d14c0f6fe1528d not found: ID does not exist" Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.903480 4997 scope.go:117] "RemoveContainer" containerID="dd7f3c25de06a420788ca688c08067d762040486fe600821caddd6828452bb4e" Dec 05 10:30:56 crc kubenswrapper[4997]: E1205 10:30:56.903995 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd7f3c25de06a420788ca688c08067d762040486fe600821caddd6828452bb4e\": container with ID starting with dd7f3c25de06a420788ca688c08067d762040486fe600821caddd6828452bb4e not found: ID does not exist" containerID="dd7f3c25de06a420788ca688c08067d762040486fe600821caddd6828452bb4e" Dec 05 10:30:56 crc kubenswrapper[4997]: I1205 10:30:56.904041 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd7f3c25de06a420788ca688c08067d762040486fe600821caddd6828452bb4e"} err="failed to get container status \"dd7f3c25de06a420788ca688c08067d762040486fe600821caddd6828452bb4e\": rpc error: code = NotFound desc = could not find container \"dd7f3c25de06a420788ca688c08067d762040486fe600821caddd6828452bb4e\": container with ID starting with dd7f3c25de06a420788ca688c08067d762040486fe600821caddd6828452bb4e not found: ID does not exist" Dec 05 10:30:57 crc kubenswrapper[4997]: I1205 10:30:57.766115 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01fae4ef-8762-4aac-8a32-db29b9f76c0d" path="/var/lib/kubelet/pods/01fae4ef-8762-4aac-8a32-db29b9f76c0d/volumes" Dec 05 10:31:19 crc kubenswrapper[4997]: I1205 10:31:19.770521 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:31:19 crc kubenswrapper[4997]: I1205 10:31:19.771065 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:31:19 crc kubenswrapper[4997]: I1205 10:31:19.771110 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 10:31:19 crc kubenswrapper[4997]: I1205 10:31:19.771882 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"570182e001027ed044c074b2d6343f7b4f00f9a0f33006cc726adf1d3379a622"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 10:31:19 crc kubenswrapper[4997]: I1205 10:31:19.771932 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://570182e001027ed044c074b2d6343f7b4f00f9a0f33006cc726adf1d3379a622" gracePeriod=600 Dec 05 10:31:20 crc kubenswrapper[4997]: I1205 10:31:20.012201 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="570182e001027ed044c074b2d6343f7b4f00f9a0f33006cc726adf1d3379a622" exitCode=0 Dec 05 10:31:20 crc kubenswrapper[4997]: I1205 10:31:20.012277 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"570182e001027ed044c074b2d6343f7b4f00f9a0f33006cc726adf1d3379a622"} Dec 05 10:31:20 crc kubenswrapper[4997]: I1205 10:31:20.012508 4997 scope.go:117] "RemoveContainer" containerID="fe88d2a245b445a9666ff72081d7ab498c7df01a3175625b1e25ad0792bbe6bc" Dec 05 10:31:21 crc kubenswrapper[4997]: I1205 10:31:21.025160 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648"} Dec 05 10:31:26 crc kubenswrapper[4997]: I1205 10:31:26.896742 4997 patch_prober.go:28] interesting pod/nmstate-webhook-5f6d4c5ccb-4cxth container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.44:9443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 10:31:26 crc kubenswrapper[4997]: I1205 10:31:26.897808 4997 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4cxth" podUID="9dc3f953-29af-435f-ba80-fc7d2f3b0dc6" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.44:9443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:33:38 crc kubenswrapper[4997]: I1205 10:33:38.905505 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_7aea731b-86e5-4305-954d-1de96d638b44/init-config-reloader/0.log" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.108023 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_7aea731b-86e5-4305-954d-1de96d638b44/init-config-reloader/0.log" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.118311 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_7aea731b-86e5-4305-954d-1de96d638b44/config-reloader/0.log" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.147390 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_7aea731b-86e5-4305-954d-1de96d638b44/alertmanager/0.log" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.185433 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-56khv"] Dec 05 10:33:39 crc kubenswrapper[4997]: E1205 10:33:39.185893 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01fae4ef-8762-4aac-8a32-db29b9f76c0d" containerName="extract-content" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.185917 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="01fae4ef-8762-4aac-8a32-db29b9f76c0d" containerName="extract-content" Dec 05 10:33:39 crc kubenswrapper[4997]: E1205 10:33:39.185933 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01fae4ef-8762-4aac-8a32-db29b9f76c0d" containerName="extract-utilities" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.185940 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="01fae4ef-8762-4aac-8a32-db29b9f76c0d" containerName="extract-utilities" Dec 05 10:33:39 crc kubenswrapper[4997]: E1205 10:33:39.185950 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01fae4ef-8762-4aac-8a32-db29b9f76c0d" containerName="registry-server" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.185956 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="01fae4ef-8762-4aac-8a32-db29b9f76c0d" containerName="registry-server" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.186159 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="01fae4ef-8762-4aac-8a32-db29b9f76c0d" containerName="registry-server" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.187586 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56khv" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.207035 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-56khv"] Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.298569 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e341343e-c6c2-43b0-a712-77bc2303f4e0-utilities\") pod \"redhat-marketplace-56khv\" (UID: \"e341343e-c6c2-43b0-a712-77bc2303f4e0\") " pod="openshift-marketplace/redhat-marketplace-56khv" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.298655 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e341343e-c6c2-43b0-a712-77bc2303f4e0-catalog-content\") pod \"redhat-marketplace-56khv\" (UID: \"e341343e-c6c2-43b0-a712-77bc2303f4e0\") " pod="openshift-marketplace/redhat-marketplace-56khv" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.298706 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n8dd\" (UniqueName: \"kubernetes.io/projected/e341343e-c6c2-43b0-a712-77bc2303f4e0-kube-api-access-6n8dd\") pod \"redhat-marketplace-56khv\" (UID: \"e341343e-c6c2-43b0-a712-77bc2303f4e0\") " pod="openshift-marketplace/redhat-marketplace-56khv" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.393872 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_43128621-bd06-4369-8b56-eda49b797194/aodh-api/0.log" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.400904 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e341343e-c6c2-43b0-a712-77bc2303f4e0-utilities\") pod \"redhat-marketplace-56khv\" (UID: \"e341343e-c6c2-43b0-a712-77bc2303f4e0\") " pod="openshift-marketplace/redhat-marketplace-56khv" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.400966 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e341343e-c6c2-43b0-a712-77bc2303f4e0-catalog-content\") pod \"redhat-marketplace-56khv\" (UID: \"e341343e-c6c2-43b0-a712-77bc2303f4e0\") " pod="openshift-marketplace/redhat-marketplace-56khv" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.401006 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n8dd\" (UniqueName: \"kubernetes.io/projected/e341343e-c6c2-43b0-a712-77bc2303f4e0-kube-api-access-6n8dd\") pod \"redhat-marketplace-56khv\" (UID: \"e341343e-c6c2-43b0-a712-77bc2303f4e0\") " pod="openshift-marketplace/redhat-marketplace-56khv" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.401905 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e341343e-c6c2-43b0-a712-77bc2303f4e0-catalog-content\") pod \"redhat-marketplace-56khv\" (UID: \"e341343e-c6c2-43b0-a712-77bc2303f4e0\") " pod="openshift-marketplace/redhat-marketplace-56khv" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.402942 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e341343e-c6c2-43b0-a712-77bc2303f4e0-utilities\") pod \"redhat-marketplace-56khv\" (UID: \"e341343e-c6c2-43b0-a712-77bc2303f4e0\") " pod="openshift-marketplace/redhat-marketplace-56khv" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.467362 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n8dd\" (UniqueName: \"kubernetes.io/projected/e341343e-c6c2-43b0-a712-77bc2303f4e0-kube-api-access-6n8dd\") pod \"redhat-marketplace-56khv\" (UID: \"e341343e-c6c2-43b0-a712-77bc2303f4e0\") " pod="openshift-marketplace/redhat-marketplace-56khv" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.499667 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_43128621-bd06-4369-8b56-eda49b797194/aodh-notifier/0.log" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.517532 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56khv" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.556482 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_43128621-bd06-4369-8b56-eda49b797194/aodh-evaluator/0.log" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.768134 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_43128621-bd06-4369-8b56-eda49b797194/aodh-listener/0.log" Dec 05 10:33:39 crc kubenswrapper[4997]: I1205 10:33:39.824130 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9f58cdf4d-fz24c_08bdbc96-88c5-4744-bc48-bc2d232a8d6a/barbican-api/0.log" Dec 05 10:33:40 crc kubenswrapper[4997]: I1205 10:33:40.019706 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9f58cdf4d-fz24c_08bdbc96-88c5-4744-bc48-bc2d232a8d6a/barbican-api-log/0.log" Dec 05 10:33:40 crc kubenswrapper[4997]: I1205 10:33:40.082823 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-56khv"] Dec 05 10:33:40 crc kubenswrapper[4997]: I1205 10:33:40.121882 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-77b797d5d4-vh4h4_6c326c07-9fee-4b86-b7c1-bd9ce32c95f3/barbican-keystone-listener/0.log" Dec 05 10:33:40 crc kubenswrapper[4997]: I1205 10:33:40.261338 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56khv" event={"ID":"e341343e-c6c2-43b0-a712-77bc2303f4e0","Type":"ContainerStarted","Data":"8715ce59f0a6af0405996b59de0dd38548636b5d9d177e9a0407b2b3347f7e08"} Dec 05 10:33:40 crc kubenswrapper[4997]: I1205 10:33:40.357179 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5756b5f4c9-wj8k4_0666a02f-d3ed-4403-b2af-5450e2901d82/barbican-worker/0.log" Dec 05 10:33:40 crc kubenswrapper[4997]: I1205 10:33:40.425207 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5756b5f4c9-wj8k4_0666a02f-d3ed-4403-b2af-5450e2901d82/barbican-worker-log/0.log" Dec 05 10:33:40 crc kubenswrapper[4997]: I1205 10:33:40.625316 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-97rb6_e45c1dd1-eab5-4852-bb9d-ecee8ab68e00/bootstrap-openstack-openstack-cell1/0.log" Dec 05 10:33:40 crc kubenswrapper[4997]: I1205 10:33:40.951539 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2a2df928-f920-4e13-87ae-a412539dc001/ceilometer-central-agent/0.log" Dec 05 10:33:40 crc kubenswrapper[4997]: I1205 10:33:40.973906 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-networker-r2z6x_95a0bda3-99d3-46df-8ca6-fd85db893d66/bootstrap-openstack-openstack-networker/0.log" Dec 05 10:33:41 crc kubenswrapper[4997]: I1205 10:33:41.000136 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-77b797d5d4-vh4h4_6c326c07-9fee-4b86-b7c1-bd9ce32c95f3/barbican-keystone-listener-log/0.log" Dec 05 10:33:41 crc kubenswrapper[4997]: I1205 10:33:41.160877 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2a2df928-f920-4e13-87ae-a412539dc001/ceilometer-notification-agent/0.log" Dec 05 10:33:41 crc kubenswrapper[4997]: I1205 10:33:41.175729 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2a2df928-f920-4e13-87ae-a412539dc001/proxy-httpd/0.log" Dec 05 10:33:41 crc kubenswrapper[4997]: I1205 10:33:41.214755 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2a2df928-f920-4e13-87ae-a412539dc001/sg-core/0.log" Dec 05 10:33:41 crc kubenswrapper[4997]: I1205 10:33:41.272228 4997 generic.go:334] "Generic (PLEG): container finished" podID="e341343e-c6c2-43b0-a712-77bc2303f4e0" containerID="0a0f187f033b476a338f65809dda43a29acd266dde50fdeaf1e08c0752232edd" exitCode=0 Dec 05 10:33:41 crc kubenswrapper[4997]: I1205 10:33:41.272273 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56khv" event={"ID":"e341343e-c6c2-43b0-a712-77bc2303f4e0","Type":"ContainerDied","Data":"0a0f187f033b476a338f65809dda43a29acd266dde50fdeaf1e08c0752232edd"} Dec 05 10:33:41 crc kubenswrapper[4997]: I1205 10:33:41.274646 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:33:41 crc kubenswrapper[4997]: I1205 10:33:41.385948 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-cfqnz_f8efbd91-b09a-4dac-bf7d-b0b0433023ff/ceph-client-openstack-openstack-cell1/0.log" Dec 05 10:33:41 crc kubenswrapper[4997]: I1205 10:33:41.774094 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9de7a6bf-87ed-44b2-9a59-e91ef6f53740/cinder-api-log/0.log" Dec 05 10:33:41 crc kubenswrapper[4997]: I1205 10:33:41.808362 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9de7a6bf-87ed-44b2-9a59-e91ef6f53740/cinder-api/0.log" Dec 05 10:33:41 crc kubenswrapper[4997]: I1205 10:33:41.977528 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_7990da72-c93f-40dd-ae1c-2e5af98502ad/probe/0.log" Dec 05 10:33:42 crc kubenswrapper[4997]: I1205 10:33:42.115601 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_958ea1bf-c540-4450-a408-c225c3ba1dd3/cinder-scheduler/0.log" Dec 05 10:33:42 crc kubenswrapper[4997]: I1205 10:33:42.290025 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56khv" event={"ID":"e341343e-c6c2-43b0-a712-77bc2303f4e0","Type":"ContainerStarted","Data":"a3450a3b2a58b9453b1532b11d497b8b89946f64f8e8b5f102fe6a3173900f57"} Dec 05 10:33:42 crc kubenswrapper[4997]: I1205 10:33:42.363703 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_958ea1bf-c540-4450-a408-c225c3ba1dd3/probe/0.log" Dec 05 10:33:42 crc kubenswrapper[4997]: I1205 10:33:42.848266 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_eba14735-b01b-4358-918f-f8f2735f2925/probe/0.log" Dec 05 10:33:43 crc kubenswrapper[4997]: I1205 10:33:43.022564 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-rlp6r_efdefcdd-f76f-4767-af25-b730587e394b/configure-network-openstack-openstack-cell1/0.log" Dec 05 10:33:43 crc kubenswrapper[4997]: I1205 10:33:43.194106 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-networker-6pc2p_10cd3056-9662-42dd-b5c4-9b4aeed3da3a/configure-network-openstack-openstack-networker/0.log" Dec 05 10:33:43 crc kubenswrapper[4997]: I1205 10:33:43.321507 4997 generic.go:334] "Generic (PLEG): container finished" podID="e341343e-c6c2-43b0-a712-77bc2303f4e0" containerID="a3450a3b2a58b9453b1532b11d497b8b89946f64f8e8b5f102fe6a3173900f57" exitCode=0 Dec 05 10:33:43 crc kubenswrapper[4997]: I1205 10:33:43.321586 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56khv" event={"ID":"e341343e-c6c2-43b0-a712-77bc2303f4e0","Type":"ContainerDied","Data":"a3450a3b2a58b9453b1532b11d497b8b89946f64f8e8b5f102fe6a3173900f57"} Dec 05 10:33:43 crc kubenswrapper[4997]: I1205 10:33:43.405352 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-jsbwt_cfe38bed-b2b8-4d90-b74b-d82cea4b63a4/configure-os-openstack-openstack-cell1/0.log" Dec 05 10:33:43 crc kubenswrapper[4997]: I1205 10:33:43.785323 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-networker-29djd_888fe326-b46b-45f0-9dfe-6875d4ef6def/configure-os-openstack-openstack-networker/0.log" Dec 05 10:33:43 crc kubenswrapper[4997]: I1205 10:33:43.964854 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78dc8dd569-nzqx8_3b87119b-5ad1-435b-9289-5a5a71765f8a/init/0.log" Dec 05 10:33:44 crc kubenswrapper[4997]: I1205 10:33:44.253505 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78dc8dd569-nzqx8_3b87119b-5ad1-435b-9289-5a5a71765f8a/init/0.log" Dec 05 10:33:44 crc kubenswrapper[4997]: I1205 10:33:44.344430 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56khv" event={"ID":"e341343e-c6c2-43b0-a712-77bc2303f4e0","Type":"ContainerStarted","Data":"3956e95cbc4f849d7000f93633202791a96bb0e65c92c0341a5cb735cb015568"} Dec 05 10:33:44 crc kubenswrapper[4997]: I1205 10:33:44.379845 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-56khv" podStartSLOduration=2.96338022 podStartE2EDuration="5.379820908s" podCreationTimestamp="2025-12-05 10:33:39 +0000 UTC" firstStartedPulling="2025-12-05 10:33:41.274414134 +0000 UTC m=+13121.803321405" lastFinishedPulling="2025-12-05 10:33:43.690854832 +0000 UTC m=+13124.219762093" observedRunningTime="2025-12-05 10:33:44.365489981 +0000 UTC m=+13124.894397252" watchObservedRunningTime="2025-12-05 10:33:44.379820908 +0000 UTC m=+13124.908728169" Dec 05 10:33:44 crc kubenswrapper[4997]: I1205 10:33:44.445244 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78dc8dd569-nzqx8_3b87119b-5ad1-435b-9289-5a5a71765f8a/dnsmasq-dns/0.log" Dec 05 10:33:44 crc kubenswrapper[4997]: I1205 10:33:44.539442 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-b22xv_999c70de-3370-4ce0-947a-b585dcd879d7/download-cache-openstack-openstack-cell1/0.log" Dec 05 10:33:44 crc kubenswrapper[4997]: I1205 10:33:44.645775 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_7990da72-c93f-40dd-ae1c-2e5af98502ad/cinder-backup/0.log" Dec 05 10:33:44 crc kubenswrapper[4997]: I1205 10:33:44.712093 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-networker-nsk64_73bb544f-2408-4f58-94eb-7b5116d41848/download-cache-openstack-openstack-networker/0.log" Dec 05 10:33:44 crc kubenswrapper[4997]: I1205 10:33:44.931439 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1c47845d-3866-47ce-9242-5b315f1a565c/glance-httpd/0.log" Dec 05 10:33:44 crc kubenswrapper[4997]: I1205 10:33:44.952797 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1c47845d-3866-47ce-9242-5b315f1a565c/glance-log/0.log" Dec 05 10:33:45 crc kubenswrapper[4997]: I1205 10:33:45.104801 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_26891501-25d1-4d26-947e-b23b45116b35/glance-httpd/0.log" Dec 05 10:33:45 crc kubenswrapper[4997]: I1205 10:33:45.188725 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_26891501-25d1-4d26-947e-b23b45116b35/glance-log/0.log" Dec 05 10:33:45 crc kubenswrapper[4997]: I1205 10:33:45.484506 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_eba14735-b01b-4358-918f-f8f2735f2925/cinder-volume/0.log" Dec 05 10:33:45 crc kubenswrapper[4997]: I1205 10:33:45.506236 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-65bb7ccfdf-b9g84_cf60578b-47bb-474a-b296-56c42105680d/heat-api/0.log" Dec 05 10:33:45 crc kubenswrapper[4997]: I1205 10:33:45.592580 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-5774948dd5-879rv_5f69f8d5-e82e-4db6-aad0-e302e807b2b0/heat-cfnapi/0.log" Dec 05 10:33:45 crc kubenswrapper[4997]: I1205 10:33:45.647808 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-54f49697fd-pwnb8_fcb82741-d0c0-42af-ad5b-ac6a840393b8/heat-engine/0.log" Dec 05 10:33:45 crc kubenswrapper[4997]: I1205 10:33:45.771331 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-678ddff49c-bvhj8_2da24366-b581-40f0-94bf-2a9652e2840a/horizon/0.log" Dec 05 10:33:45 crc kubenswrapper[4997]: I1205 10:33:45.918954 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-2kzxz_b47da422-d6ad-4a06-8f9d-f6e0d6c80d57/install-certs-openstack-openstack-cell1/0.log" Dec 05 10:33:45 crc kubenswrapper[4997]: I1205 10:33:45.937054 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-678ddff49c-bvhj8_2da24366-b581-40f0-94bf-2a9652e2840a/horizon-log/0.log" Dec 05 10:33:46 crc kubenswrapper[4997]: I1205 10:33:46.008754 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-networker-zzx85_5e5fc186-cf3d-444f-9f4d-3dc494faaa7e/install-certs-openstack-openstack-networker/0.log" Dec 05 10:33:46 crc kubenswrapper[4997]: I1205 10:33:46.144702 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-bxss2_a9704a5b-2b0a-4786-b151-29cddb30973f/install-os-openstack-openstack-cell1/0.log" Dec 05 10:33:46 crc kubenswrapper[4997]: I1205 10:33:46.262688 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-networker-xmfpv_d81a1b59-323c-4254-8e95-b50acb2eb309/install-os-openstack-openstack-networker/0.log" Dec 05 10:33:46 crc kubenswrapper[4997]: I1205 10:33:46.359897 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29415421-f7jhc_6787bbc7-67b4-4f03-9302-470b4f539214/keystone-cron/0.log" Dec 05 10:33:46 crc kubenswrapper[4997]: I1205 10:33:46.534663 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29415481-fnjnr_08b4289a-1b86-4057-860a-4f5b4bd157b1/keystone-cron/0.log" Dec 05 10:33:46 crc kubenswrapper[4997]: I1205 10:33:46.646192 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_53022e5c-c432-461a-bb5a-31844df7c3e9/kube-state-metrics/0.log" Dec 05 10:33:46 crc kubenswrapper[4997]: I1205 10:33:46.853288 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-5q88v_93e1015b-7d40-4ad7-a1b8-78f9a243223e/libvirt-openstack-openstack-cell1/0.log" Dec 05 10:33:47 crc kubenswrapper[4997]: I1205 10:33:47.288911 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_ed63e934-a1ac-479c-96cc-0100c0869765/manila-scheduler/0.log" Dec 05 10:33:47 crc kubenswrapper[4997]: I1205 10:33:47.394188 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-56858ddfd7-wzhwt_56e0b4d4-eadc-456f-b9b1-30db806b94cc/keystone-api/0.log" Dec 05 10:33:47 crc kubenswrapper[4997]: I1205 10:33:47.406332 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_bcc15490-e838-4503-8318-f79aafc8c961/manila-api/0.log" Dec 05 10:33:47 crc kubenswrapper[4997]: I1205 10:33:47.443797 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_ed63e934-a1ac-479c-96cc-0100c0869765/probe/0.log" Dec 05 10:33:47 crc kubenswrapper[4997]: I1205 10:33:47.533867 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_bcc15490-e838-4503-8318-f79aafc8c961/manila-api-log/0.log" Dec 05 10:33:47 crc kubenswrapper[4997]: I1205 10:33:47.675227 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_799f7992-40e0-4fea-8e26-84a596b0500f/probe/0.log" Dec 05 10:33:47 crc kubenswrapper[4997]: I1205 10:33:47.689790 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_799f7992-40e0-4fea-8e26-84a596b0500f/manila-share/0.log" Dec 05 10:33:48 crc kubenswrapper[4997]: I1205 10:33:48.363889 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7f58c8c5b5-bd9p4_45add928-5cda-478a-9f33-ea2c2a09f99e/neutron-httpd/0.log" Dec 05 10:33:48 crc kubenswrapper[4997]: I1205 10:33:48.703638 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-qbcxc_f6398c5c-8fdf-49dc-8de8-c0bc70637ac8/neutron-dhcp-openstack-openstack-cell1/0.log" Dec 05 10:33:48 crc kubenswrapper[4997]: I1205 10:33:48.830551 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-v7mv9_4063a721-9efb-4e0b-a7c6-93c9bdaf3ff4/neutron-metadata-openstack-openstack-cell1/0.log" Dec 05 10:33:48 crc kubenswrapper[4997]: I1205 10:33:48.912931 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7f58c8c5b5-bd9p4_45add928-5cda-478a-9f33-ea2c2a09f99e/neutron-api/0.log" Dec 05 10:33:49 crc kubenswrapper[4997]: I1205 10:33:49.083043 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-networker-n7m22_ba5f43cb-d3c6-49e8-9f83-551cd860f736/neutron-metadata-openstack-openstack-networker/0.log" Dec 05 10:33:49 crc kubenswrapper[4997]: I1205 10:33:49.308756 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-qmm94_8f306f28-44f2-4a8c-95e2-9b6d1c7a5337/neutron-sriov-openstack-openstack-cell1/0.log" Dec 05 10:33:49 crc kubenswrapper[4997]: I1205 10:33:49.517906 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-56khv" Dec 05 10:33:49 crc kubenswrapper[4997]: I1205 10:33:49.521974 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-56khv" Dec 05 10:33:49 crc kubenswrapper[4997]: I1205 10:33:49.544000 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_5db61aad-f627-47e0-9e6c-4336e1a723b3/nova-api-api/0.log" Dec 05 10:33:49 crc kubenswrapper[4997]: I1205 10:33:49.576184 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-56khv" Dec 05 10:33:49 crc kubenswrapper[4997]: I1205 10:33:49.708484 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_b3180318-e875-4261-bc09-b10c13e79ca5/nova-cell0-conductor-conductor/0.log" Dec 05 10:33:49 crc kubenswrapper[4997]: I1205 10:33:49.769551 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:33:49 crc kubenswrapper[4997]: I1205 10:33:49.769700 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:33:49 crc kubenswrapper[4997]: I1205 10:33:49.865933 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_5db61aad-f627-47e0-9e6c-4336e1a723b3/nova-api-log/0.log" Dec 05 10:33:49 crc kubenswrapper[4997]: I1205 10:33:49.918804 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0fbc803d-3d9a-425a-9ac3-65e7010d30a9/nova-cell1-conductor-conductor/0.log" Dec 05 10:33:50 crc kubenswrapper[4997]: I1205 10:33:50.168332 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_4d30eaf1-eb47-46c2-b87a-18fe74b003e9/nova-cell1-novncproxy-novncproxy/0.log" Dec 05 10:33:50 crc kubenswrapper[4997]: I1205 10:33:50.409124 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellfk4zh_329e73d1-df45-481a-b897-1bde9107da2b/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Dec 05 10:33:50 crc kubenswrapper[4997]: I1205 10:33:50.463144 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-lltrv_c684f910-aebf-42e8-940b-5abd6b46c2a2/nova-cell1-openstack-openstack-cell1/0.log" Dec 05 10:33:50 crc kubenswrapper[4997]: I1205 10:33:50.512525 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-56khv" Dec 05 10:33:50 crc kubenswrapper[4997]: I1205 10:33:50.600200 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-56khv"] Dec 05 10:33:50 crc kubenswrapper[4997]: I1205 10:33:50.697302 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_481e9fdd-91aa-4306-bd21-27fc9bbf9608/nova-metadata-log/0.log" Dec 05 10:33:50 crc kubenswrapper[4997]: I1205 10:33:50.858973 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_481e9fdd-91aa-4306-bd21-27fc9bbf9608/nova-metadata-metadata/0.log" Dec 05 10:33:51 crc kubenswrapper[4997]: I1205 10:33:51.002238 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_396b6f95-2e98-4e5d-9a32-a0b6758b09db/nova-scheduler-scheduler/0.log" Dec 05 10:33:51 crc kubenswrapper[4997]: I1205 10:33:51.008121 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_351c29ac-7ec2-4c65-95a1-040707a8d01f/mysql-bootstrap/0.log" Dec 05 10:33:51 crc kubenswrapper[4997]: I1205 10:33:51.216514 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_351c29ac-7ec2-4c65-95a1-040707a8d01f/mysql-bootstrap/0.log" Dec 05 10:33:51 crc kubenswrapper[4997]: I1205 10:33:51.227998 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_351c29ac-7ec2-4c65-95a1-040707a8d01f/galera/0.log" Dec 05 10:33:51 crc kubenswrapper[4997]: I1205 10:33:51.257086 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_108e77e9-9f51-4686-bac0-1e0397f33c28/mysql-bootstrap/0.log" Dec 05 10:33:51 crc kubenswrapper[4997]: I1205 10:33:51.550821 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_174c1aed-3954-4867-ad3e-0a5d51579490/openstackclient/0.log" Dec 05 10:33:51 crc kubenswrapper[4997]: I1205 10:33:51.555358 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_108e77e9-9f51-4686-bac0-1e0397f33c28/mysql-bootstrap/0.log" Dec 05 10:33:51 crc kubenswrapper[4997]: I1205 10:33:51.617591 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_108e77e9-9f51-4686-bac0-1e0397f33c28/galera/0.log" Dec 05 10:33:51 crc kubenswrapper[4997]: I1205 10:33:51.745084 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a10d62a7-cf49-42bd-8d9b-72cba21becd8/openstack-network-exporter/0.log" Dec 05 10:33:51 crc kubenswrapper[4997]: I1205 10:33:51.784119 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a10d62a7-cf49-42bd-8d9b-72cba21becd8/ovn-northd/0.log" Dec 05 10:33:52 crc kubenswrapper[4997]: I1205 10:33:52.142469 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-9dg4x_d9179dc6-dc8e-493d-8120-27d7a3e723e6/ovn-openstack-openstack-cell1/0.log" Dec 05 10:33:52 crc kubenswrapper[4997]: I1205 10:33:52.311230 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bf6dd0ec-b448-408c-8b0f-7165be31147f/openstack-network-exporter/0.log" Dec 05 10:33:52 crc kubenswrapper[4997]: I1205 10:33:52.398312 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bf6dd0ec-b448-408c-8b0f-7165be31147f/ovsdbserver-nb/0.log" Dec 05 10:33:52 crc kubenswrapper[4997]: I1205 10:33:52.411846 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-networker-9mbvl_51661db5-879f-4b6c-a563-ff1c84e47f0d/ovn-openstack-openstack-networker/0.log" Dec 05 10:33:52 crc kubenswrapper[4997]: I1205 10:33:52.469521 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-56khv" podUID="e341343e-c6c2-43b0-a712-77bc2303f4e0" containerName="registry-server" containerID="cri-o://3956e95cbc4f849d7000f93633202791a96bb0e65c92c0341a5cb735cb015568" gracePeriod=2 Dec 05 10:33:52 crc kubenswrapper[4997]: I1205 10:33:52.567338 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_43590576-9b6a-436e-9022-d9af8d390cda/openstack-network-exporter/0.log" Dec 05 10:33:52 crc kubenswrapper[4997]: I1205 10:33:52.629951 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_43590576-9b6a-436e-9022-d9af8d390cda/ovsdbserver-nb/0.log" Dec 05 10:33:52 crc kubenswrapper[4997]: I1205 10:33:52.788698 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_9c57e847-90d0-4769-9eb8-8c90080aee26/openstack-network-exporter/0.log" Dec 05 10:33:52 crc kubenswrapper[4997]: I1205 10:33:52.877278 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_9c57e847-90d0-4769-9eb8-8c90080aee26/ovsdbserver-nb/0.log" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.003222 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7bd459cc-346a-4432-ae3d-4df900035da3/openstack-network-exporter/0.log" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.062182 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56khv" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.089697 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7bd459cc-346a-4432-ae3d-4df900035da3/ovsdbserver-sb/0.log" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.143428 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e341343e-c6c2-43b0-a712-77bc2303f4e0-catalog-content\") pod \"e341343e-c6c2-43b0-a712-77bc2303f4e0\" (UID: \"e341343e-c6c2-43b0-a712-77bc2303f4e0\") " Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.143586 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n8dd\" (UniqueName: \"kubernetes.io/projected/e341343e-c6c2-43b0-a712-77bc2303f4e0-kube-api-access-6n8dd\") pod \"e341343e-c6c2-43b0-a712-77bc2303f4e0\" (UID: \"e341343e-c6c2-43b0-a712-77bc2303f4e0\") " Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.143824 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e341343e-c6c2-43b0-a712-77bc2303f4e0-utilities\") pod \"e341343e-c6c2-43b0-a712-77bc2303f4e0\" (UID: \"e341343e-c6c2-43b0-a712-77bc2303f4e0\") " Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.145146 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e341343e-c6c2-43b0-a712-77bc2303f4e0-utilities" (OuterVolumeSpecName: "utilities") pod "e341343e-c6c2-43b0-a712-77bc2303f4e0" (UID: "e341343e-c6c2-43b0-a712-77bc2303f4e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.190530 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e341343e-c6c2-43b0-a712-77bc2303f4e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e341343e-c6c2-43b0-a712-77bc2303f4e0" (UID: "e341343e-c6c2-43b0-a712-77bc2303f4e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.202072 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e341343e-c6c2-43b0-a712-77bc2303f4e0-kube-api-access-6n8dd" (OuterVolumeSpecName: "kube-api-access-6n8dd") pod "e341343e-c6c2-43b0-a712-77bc2303f4e0" (UID: "e341343e-c6c2-43b0-a712-77bc2303f4e0"). InnerVolumeSpecName "kube-api-access-6n8dd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.246527 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e341343e-c6c2-43b0-a712-77bc2303f4e0-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.246558 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e341343e-c6c2-43b0-a712-77bc2303f4e0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.246569 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n8dd\" (UniqueName: \"kubernetes.io/projected/e341343e-c6c2-43b0-a712-77bc2303f4e0-kube-api-access-6n8dd\") on node \"crc\" DevicePath \"\"" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.445253 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_def832ab-6d2d-474e-ae28-f1e8e868f07e/openstack-network-exporter/0.log" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.481211 4997 generic.go:334] "Generic (PLEG): container finished" podID="e341343e-c6c2-43b0-a712-77bc2303f4e0" containerID="3956e95cbc4f849d7000f93633202791a96bb0e65c92c0341a5cb735cb015568" exitCode=0 Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.481333 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56khv" event={"ID":"e341343e-c6c2-43b0-a712-77bc2303f4e0","Type":"ContainerDied","Data":"3956e95cbc4f849d7000f93633202791a96bb0e65c92c0341a5cb735cb015568"} Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.481367 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56khv" event={"ID":"e341343e-c6c2-43b0-a712-77bc2303f4e0","Type":"ContainerDied","Data":"8715ce59f0a6af0405996b59de0dd38548636b5d9d177e9a0407b2b3347f7e08"} Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.481390 4997 scope.go:117] "RemoveContainer" containerID="3956e95cbc4f849d7000f93633202791a96bb0e65c92c0341a5cb735cb015568" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.481590 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56khv" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.524219 4997 scope.go:117] "RemoveContainer" containerID="a3450a3b2a58b9453b1532b11d497b8b89946f64f8e8b5f102fe6a3173900f57" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.532583 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-56khv"] Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.537087 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_def832ab-6d2d-474e-ae28-f1e8e868f07e/ovsdbserver-sb/0.log" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.555184 4997 scope.go:117] "RemoveContainer" containerID="0a0f187f033b476a338f65809dda43a29acd266dde50fdeaf1e08c0752232edd" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.602635 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-56khv"] Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.619464 4997 scope.go:117] "RemoveContainer" containerID="3956e95cbc4f849d7000f93633202791a96bb0e65c92c0341a5cb735cb015568" Dec 05 10:33:53 crc kubenswrapper[4997]: E1205 10:33:53.619886 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3956e95cbc4f849d7000f93633202791a96bb0e65c92c0341a5cb735cb015568\": container with ID starting with 3956e95cbc4f849d7000f93633202791a96bb0e65c92c0341a5cb735cb015568 not found: ID does not exist" containerID="3956e95cbc4f849d7000f93633202791a96bb0e65c92c0341a5cb735cb015568" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.619910 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3956e95cbc4f849d7000f93633202791a96bb0e65c92c0341a5cb735cb015568"} err="failed to get container status \"3956e95cbc4f849d7000f93633202791a96bb0e65c92c0341a5cb735cb015568\": rpc error: code = NotFound desc = could not find container \"3956e95cbc4f849d7000f93633202791a96bb0e65c92c0341a5cb735cb015568\": container with ID starting with 3956e95cbc4f849d7000f93633202791a96bb0e65c92c0341a5cb735cb015568 not found: ID does not exist" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.619928 4997 scope.go:117] "RemoveContainer" containerID="a3450a3b2a58b9453b1532b11d497b8b89946f64f8e8b5f102fe6a3173900f57" Dec 05 10:33:53 crc kubenswrapper[4997]: E1205 10:33:53.620119 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3450a3b2a58b9453b1532b11d497b8b89946f64f8e8b5f102fe6a3173900f57\": container with ID starting with a3450a3b2a58b9453b1532b11d497b8b89946f64f8e8b5f102fe6a3173900f57 not found: ID does not exist" containerID="a3450a3b2a58b9453b1532b11d497b8b89946f64f8e8b5f102fe6a3173900f57" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.620137 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3450a3b2a58b9453b1532b11d497b8b89946f64f8e8b5f102fe6a3173900f57"} err="failed to get container status \"a3450a3b2a58b9453b1532b11d497b8b89946f64f8e8b5f102fe6a3173900f57\": rpc error: code = NotFound desc = could not find container \"a3450a3b2a58b9453b1532b11d497b8b89946f64f8e8b5f102fe6a3173900f57\": container with ID starting with a3450a3b2a58b9453b1532b11d497b8b89946f64f8e8b5f102fe6a3173900f57 not found: ID does not exist" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.620150 4997 scope.go:117] "RemoveContainer" containerID="0a0f187f033b476a338f65809dda43a29acd266dde50fdeaf1e08c0752232edd" Dec 05 10:33:53 crc kubenswrapper[4997]: E1205 10:33:53.620311 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a0f187f033b476a338f65809dda43a29acd266dde50fdeaf1e08c0752232edd\": container with ID starting with 0a0f187f033b476a338f65809dda43a29acd266dde50fdeaf1e08c0752232edd not found: ID does not exist" containerID="0a0f187f033b476a338f65809dda43a29acd266dde50fdeaf1e08c0752232edd" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.620325 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a0f187f033b476a338f65809dda43a29acd266dde50fdeaf1e08c0752232edd"} err="failed to get container status \"0a0f187f033b476a338f65809dda43a29acd266dde50fdeaf1e08c0752232edd\": rpc error: code = NotFound desc = could not find container \"0a0f187f033b476a338f65809dda43a29acd266dde50fdeaf1e08c0752232edd\": container with ID starting with 0a0f187f033b476a338f65809dda43a29acd266dde50fdeaf1e08c0752232edd not found: ID does not exist" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.684116 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_0660b4d6-650f-47a4-8f78-b6a4fba01954/openstack-network-exporter/0.log" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.753653 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_0660b4d6-650f-47a4-8f78-b6a4fba01954/ovsdbserver-sb/0.log" Dec 05 10:33:53 crc kubenswrapper[4997]: I1205 10:33:53.762269 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e341343e-c6c2-43b0-a712-77bc2303f4e0" path="/var/lib/kubelet/pods/e341343e-c6c2-43b0-a712-77bc2303f4e0/volumes" Dec 05 10:33:54 crc kubenswrapper[4997]: I1205 10:33:54.094727 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-565c9dc47b-g8s9d_bd651c2b-8f0e-4caf-967b-36b0ccdbe54a/placement-api/0.log" Dec 05 10:33:54 crc kubenswrapper[4997]: I1205 10:33:54.105077 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-cx8r7m_10a290ee-4d1b-4e9d-b747-9a4e9698a0cf/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Dec 05 10:33:54 crc kubenswrapper[4997]: I1205 10:33:54.183039 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-565c9dc47b-g8s9d_bd651c2b-8f0e-4caf-967b-36b0ccdbe54a/placement-log/0.log" Dec 05 10:33:54 crc kubenswrapper[4997]: I1205 10:33:54.326934 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-n9vk4n_a364f0ce-49a9-4284-982f-d303bddb5685/pre-adoption-validation-openstack-pre-adoption-openstack-networ/0.log" Dec 05 10:33:54 crc kubenswrapper[4997]: I1205 10:33:54.935924 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_226e3599-1e1d-4216-b76a-7161e5c57556/init-config-reloader/0.log" Dec 05 10:33:54 crc kubenswrapper[4997]: I1205 10:33:54.935985 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_226e3599-1e1d-4216-b76a-7161e5c57556/init-config-reloader/0.log" Dec 05 10:33:54 crc kubenswrapper[4997]: I1205 10:33:54.938568 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_226e3599-1e1d-4216-b76a-7161e5c57556/config-reloader/0.log" Dec 05 10:33:54 crc kubenswrapper[4997]: I1205 10:33:54.939000 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_226e3599-1e1d-4216-b76a-7161e5c57556/prometheus/0.log" Dec 05 10:33:55 crc kubenswrapper[4997]: I1205 10:33:55.157832 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_522f749b-2397-472f-bde2-b0f8fb8419a8/setup-container/0.log" Dec 05 10:33:55 crc kubenswrapper[4997]: I1205 10:33:55.186103 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_226e3599-1e1d-4216-b76a-7161e5c57556/thanos-sidecar/0.log" Dec 05 10:33:55 crc kubenswrapper[4997]: I1205 10:33:55.408427 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_522f749b-2397-472f-bde2-b0f8fb8419a8/setup-container/0.log" Dec 05 10:33:55 crc kubenswrapper[4997]: I1205 10:33:55.415347 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_522f749b-2397-472f-bde2-b0f8fb8419a8/rabbitmq/0.log" Dec 05 10:33:55 crc kubenswrapper[4997]: I1205 10:33:55.512256 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a3c52965-15ba-45b0-b0fc-b447ae56547e/setup-container/0.log" Dec 05 10:33:55 crc kubenswrapper[4997]: I1205 10:33:55.723559 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a3c52965-15ba-45b0-b0fc-b447ae56547e/setup-container/0.log" Dec 05 10:33:55 crc kubenswrapper[4997]: I1205 10:33:55.807320 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-bmkjw_6e5fc8d2-9421-49e7-9007-697047cddecd/reboot-os-openstack-openstack-cell1/0.log" Dec 05 10:33:56 crc kubenswrapper[4997]: I1205 10:33:56.047423 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-networker-ghd89_8a7a8d6a-c2df-4df5-8079-695c04523bf9/reboot-os-openstack-openstack-networker/0.log" Dec 05 10:33:56 crc kubenswrapper[4997]: I1205 10:33:56.203759 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-x8tvh_fa02f03a-593b-447a-a94b-99484b6cd084/run-os-openstack-openstack-cell1/0.log" Dec 05 10:33:56 crc kubenswrapper[4997]: I1205 10:33:56.313831 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-networker-spxhg_9c036120-5481-45e6-8575-5df04a770fd1/run-os-openstack-openstack-networker/0.log" Dec 05 10:33:56 crc kubenswrapper[4997]: I1205 10:33:56.495215 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-26zr2_1fc2d23b-1a50-4016-81ab-9df0a6c8d021/ssh-known-hosts-openstack/0.log" Dec 05 10:33:56 crc kubenswrapper[4997]: I1205 10:33:56.774984 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-q9b5m_746846bb-be60-4333-a253-f1f3057ea2d7/telemetry-openstack-openstack-cell1/0.log" Dec 05 10:33:56 crc kubenswrapper[4997]: I1205 10:33:56.933916 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_8688d151-9d21-4107-86cd-f9cee40d08b8/tempest-tests-tempest-tests-runner/0.log" Dec 05 10:33:56 crc kubenswrapper[4997]: I1205 10:33:56.976783 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_d45c99f0-7cf6-4abc-85d3-529596e9f01a/test-operator-logs-container/0.log" Dec 05 10:33:57 crc kubenswrapper[4997]: I1205 10:33:57.314417 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-tnbnx_82e1e5f9-2d07-4e81-a32a-a35535dd9d2d/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Dec 05 10:33:57 crc kubenswrapper[4997]: I1205 10:33:57.471963 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a3c52965-15ba-45b0-b0fc-b447ae56547e/rabbitmq/0.log" Dec 05 10:33:57 crc kubenswrapper[4997]: I1205 10:33:57.497059 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-networker-6n4l2_6d475848-9ccb-482d-aa85-f6c9e3ef4ef8/tripleo-cleanup-tripleo-cleanup-openstack-networker/0.log" Dec 05 10:33:57 crc kubenswrapper[4997]: I1205 10:33:57.527644 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-4gmlh_8c00bc35-9325-4d86-b9d5-0871b4842954/validate-network-openstack-openstack-cell1/0.log" Dec 05 10:33:57 crc kubenswrapper[4997]: I1205 10:33:57.748800 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-networker-s9llr_10ee156b-3801-425b-b9cc-33c89f2912cd/validate-network-openstack-openstack-networker/0.log" Dec 05 10:34:10 crc kubenswrapper[4997]: I1205 10:34:10.289921 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7l2nz"] Dec 05 10:34:10 crc kubenswrapper[4997]: E1205 10:34:10.291211 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e341343e-c6c2-43b0-a712-77bc2303f4e0" containerName="extract-utilities" Dec 05 10:34:10 crc kubenswrapper[4997]: I1205 10:34:10.291229 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e341343e-c6c2-43b0-a712-77bc2303f4e0" containerName="extract-utilities" Dec 05 10:34:10 crc kubenswrapper[4997]: E1205 10:34:10.291259 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e341343e-c6c2-43b0-a712-77bc2303f4e0" containerName="extract-content" Dec 05 10:34:10 crc kubenswrapper[4997]: I1205 10:34:10.291265 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e341343e-c6c2-43b0-a712-77bc2303f4e0" containerName="extract-content" Dec 05 10:34:10 crc kubenswrapper[4997]: E1205 10:34:10.291292 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e341343e-c6c2-43b0-a712-77bc2303f4e0" containerName="registry-server" Dec 05 10:34:10 crc kubenswrapper[4997]: I1205 10:34:10.291298 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="e341343e-c6c2-43b0-a712-77bc2303f4e0" containerName="registry-server" Dec 05 10:34:10 crc kubenswrapper[4997]: I1205 10:34:10.291500 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="e341343e-c6c2-43b0-a712-77bc2303f4e0" containerName="registry-server" Dec 05 10:34:10 crc kubenswrapper[4997]: I1205 10:34:10.293010 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7l2nz" Dec 05 10:34:10 crc kubenswrapper[4997]: I1205 10:34:10.305352 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7l2nz"] Dec 05 10:34:10 crc kubenswrapper[4997]: I1205 10:34:10.404915 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mss6h\" (UniqueName: \"kubernetes.io/projected/84acfaab-2ce8-4171-ad49-7f316f7faa61-kube-api-access-mss6h\") pod \"community-operators-7l2nz\" (UID: \"84acfaab-2ce8-4171-ad49-7f316f7faa61\") " pod="openshift-marketplace/community-operators-7l2nz" Dec 05 10:34:10 crc kubenswrapper[4997]: I1205 10:34:10.405131 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84acfaab-2ce8-4171-ad49-7f316f7faa61-utilities\") pod \"community-operators-7l2nz\" (UID: \"84acfaab-2ce8-4171-ad49-7f316f7faa61\") " pod="openshift-marketplace/community-operators-7l2nz" Dec 05 10:34:10 crc kubenswrapper[4997]: I1205 10:34:10.405347 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84acfaab-2ce8-4171-ad49-7f316f7faa61-catalog-content\") pod \"community-operators-7l2nz\" (UID: \"84acfaab-2ce8-4171-ad49-7f316f7faa61\") " pod="openshift-marketplace/community-operators-7l2nz" Dec 05 10:34:10 crc kubenswrapper[4997]: I1205 10:34:10.506923 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mss6h\" (UniqueName: \"kubernetes.io/projected/84acfaab-2ce8-4171-ad49-7f316f7faa61-kube-api-access-mss6h\") pod \"community-operators-7l2nz\" (UID: \"84acfaab-2ce8-4171-ad49-7f316f7faa61\") " pod="openshift-marketplace/community-operators-7l2nz" Dec 05 10:34:10 crc kubenswrapper[4997]: I1205 10:34:10.508546 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84acfaab-2ce8-4171-ad49-7f316f7faa61-utilities\") pod \"community-operators-7l2nz\" (UID: \"84acfaab-2ce8-4171-ad49-7f316f7faa61\") " pod="openshift-marketplace/community-operators-7l2nz" Dec 05 10:34:10 crc kubenswrapper[4997]: I1205 10:34:10.508699 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84acfaab-2ce8-4171-ad49-7f316f7faa61-catalog-content\") pod \"community-operators-7l2nz\" (UID: \"84acfaab-2ce8-4171-ad49-7f316f7faa61\") " pod="openshift-marketplace/community-operators-7l2nz" Dec 05 10:34:10 crc kubenswrapper[4997]: I1205 10:34:10.509097 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84acfaab-2ce8-4171-ad49-7f316f7faa61-utilities\") pod \"community-operators-7l2nz\" (UID: \"84acfaab-2ce8-4171-ad49-7f316f7faa61\") " pod="openshift-marketplace/community-operators-7l2nz" Dec 05 10:34:10 crc kubenswrapper[4997]: I1205 10:34:10.509377 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84acfaab-2ce8-4171-ad49-7f316f7faa61-catalog-content\") pod \"community-operators-7l2nz\" (UID: \"84acfaab-2ce8-4171-ad49-7f316f7faa61\") " pod="openshift-marketplace/community-operators-7l2nz" Dec 05 10:34:10 crc kubenswrapper[4997]: I1205 10:34:10.538287 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mss6h\" (UniqueName: \"kubernetes.io/projected/84acfaab-2ce8-4171-ad49-7f316f7faa61-kube-api-access-mss6h\") pod \"community-operators-7l2nz\" (UID: \"84acfaab-2ce8-4171-ad49-7f316f7faa61\") " pod="openshift-marketplace/community-operators-7l2nz" Dec 05 10:34:10 crc kubenswrapper[4997]: I1205 10:34:10.624063 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7l2nz" Dec 05 10:34:11 crc kubenswrapper[4997]: I1205 10:34:11.211866 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7l2nz"] Dec 05 10:34:11 crc kubenswrapper[4997]: I1205 10:34:11.754224 4997 generic.go:334] "Generic (PLEG): container finished" podID="84acfaab-2ce8-4171-ad49-7f316f7faa61" containerID="5219be148cee3359e5a72b4db92526573d11d6a38a59df61c0deb154b6a92a03" exitCode=0 Dec 05 10:34:11 crc kubenswrapper[4997]: I1205 10:34:11.759816 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7l2nz" event={"ID":"84acfaab-2ce8-4171-ad49-7f316f7faa61","Type":"ContainerDied","Data":"5219be148cee3359e5a72b4db92526573d11d6a38a59df61c0deb154b6a92a03"} Dec 05 10:34:11 crc kubenswrapper[4997]: I1205 10:34:11.759856 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7l2nz" event={"ID":"84acfaab-2ce8-4171-ad49-7f316f7faa61","Type":"ContainerStarted","Data":"045793257c6e416eb1b0511781d08db41ade24543d08b1fd8a34e25d32f41e5f"} Dec 05 10:34:12 crc kubenswrapper[4997]: I1205 10:34:12.769752 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_2800a118-2688-4e45-8554-6d613c61d3ad/memcached/0.log" Dec 05 10:34:13 crc kubenswrapper[4997]: I1205 10:34:13.805044 4997 generic.go:334] "Generic (PLEG): container finished" podID="84acfaab-2ce8-4171-ad49-7f316f7faa61" containerID="5301ea92725f4625d151cc084227220dc6952f56021a1185da407481cd5703f3" exitCode=0 Dec 05 10:34:13 crc kubenswrapper[4997]: I1205 10:34:13.805444 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7l2nz" event={"ID":"84acfaab-2ce8-4171-ad49-7f316f7faa61","Type":"ContainerDied","Data":"5301ea92725f4625d151cc084227220dc6952f56021a1185da407481cd5703f3"} Dec 05 10:34:14 crc kubenswrapper[4997]: I1205 10:34:14.818543 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7l2nz" event={"ID":"84acfaab-2ce8-4171-ad49-7f316f7faa61","Type":"ContainerStarted","Data":"16b2c7c20f8dd0d1540bcff68c65a3de2970dea85ebc0d80900f860bb39574dd"} Dec 05 10:34:14 crc kubenswrapper[4997]: I1205 10:34:14.862233 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7l2nz" podStartSLOduration=2.393640484 podStartE2EDuration="4.862204371s" podCreationTimestamp="2025-12-05 10:34:10 +0000 UTC" firstStartedPulling="2025-12-05 10:34:11.756087058 +0000 UTC m=+13152.284994319" lastFinishedPulling="2025-12-05 10:34:14.224650945 +0000 UTC m=+13154.753558206" observedRunningTime="2025-12-05 10:34:14.839445626 +0000 UTC m=+13155.368352927" watchObservedRunningTime="2025-12-05 10:34:14.862204371 +0000 UTC m=+13155.391111652" Dec 05 10:34:19 crc kubenswrapper[4997]: I1205 10:34:19.770380 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:34:19 crc kubenswrapper[4997]: I1205 10:34:19.771528 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:34:20 crc kubenswrapper[4997]: I1205 10:34:20.624358 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7l2nz" Dec 05 10:34:20 crc kubenswrapper[4997]: I1205 10:34:20.624426 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7l2nz" Dec 05 10:34:20 crc kubenswrapper[4997]: I1205 10:34:20.676935 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7l2nz" Dec 05 10:34:20 crc kubenswrapper[4997]: I1205 10:34:20.942247 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7l2nz" Dec 05 10:34:20 crc kubenswrapper[4997]: I1205 10:34:20.990162 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7l2nz"] Dec 05 10:34:22 crc kubenswrapper[4997]: I1205 10:34:22.907348 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7l2nz" podUID="84acfaab-2ce8-4171-ad49-7f316f7faa61" containerName="registry-server" containerID="cri-o://16b2c7c20f8dd0d1540bcff68c65a3de2970dea85ebc0d80900f860bb39574dd" gracePeriod=2 Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.395667 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7l2nz" Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.487133 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84acfaab-2ce8-4171-ad49-7f316f7faa61-catalog-content\") pod \"84acfaab-2ce8-4171-ad49-7f316f7faa61\" (UID: \"84acfaab-2ce8-4171-ad49-7f316f7faa61\") " Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.487235 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mss6h\" (UniqueName: \"kubernetes.io/projected/84acfaab-2ce8-4171-ad49-7f316f7faa61-kube-api-access-mss6h\") pod \"84acfaab-2ce8-4171-ad49-7f316f7faa61\" (UID: \"84acfaab-2ce8-4171-ad49-7f316f7faa61\") " Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.487515 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84acfaab-2ce8-4171-ad49-7f316f7faa61-utilities\") pod \"84acfaab-2ce8-4171-ad49-7f316f7faa61\" (UID: \"84acfaab-2ce8-4171-ad49-7f316f7faa61\") " Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.488324 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84acfaab-2ce8-4171-ad49-7f316f7faa61-utilities" (OuterVolumeSpecName: "utilities") pod "84acfaab-2ce8-4171-ad49-7f316f7faa61" (UID: "84acfaab-2ce8-4171-ad49-7f316f7faa61"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.492344 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84acfaab-2ce8-4171-ad49-7f316f7faa61-kube-api-access-mss6h" (OuterVolumeSpecName: "kube-api-access-mss6h") pod "84acfaab-2ce8-4171-ad49-7f316f7faa61" (UID: "84acfaab-2ce8-4171-ad49-7f316f7faa61"). InnerVolumeSpecName "kube-api-access-mss6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.537370 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84acfaab-2ce8-4171-ad49-7f316f7faa61-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84acfaab-2ce8-4171-ad49-7f316f7faa61" (UID: "84acfaab-2ce8-4171-ad49-7f316f7faa61"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.589957 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84acfaab-2ce8-4171-ad49-7f316f7faa61-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.590266 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mss6h\" (UniqueName: \"kubernetes.io/projected/84acfaab-2ce8-4171-ad49-7f316f7faa61-kube-api-access-mss6h\") on node \"crc\" DevicePath \"\"" Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.590282 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84acfaab-2ce8-4171-ad49-7f316f7faa61-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.921815 4997 generic.go:334] "Generic (PLEG): container finished" podID="84acfaab-2ce8-4171-ad49-7f316f7faa61" containerID="16b2c7c20f8dd0d1540bcff68c65a3de2970dea85ebc0d80900f860bb39574dd" exitCode=0 Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.921867 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7l2nz" event={"ID":"84acfaab-2ce8-4171-ad49-7f316f7faa61","Type":"ContainerDied","Data":"16b2c7c20f8dd0d1540bcff68c65a3de2970dea85ebc0d80900f860bb39574dd"} Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.921944 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7l2nz" event={"ID":"84acfaab-2ce8-4171-ad49-7f316f7faa61","Type":"ContainerDied","Data":"045793257c6e416eb1b0511781d08db41ade24543d08b1fd8a34e25d32f41e5f"} Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.921962 4997 scope.go:117] "RemoveContainer" containerID="16b2c7c20f8dd0d1540bcff68c65a3de2970dea85ebc0d80900f860bb39574dd" Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.922217 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7l2nz" Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.946801 4997 scope.go:117] "RemoveContainer" containerID="5301ea92725f4625d151cc084227220dc6952f56021a1185da407481cd5703f3" Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.947199 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7l2nz"] Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.961133 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7l2nz"] Dec 05 10:34:23 crc kubenswrapper[4997]: I1205 10:34:23.985232 4997 scope.go:117] "RemoveContainer" containerID="5219be148cee3359e5a72b4db92526573d11d6a38a59df61c0deb154b6a92a03" Dec 05 10:34:24 crc kubenswrapper[4997]: I1205 10:34:24.034230 4997 scope.go:117] "RemoveContainer" containerID="16b2c7c20f8dd0d1540bcff68c65a3de2970dea85ebc0d80900f860bb39574dd" Dec 05 10:34:24 crc kubenswrapper[4997]: E1205 10:34:24.034928 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16b2c7c20f8dd0d1540bcff68c65a3de2970dea85ebc0d80900f860bb39574dd\": container with ID starting with 16b2c7c20f8dd0d1540bcff68c65a3de2970dea85ebc0d80900f860bb39574dd not found: ID does not exist" containerID="16b2c7c20f8dd0d1540bcff68c65a3de2970dea85ebc0d80900f860bb39574dd" Dec 05 10:34:24 crc kubenswrapper[4997]: I1205 10:34:24.034966 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16b2c7c20f8dd0d1540bcff68c65a3de2970dea85ebc0d80900f860bb39574dd"} err="failed to get container status \"16b2c7c20f8dd0d1540bcff68c65a3de2970dea85ebc0d80900f860bb39574dd\": rpc error: code = NotFound desc = could not find container \"16b2c7c20f8dd0d1540bcff68c65a3de2970dea85ebc0d80900f860bb39574dd\": container with ID starting with 16b2c7c20f8dd0d1540bcff68c65a3de2970dea85ebc0d80900f860bb39574dd not found: ID does not exist" Dec 05 10:34:24 crc kubenswrapper[4997]: I1205 10:34:24.034995 4997 scope.go:117] "RemoveContainer" containerID="5301ea92725f4625d151cc084227220dc6952f56021a1185da407481cd5703f3" Dec 05 10:34:24 crc kubenswrapper[4997]: E1205 10:34:24.035400 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5301ea92725f4625d151cc084227220dc6952f56021a1185da407481cd5703f3\": container with ID starting with 5301ea92725f4625d151cc084227220dc6952f56021a1185da407481cd5703f3 not found: ID does not exist" containerID="5301ea92725f4625d151cc084227220dc6952f56021a1185da407481cd5703f3" Dec 05 10:34:24 crc kubenswrapper[4997]: I1205 10:34:24.035428 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5301ea92725f4625d151cc084227220dc6952f56021a1185da407481cd5703f3"} err="failed to get container status \"5301ea92725f4625d151cc084227220dc6952f56021a1185da407481cd5703f3\": rpc error: code = NotFound desc = could not find container \"5301ea92725f4625d151cc084227220dc6952f56021a1185da407481cd5703f3\": container with ID starting with 5301ea92725f4625d151cc084227220dc6952f56021a1185da407481cd5703f3 not found: ID does not exist" Dec 05 10:34:24 crc kubenswrapper[4997]: I1205 10:34:24.035447 4997 scope.go:117] "RemoveContainer" containerID="5219be148cee3359e5a72b4db92526573d11d6a38a59df61c0deb154b6a92a03" Dec 05 10:34:24 crc kubenswrapper[4997]: E1205 10:34:24.035790 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5219be148cee3359e5a72b4db92526573d11d6a38a59df61c0deb154b6a92a03\": container with ID starting with 5219be148cee3359e5a72b4db92526573d11d6a38a59df61c0deb154b6a92a03 not found: ID does not exist" containerID="5219be148cee3359e5a72b4db92526573d11d6a38a59df61c0deb154b6a92a03" Dec 05 10:34:24 crc kubenswrapper[4997]: I1205 10:34:24.035822 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5219be148cee3359e5a72b4db92526573d11d6a38a59df61c0deb154b6a92a03"} err="failed to get container status \"5219be148cee3359e5a72b4db92526573d11d6a38a59df61c0deb154b6a92a03\": rpc error: code = NotFound desc = could not find container \"5219be148cee3359e5a72b4db92526573d11d6a38a59df61c0deb154b6a92a03\": container with ID starting with 5219be148cee3359e5a72b4db92526573d11d6a38a59df61c0deb154b6a92a03 not found: ID does not exist" Dec 05 10:34:25 crc kubenswrapper[4997]: I1205 10:34:25.762467 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84acfaab-2ce8-4171-ad49-7f316f7faa61" path="/var/lib/kubelet/pods/84acfaab-2ce8-4171-ad49-7f316f7faa61/volumes" Dec 05 10:34:26 crc kubenswrapper[4997]: I1205 10:34:26.176061 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9_aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5/util/0.log" Dec 05 10:34:26 crc kubenswrapper[4997]: I1205 10:34:26.283873 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9_aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5/pull/0.log" Dec 05 10:34:26 crc kubenswrapper[4997]: I1205 10:34:26.362496 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9_aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5/pull/0.log" Dec 05 10:34:26 crc kubenswrapper[4997]: I1205 10:34:26.377881 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9_aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5/util/0.log" Dec 05 10:34:26 crc kubenswrapper[4997]: I1205 10:34:26.511052 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9_aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5/util/0.log" Dec 05 10:34:26 crc kubenswrapper[4997]: I1205 10:34:26.559377 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9_aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5/pull/0.log" Dec 05 10:34:26 crc kubenswrapper[4997]: I1205 10:34:26.568206 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafn5gm9_aa6cb2d5-6fc0-4049-a6a0-bd081ee0aee5/extract/0.log" Dec 05 10:34:26 crc kubenswrapper[4997]: I1205 10:34:26.753226 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-r4m8t_fe84dde4-76f7-4df6-a6fd-d1b45d097e9b/kube-rbac-proxy/0.log" Dec 05 10:34:26 crc kubenswrapper[4997]: I1205 10:34:26.860856 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-rnq5d_27e5617b-9979-45b2-a526-bb671edf63c2/kube-rbac-proxy/0.log" Dec 05 10:34:26 crc kubenswrapper[4997]: I1205 10:34:26.875481 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-r4m8t_fe84dde4-76f7-4df6-a6fd-d1b45d097e9b/manager/0.log" Dec 05 10:34:27 crc kubenswrapper[4997]: I1205 10:34:27.067555 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-rnq5d_27e5617b-9979-45b2-a526-bb671edf63c2/manager/0.log" Dec 05 10:34:27 crc kubenswrapper[4997]: I1205 10:34:27.147334 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-r9mh5_351c9912-1ac6-4b8b-b009-d073b1bd28a5/manager/0.log" Dec 05 10:34:27 crc kubenswrapper[4997]: I1205 10:34:27.185697 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-r9mh5_351c9912-1ac6-4b8b-b009-d073b1bd28a5/kube-rbac-proxy/0.log" Dec 05 10:34:27 crc kubenswrapper[4997]: I1205 10:34:27.288742 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-lpx4g_d5054dc6-123a-4d45-8340-5315c0c57e32/kube-rbac-proxy/0.log" Dec 05 10:34:27 crc kubenswrapper[4997]: I1205 10:34:27.567057 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-lpx4g_d5054dc6-123a-4d45-8340-5315c0c57e32/manager/0.log" Dec 05 10:34:27 crc kubenswrapper[4997]: I1205 10:34:27.608337 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-5zk9v_c488dae2-6bbc-434a-a1c8-d3bd6ba5a549/kube-rbac-proxy/0.log" Dec 05 10:34:27 crc kubenswrapper[4997]: I1205 10:34:27.681556 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-5zk9v_c488dae2-6bbc-434a-a1c8-d3bd6ba5a549/manager/0.log" Dec 05 10:34:27 crc kubenswrapper[4997]: I1205 10:34:27.883202 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-jk2cb_7db21a3e-655e-4874-8366-dfdb5d2521f5/manager/0.log" Dec 05 10:34:27 crc kubenswrapper[4997]: I1205 10:34:27.930377 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-jk2cb_7db21a3e-655e-4874-8366-dfdb5d2521f5/kube-rbac-proxy/0.log" Dec 05 10:34:28 crc kubenswrapper[4997]: I1205 10:34:28.107454 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-pppgg_6aac09d2-991d-45b9-bba9-4cf2d70278ab/kube-rbac-proxy/0.log" Dec 05 10:34:28 crc kubenswrapper[4997]: I1205 10:34:28.303201 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-q62zq_c9d46c14-6c5e-42d4-81f3-43e7169d23a0/kube-rbac-proxy/0.log" Dec 05 10:34:28 crc kubenswrapper[4997]: I1205 10:34:28.444307 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-q62zq_c9d46c14-6c5e-42d4-81f3-43e7169d23a0/manager/0.log" Dec 05 10:34:28 crc kubenswrapper[4997]: I1205 10:34:28.515129 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-pppgg_6aac09d2-991d-45b9-bba9-4cf2d70278ab/manager/0.log" Dec 05 10:34:28 crc kubenswrapper[4997]: I1205 10:34:28.618789 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-88hm7_40df044a-6922-43cf-ae37-ebc8dd637119/kube-rbac-proxy/0.log" Dec 05 10:34:28 crc kubenswrapper[4997]: I1205 10:34:28.805554 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-88hm7_40df044a-6922-43cf-ae37-ebc8dd637119/manager/0.log" Dec 05 10:34:28 crc kubenswrapper[4997]: I1205 10:34:28.845336 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-sbfxr_3580ba04-55a0-4e33-aa08-932a2648bf2d/kube-rbac-proxy/0.log" Dec 05 10:34:28 crc kubenswrapper[4997]: I1205 10:34:28.896998 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-sbfxr_3580ba04-55a0-4e33-aa08-932a2648bf2d/manager/0.log" Dec 05 10:34:29 crc kubenswrapper[4997]: I1205 10:34:29.019318 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-tpg29_e5f7e30e-e2a8-4131-9c10-1dddfe528684/kube-rbac-proxy/0.log" Dec 05 10:34:29 crc kubenswrapper[4997]: I1205 10:34:29.164637 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-tpg29_e5f7e30e-e2a8-4131-9c10-1dddfe528684/manager/0.log" Dec 05 10:34:29 crc kubenswrapper[4997]: I1205 10:34:29.269604 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-rq99q_7a8b9284-3f13-4396-8f85-36f0c59e1ba3/kube-rbac-proxy/0.log" Dec 05 10:34:29 crc kubenswrapper[4997]: I1205 10:34:29.305221 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-rq99q_7a8b9284-3f13-4396-8f85-36f0c59e1ba3/manager/0.log" Dec 05 10:34:29 crc kubenswrapper[4997]: I1205 10:34:29.417498 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-qppfb_7afb5248-e41e-46ec-9e46-ed9fcc7cf8d3/kube-rbac-proxy/0.log" Dec 05 10:34:29 crc kubenswrapper[4997]: I1205 10:34:29.622451 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-qppfb_7afb5248-e41e-46ec-9e46-ed9fcc7cf8d3/manager/0.log" Dec 05 10:34:29 crc kubenswrapper[4997]: I1205 10:34:29.905143 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-r9knp_9cc09cce-bafa-4f17-bd45-b767fa576e54/manager/0.log" Dec 05 10:34:29 crc kubenswrapper[4997]: I1205 10:34:29.936051 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-r9knp_9cc09cce-bafa-4f17-bd45-b767fa576e54/kube-rbac-proxy/0.log" Dec 05 10:34:30 crc kubenswrapper[4997]: I1205 10:34:30.090357 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55c85496f56jjnr_64d88cd2-49b5-439a-9344-bb9dc4ec5730/kube-rbac-proxy/0.log" Dec 05 10:34:30 crc kubenswrapper[4997]: I1205 10:34:30.172856 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55c85496f56jjnr_64d88cd2-49b5-439a-9344-bb9dc4ec5730/manager/0.log" Dec 05 10:34:30 crc kubenswrapper[4997]: I1205 10:34:30.719295 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-hrs7p_3f6028fa-7b09-49ec-a8ed-4f4b25419e19/kube-rbac-proxy/0.log" Dec 05 10:34:30 crc kubenswrapper[4997]: I1205 10:34:30.765186 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-55b6fb9447-259kk_90da0714-e130-42be-acbb-845291fb3d9a/operator/0.log" Dec 05 10:34:30 crc kubenswrapper[4997]: I1205 10:34:30.876438 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-kc298_395fd5d1-b4de-4859-ba11-2f2fa89395d2/registry-server/0.log" Dec 05 10:34:31 crc kubenswrapper[4997]: I1205 10:34:31.030478 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-ztwsn_905ae077-9cc3-41a9-b03f-77b3eeb052b3/kube-rbac-proxy/0.log" Dec 05 10:34:31 crc kubenswrapper[4997]: I1205 10:34:31.068998 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-hrs7p_3f6028fa-7b09-49ec-a8ed-4f4b25419e19/manager/0.log" Dec 05 10:34:31 crc kubenswrapper[4997]: I1205 10:34:31.174205 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-ztwsn_905ae077-9cc3-41a9-b03f-77b3eeb052b3/manager/0.log" Dec 05 10:34:31 crc kubenswrapper[4997]: I1205 10:34:31.328493 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-qsqst_8d31795c-bf5a-48fc-85cc-0c1385c2d7b0/operator/0.log" Dec 05 10:34:31 crc kubenswrapper[4997]: I1205 10:34:31.444680 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-jtjgd_9597d662-50ea-4066-8bbc-232cc1d90dbd/kube-rbac-proxy/0.log" Dec 05 10:34:31 crc kubenswrapper[4997]: I1205 10:34:31.494366 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-jtjgd_9597d662-50ea-4066-8bbc-232cc1d90dbd/manager/0.log" Dec 05 10:34:31 crc kubenswrapper[4997]: I1205 10:34:31.630165 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-cjggt_e1874b23-363f-4303-bfeb-86282b20963d/kube-rbac-proxy/0.log" Dec 05 10:34:31 crc kubenswrapper[4997]: I1205 10:34:31.876921 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-kbfbg_25ca1d60-a629-447e-afa7-7a9f0e6f0b7e/kube-rbac-proxy/0.log" Dec 05 10:34:31 crc kubenswrapper[4997]: I1205 10:34:31.921969 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-kbfbg_25ca1d60-a629-447e-afa7-7a9f0e6f0b7e/manager/0.log" Dec 05 10:34:32 crc kubenswrapper[4997]: I1205 10:34:32.079967 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-cjggt_e1874b23-363f-4303-bfeb-86282b20963d/manager/0.log" Dec 05 10:34:32 crc kubenswrapper[4997]: I1205 10:34:32.145527 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-skcxd_f309b30b-9712-4175-b237-7c35ac2e5dee/kube-rbac-proxy/0.log" Dec 05 10:34:32 crc kubenswrapper[4997]: I1205 10:34:32.242369 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-skcxd_f309b30b-9712-4175-b237-7c35ac2e5dee/manager/0.log" Dec 05 10:34:33 crc kubenswrapper[4997]: I1205 10:34:33.045600 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-54bdf956c4-7wbfk_6c477704-084c-4d28-8ab3-cd0f29ca488f/manager/0.log" Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.040931 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7ggp6"] Dec 05 10:34:48 crc kubenswrapper[4997]: E1205 10:34:48.042177 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84acfaab-2ce8-4171-ad49-7f316f7faa61" containerName="registry-server" Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.042195 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="84acfaab-2ce8-4171-ad49-7f316f7faa61" containerName="registry-server" Dec 05 10:34:48 crc kubenswrapper[4997]: E1205 10:34:48.042210 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84acfaab-2ce8-4171-ad49-7f316f7faa61" containerName="extract-utilities" Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.042216 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="84acfaab-2ce8-4171-ad49-7f316f7faa61" containerName="extract-utilities" Dec 05 10:34:48 crc kubenswrapper[4997]: E1205 10:34:48.042226 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84acfaab-2ce8-4171-ad49-7f316f7faa61" containerName="extract-content" Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.042231 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="84acfaab-2ce8-4171-ad49-7f316f7faa61" containerName="extract-content" Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.042493 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="84acfaab-2ce8-4171-ad49-7f316f7faa61" containerName="registry-server" Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.044315 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7ggp6" Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.060441 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7ggp6"] Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.085076 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-catalog-content\") pod \"redhat-operators-7ggp6\" (UID: \"a0c8978e-08e9-4c27-8f13-d83b35a02ca7\") " pod="openshift-marketplace/redhat-operators-7ggp6" Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.085331 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-utilities\") pod \"redhat-operators-7ggp6\" (UID: \"a0c8978e-08e9-4c27-8f13-d83b35a02ca7\") " pod="openshift-marketplace/redhat-operators-7ggp6" Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.085476 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpx2h\" (UniqueName: \"kubernetes.io/projected/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-kube-api-access-jpx2h\") pod \"redhat-operators-7ggp6\" (UID: \"a0c8978e-08e9-4c27-8f13-d83b35a02ca7\") " pod="openshift-marketplace/redhat-operators-7ggp6" Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.187176 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpx2h\" (UniqueName: \"kubernetes.io/projected/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-kube-api-access-jpx2h\") pod \"redhat-operators-7ggp6\" (UID: \"a0c8978e-08e9-4c27-8f13-d83b35a02ca7\") " pod="openshift-marketplace/redhat-operators-7ggp6" Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.187557 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-catalog-content\") pod \"redhat-operators-7ggp6\" (UID: \"a0c8978e-08e9-4c27-8f13-d83b35a02ca7\") " pod="openshift-marketplace/redhat-operators-7ggp6" Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.188354 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-catalog-content\") pod \"redhat-operators-7ggp6\" (UID: \"a0c8978e-08e9-4c27-8f13-d83b35a02ca7\") " pod="openshift-marketplace/redhat-operators-7ggp6" Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.188501 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-utilities\") pod \"redhat-operators-7ggp6\" (UID: \"a0c8978e-08e9-4c27-8f13-d83b35a02ca7\") " pod="openshift-marketplace/redhat-operators-7ggp6" Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.189091 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-utilities\") pod \"redhat-operators-7ggp6\" (UID: \"a0c8978e-08e9-4c27-8f13-d83b35a02ca7\") " pod="openshift-marketplace/redhat-operators-7ggp6" Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.223571 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpx2h\" (UniqueName: \"kubernetes.io/projected/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-kube-api-access-jpx2h\") pod \"redhat-operators-7ggp6\" (UID: \"a0c8978e-08e9-4c27-8f13-d83b35a02ca7\") " pod="openshift-marketplace/redhat-operators-7ggp6" Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.366400 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7ggp6" Dec 05 10:34:48 crc kubenswrapper[4997]: I1205 10:34:48.898497 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7ggp6"] Dec 05 10:34:49 crc kubenswrapper[4997]: I1205 10:34:49.184570 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ggp6" event={"ID":"a0c8978e-08e9-4c27-8f13-d83b35a02ca7","Type":"ContainerStarted","Data":"3f7b90b32962f165827948736aa7e53f93bc5036c172a002d4011f0990cb32dd"} Dec 05 10:34:49 crc kubenswrapper[4997]: I1205 10:34:49.770427 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:34:49 crc kubenswrapper[4997]: I1205 10:34:49.770984 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:34:49 crc kubenswrapper[4997]: I1205 10:34:49.771047 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 10:34:49 crc kubenswrapper[4997]: I1205 10:34:49.772292 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 10:34:49 crc kubenswrapper[4997]: I1205 10:34:49.772370 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" gracePeriod=600 Dec 05 10:34:49 crc kubenswrapper[4997]: E1205 10:34:49.905146 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:34:50 crc kubenswrapper[4997]: I1205 10:34:50.217981 4997 generic.go:334] "Generic (PLEG): container finished" podID="a0c8978e-08e9-4c27-8f13-d83b35a02ca7" containerID="97735aef0c0d17666f057fc9cf957ec68a6c08acac0d1869780773c083e06c7a" exitCode=0 Dec 05 10:34:50 crc kubenswrapper[4997]: I1205 10:34:50.218092 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ggp6" event={"ID":"a0c8978e-08e9-4c27-8f13-d83b35a02ca7","Type":"ContainerDied","Data":"97735aef0c0d17666f057fc9cf957ec68a6c08acac0d1869780773c083e06c7a"} Dec 05 10:34:50 crc kubenswrapper[4997]: I1205 10:34:50.221988 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" exitCode=0 Dec 05 10:34:50 crc kubenswrapper[4997]: I1205 10:34:50.222081 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648"} Dec 05 10:34:50 crc kubenswrapper[4997]: I1205 10:34:50.222145 4997 scope.go:117] "RemoveContainer" containerID="570182e001027ed044c074b2d6343f7b4f00f9a0f33006cc726adf1d3379a622" Dec 05 10:34:50 crc kubenswrapper[4997]: I1205 10:34:50.236363 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:34:50 crc kubenswrapper[4997]: E1205 10:34:50.279798 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:34:52 crc kubenswrapper[4997]: I1205 10:34:52.255526 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ggp6" event={"ID":"a0c8978e-08e9-4c27-8f13-d83b35a02ca7","Type":"ContainerStarted","Data":"95f65eba53dc07fb22930c242fa66a9cf01baf75ae059a1d1c40cab2a97d0583"} Dec 05 10:34:54 crc kubenswrapper[4997]: I1205 10:34:54.584819 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zs9km_ab4d334f-b4a4-4d04-bf90-d2178f991f5f/control-plane-machine-set-operator/0.log" Dec 05 10:34:54 crc kubenswrapper[4997]: I1205 10:34:54.924036 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wpl2k_f3993123-571c-4be4-a525-d8e0481b1aa5/kube-rbac-proxy/0.log" Dec 05 10:34:54 crc kubenswrapper[4997]: I1205 10:34:54.924754 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wpl2k_f3993123-571c-4be4-a525-d8e0481b1aa5/machine-api-operator/0.log" Dec 05 10:34:57 crc kubenswrapper[4997]: I1205 10:34:57.307345 4997 generic.go:334] "Generic (PLEG): container finished" podID="a0c8978e-08e9-4c27-8f13-d83b35a02ca7" containerID="95f65eba53dc07fb22930c242fa66a9cf01baf75ae059a1d1c40cab2a97d0583" exitCode=0 Dec 05 10:34:57 crc kubenswrapper[4997]: I1205 10:34:57.307444 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ggp6" event={"ID":"a0c8978e-08e9-4c27-8f13-d83b35a02ca7","Type":"ContainerDied","Data":"95f65eba53dc07fb22930c242fa66a9cf01baf75ae059a1d1c40cab2a97d0583"} Dec 05 10:34:58 crc kubenswrapper[4997]: I1205 10:34:58.319375 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ggp6" event={"ID":"a0c8978e-08e9-4c27-8f13-d83b35a02ca7","Type":"ContainerStarted","Data":"377c2b35631fe5cfb05d385837b4dfe7fd05b103541659954347a2b5efc30b71"} Dec 05 10:34:58 crc kubenswrapper[4997]: I1205 10:34:58.343935 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7ggp6" podStartSLOduration=2.860462012 podStartE2EDuration="10.343917326s" podCreationTimestamp="2025-12-05 10:34:48 +0000 UTC" firstStartedPulling="2025-12-05 10:34:50.221767386 +0000 UTC m=+13190.750674657" lastFinishedPulling="2025-12-05 10:34:57.70522271 +0000 UTC m=+13198.234129971" observedRunningTime="2025-12-05 10:34:58.334601375 +0000 UTC m=+13198.863508666" watchObservedRunningTime="2025-12-05 10:34:58.343917326 +0000 UTC m=+13198.872824587" Dec 05 10:34:58 crc kubenswrapper[4997]: I1205 10:34:58.367039 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7ggp6" Dec 05 10:34:58 crc kubenswrapper[4997]: I1205 10:34:58.367101 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7ggp6" Dec 05 10:34:59 crc kubenswrapper[4997]: I1205 10:34:59.413412 4997 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7ggp6" podUID="a0c8978e-08e9-4c27-8f13-d83b35a02ca7" containerName="registry-server" probeResult="failure" output=< Dec 05 10:34:59 crc kubenswrapper[4997]: timeout: failed to connect service ":50051" within 1s Dec 05 10:34:59 crc kubenswrapper[4997]: > Dec 05 10:35:02 crc kubenswrapper[4997]: I1205 10:35:02.748918 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:35:02 crc kubenswrapper[4997]: E1205 10:35:02.749556 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:35:08 crc kubenswrapper[4997]: I1205 10:35:08.424595 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7ggp6" Dec 05 10:35:08 crc kubenswrapper[4997]: I1205 10:35:08.494108 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7ggp6" Dec 05 10:35:08 crc kubenswrapper[4997]: I1205 10:35:08.665185 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7ggp6"] Dec 05 10:35:09 crc kubenswrapper[4997]: I1205 10:35:09.392839 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-p6wrp_bd787551-b226-4b7d-88db-785469feec70/cert-manager-controller/0.log" Dec 05 10:35:09 crc kubenswrapper[4997]: I1205 10:35:09.692747 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-qdq9n_d9e89c33-9251-4912-82c5-c37e1aec78bc/cert-manager-cainjector/0.log" Dec 05 10:35:09 crc kubenswrapper[4997]: I1205 10:35:09.718877 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-8z9rx_849d3f8f-a64b-4b10-a847-d9305ed4c13b/cert-manager-webhook/0.log" Dec 05 10:35:10 crc kubenswrapper[4997]: I1205 10:35:10.440844 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7ggp6" podUID="a0c8978e-08e9-4c27-8f13-d83b35a02ca7" containerName="registry-server" containerID="cri-o://377c2b35631fe5cfb05d385837b4dfe7fd05b103541659954347a2b5efc30b71" gracePeriod=2 Dec 05 10:35:11 crc kubenswrapper[4997]: I1205 10:35:11.458429 4997 generic.go:334] "Generic (PLEG): container finished" podID="a0c8978e-08e9-4c27-8f13-d83b35a02ca7" containerID="377c2b35631fe5cfb05d385837b4dfe7fd05b103541659954347a2b5efc30b71" exitCode=0 Dec 05 10:35:11 crc kubenswrapper[4997]: I1205 10:35:11.458774 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ggp6" event={"ID":"a0c8978e-08e9-4c27-8f13-d83b35a02ca7","Type":"ContainerDied","Data":"377c2b35631fe5cfb05d385837b4dfe7fd05b103541659954347a2b5efc30b71"} Dec 05 10:35:12 crc kubenswrapper[4997]: I1205 10:35:12.134505 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7ggp6" Dec 05 10:35:12 crc kubenswrapper[4997]: I1205 10:35:12.243909 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpx2h\" (UniqueName: \"kubernetes.io/projected/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-kube-api-access-jpx2h\") pod \"a0c8978e-08e9-4c27-8f13-d83b35a02ca7\" (UID: \"a0c8978e-08e9-4c27-8f13-d83b35a02ca7\") " Dec 05 10:35:12 crc kubenswrapper[4997]: I1205 10:35:12.244228 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-catalog-content\") pod \"a0c8978e-08e9-4c27-8f13-d83b35a02ca7\" (UID: \"a0c8978e-08e9-4c27-8f13-d83b35a02ca7\") " Dec 05 10:35:12 crc kubenswrapper[4997]: I1205 10:35:12.244320 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-utilities\") pod \"a0c8978e-08e9-4c27-8f13-d83b35a02ca7\" (UID: \"a0c8978e-08e9-4c27-8f13-d83b35a02ca7\") " Dec 05 10:35:12 crc kubenswrapper[4997]: I1205 10:35:12.245154 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-utilities" (OuterVolumeSpecName: "utilities") pod "a0c8978e-08e9-4c27-8f13-d83b35a02ca7" (UID: "a0c8978e-08e9-4c27-8f13-d83b35a02ca7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:35:12 crc kubenswrapper[4997]: I1205 10:35:12.245639 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:35:12 crc kubenswrapper[4997]: I1205 10:35:12.249691 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-kube-api-access-jpx2h" (OuterVolumeSpecName: "kube-api-access-jpx2h") pod "a0c8978e-08e9-4c27-8f13-d83b35a02ca7" (UID: "a0c8978e-08e9-4c27-8f13-d83b35a02ca7"). InnerVolumeSpecName "kube-api-access-jpx2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:35:12 crc kubenswrapper[4997]: I1205 10:35:12.347955 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpx2h\" (UniqueName: \"kubernetes.io/projected/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-kube-api-access-jpx2h\") on node \"crc\" DevicePath \"\"" Dec 05 10:35:12 crc kubenswrapper[4997]: I1205 10:35:12.353367 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0c8978e-08e9-4c27-8f13-d83b35a02ca7" (UID: "a0c8978e-08e9-4c27-8f13-d83b35a02ca7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:35:12 crc kubenswrapper[4997]: I1205 10:35:12.450463 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0c8978e-08e9-4c27-8f13-d83b35a02ca7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:35:12 crc kubenswrapper[4997]: I1205 10:35:12.483864 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ggp6" event={"ID":"a0c8978e-08e9-4c27-8f13-d83b35a02ca7","Type":"ContainerDied","Data":"3f7b90b32962f165827948736aa7e53f93bc5036c172a002d4011f0990cb32dd"} Dec 05 10:35:12 crc kubenswrapper[4997]: I1205 10:35:12.483942 4997 scope.go:117] "RemoveContainer" containerID="377c2b35631fe5cfb05d385837b4dfe7fd05b103541659954347a2b5efc30b71" Dec 05 10:35:12 crc kubenswrapper[4997]: I1205 10:35:12.484148 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7ggp6" Dec 05 10:35:12 crc kubenswrapper[4997]: I1205 10:35:12.508878 4997 scope.go:117] "RemoveContainer" containerID="95f65eba53dc07fb22930c242fa66a9cf01baf75ae059a1d1c40cab2a97d0583" Dec 05 10:35:12 crc kubenswrapper[4997]: I1205 10:35:12.532637 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7ggp6"] Dec 05 10:35:12 crc kubenswrapper[4997]: I1205 10:35:12.545546 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7ggp6"] Dec 05 10:35:12 crc kubenswrapper[4997]: I1205 10:35:12.558111 4997 scope.go:117] "RemoveContainer" containerID="97735aef0c0d17666f057fc9cf957ec68a6c08acac0d1869780773c083e06c7a" Dec 05 10:35:13 crc kubenswrapper[4997]: I1205 10:35:13.760976 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0c8978e-08e9-4c27-8f13-d83b35a02ca7" path="/var/lib/kubelet/pods/a0c8978e-08e9-4c27-8f13-d83b35a02ca7/volumes" Dec 05 10:35:16 crc kubenswrapper[4997]: I1205 10:35:16.748838 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:35:16 crc kubenswrapper[4997]: E1205 10:35:16.749732 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:35:23 crc kubenswrapper[4997]: I1205 10:35:23.489657 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-gl7kq_81f37659-4cca-4dc1-85cd-f755196a4ca6/nmstate-console-plugin/0.log" Dec 05 10:35:23 crc kubenswrapper[4997]: I1205 10:35:23.679511 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-j74g9_90785160-89df-4c52-b2ed-05c8740f050d/nmstate-handler/0.log" Dec 05 10:35:23 crc kubenswrapper[4997]: I1205 10:35:23.743072 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-vdrwf_5607a1c5-5883-42d9-8d2f-9bb2194ee33d/nmstate-metrics/0.log" Dec 05 10:35:23 crc kubenswrapper[4997]: I1205 10:35:23.772493 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-vdrwf_5607a1c5-5883-42d9-8d2f-9bb2194ee33d/kube-rbac-proxy/0.log" Dec 05 10:35:23 crc kubenswrapper[4997]: I1205 10:35:23.932458 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-r64sz_147e0782-52e6-4511-aa4a-27fe62b112dd/nmstate-operator/0.log" Dec 05 10:35:23 crc kubenswrapper[4997]: I1205 10:35:23.986725 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-4cxth_9dc3f953-29af-435f-ba80-fc7d2f3b0dc6/nmstate-webhook/0.log" Dec 05 10:35:28 crc kubenswrapper[4997]: I1205 10:35:28.748975 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:35:28 crc kubenswrapper[4997]: E1205 10:35:28.749730 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:35:38 crc kubenswrapper[4997]: I1205 10:35:38.842424 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-4qdnj_5e7b1770-04d8-41e3-bdfa-cafc699d03e1/kube-rbac-proxy/0.log" Dec 05 10:35:39 crc kubenswrapper[4997]: I1205 10:35:39.052921 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/cp-frr-files/0.log" Dec 05 10:35:39 crc kubenswrapper[4997]: I1205 10:35:39.083353 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-4qdnj_5e7b1770-04d8-41e3-bdfa-cafc699d03e1/controller/0.log" Dec 05 10:35:39 crc kubenswrapper[4997]: I1205 10:35:39.336476 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/cp-metrics/0.log" Dec 05 10:35:39 crc kubenswrapper[4997]: I1205 10:35:39.337366 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/cp-reloader/0.log" Dec 05 10:35:39 crc kubenswrapper[4997]: I1205 10:35:39.373441 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/cp-reloader/0.log" Dec 05 10:35:39 crc kubenswrapper[4997]: I1205 10:35:39.416983 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/cp-frr-files/0.log" Dec 05 10:35:39 crc kubenswrapper[4997]: I1205 10:35:39.564322 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/cp-frr-files/0.log" Dec 05 10:35:39 crc kubenswrapper[4997]: I1205 10:35:39.622210 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/cp-metrics/0.log" Dec 05 10:35:39 crc kubenswrapper[4997]: I1205 10:35:39.645230 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/cp-reloader/0.log" Dec 05 10:35:39 crc kubenswrapper[4997]: I1205 10:35:39.714210 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/cp-metrics/0.log" Dec 05 10:35:39 crc kubenswrapper[4997]: I1205 10:35:39.829995 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/cp-frr-files/0.log" Dec 05 10:35:39 crc kubenswrapper[4997]: I1205 10:35:39.872212 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/cp-metrics/0.log" Dec 05 10:35:39 crc kubenswrapper[4997]: I1205 10:35:39.923599 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/cp-reloader/0.log" Dec 05 10:35:39 crc kubenswrapper[4997]: I1205 10:35:39.944772 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/controller/0.log" Dec 05 10:35:40 crc kubenswrapper[4997]: I1205 10:35:40.103705 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/frr-metrics/0.log" Dec 05 10:35:40 crc kubenswrapper[4997]: I1205 10:35:40.122485 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/kube-rbac-proxy/0.log" Dec 05 10:35:40 crc kubenswrapper[4997]: I1205 10:35:40.178261 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/kube-rbac-proxy-frr/0.log" Dec 05 10:35:40 crc kubenswrapper[4997]: I1205 10:35:40.315347 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/reloader/0.log" Dec 05 10:35:40 crc kubenswrapper[4997]: I1205 10:35:40.433017 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-c9zt6_79868cb5-be74-451c-8260-04bca75ded8c/frr-k8s-webhook-server/0.log" Dec 05 10:35:40 crc kubenswrapper[4997]: I1205 10:35:40.676566 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5b6d8f454b-9vl4n_0f68e314-6e6f-40e9-b439-3f91124f9150/manager/0.log" Dec 05 10:35:41 crc kubenswrapper[4997]: I1205 10:35:41.301428 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68c5d7fff9-9gxmm_211e2b11-88f1-4660-9a8b-b43dc7ceb1ec/webhook-server/0.log" Dec 05 10:35:41 crc kubenswrapper[4997]: I1205 10:35:41.415887 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-sdv46_c0d286d5-b144-44ef-9696-21777bfad09b/kube-rbac-proxy/0.log" Dec 05 10:35:42 crc kubenswrapper[4997]: I1205 10:35:42.396750 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-sdv46_c0d286d5-b144-44ef-9696-21777bfad09b/speaker/0.log" Dec 05 10:35:43 crc kubenswrapper[4997]: I1205 10:35:43.448124 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2plrj_24697cfc-2c63-4b44-b4db-88a361476c5f/frr/0.log" Dec 05 10:35:43 crc kubenswrapper[4997]: I1205 10:35:43.750307 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:35:43 crc kubenswrapper[4997]: E1205 10:35:43.750926 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:35:55 crc kubenswrapper[4997]: I1205 10:35:55.700050 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p_5bbd9d2d-29b4-424d-8ad2-46f497377d53/util/0.log" Dec 05 10:35:55 crc kubenswrapper[4997]: I1205 10:35:55.754397 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:35:55 crc kubenswrapper[4997]: E1205 10:35:55.754937 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:35:55 crc kubenswrapper[4997]: I1205 10:35:55.870114 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p_5bbd9d2d-29b4-424d-8ad2-46f497377d53/util/0.log" Dec 05 10:35:55 crc kubenswrapper[4997]: I1205 10:35:55.903644 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p_5bbd9d2d-29b4-424d-8ad2-46f497377d53/pull/0.log" Dec 05 10:35:55 crc kubenswrapper[4997]: I1205 10:35:55.948191 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p_5bbd9d2d-29b4-424d-8ad2-46f497377d53/pull/0.log" Dec 05 10:35:56 crc kubenswrapper[4997]: I1205 10:35:56.088625 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p_5bbd9d2d-29b4-424d-8ad2-46f497377d53/pull/0.log" Dec 05 10:35:56 crc kubenswrapper[4997]: I1205 10:35:56.095585 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p_5bbd9d2d-29b4-424d-8ad2-46f497377d53/util/0.log" Dec 05 10:35:56 crc kubenswrapper[4997]: I1205 10:35:56.137735 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931aqm47p_5bbd9d2d-29b4-424d-8ad2-46f497377d53/extract/0.log" Dec 05 10:35:56 crc kubenswrapper[4997]: I1205 10:35:56.290981 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44_e6b7e3f5-ed40-4170-a421-63547bd2fb65/util/0.log" Dec 05 10:35:56 crc kubenswrapper[4997]: I1205 10:35:56.450101 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44_e6b7e3f5-ed40-4170-a421-63547bd2fb65/util/0.log" Dec 05 10:35:56 crc kubenswrapper[4997]: I1205 10:35:56.495502 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44_e6b7e3f5-ed40-4170-a421-63547bd2fb65/pull/0.log" Dec 05 10:35:56 crc kubenswrapper[4997]: I1205 10:35:56.497701 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44_e6b7e3f5-ed40-4170-a421-63547bd2fb65/pull/0.log" Dec 05 10:35:56 crc kubenswrapper[4997]: I1205 10:35:56.777972 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44_e6b7e3f5-ed40-4170-a421-63547bd2fb65/pull/0.log" Dec 05 10:35:56 crc kubenswrapper[4997]: I1205 10:35:56.781483 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44_e6b7e3f5-ed40-4170-a421-63547bd2fb65/util/0.log" Dec 05 10:35:56 crc kubenswrapper[4997]: I1205 10:35:56.816197 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fmnv44_e6b7e3f5-ed40-4170-a421-63547bd2fb65/extract/0.log" Dec 05 10:35:56 crc kubenswrapper[4997]: I1205 10:35:56.962338 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm_2414755f-2b83-4d3e-a9d2-16147925fc44/util/0.log" Dec 05 10:35:57 crc kubenswrapper[4997]: I1205 10:35:57.135895 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm_2414755f-2b83-4d3e-a9d2-16147925fc44/pull/0.log" Dec 05 10:35:57 crc kubenswrapper[4997]: I1205 10:35:57.146473 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm_2414755f-2b83-4d3e-a9d2-16147925fc44/util/0.log" Dec 05 10:35:57 crc kubenswrapper[4997]: I1205 10:35:57.178794 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm_2414755f-2b83-4d3e-a9d2-16147925fc44/pull/0.log" Dec 05 10:35:57 crc kubenswrapper[4997]: I1205 10:35:57.414087 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm_2414755f-2b83-4d3e-a9d2-16147925fc44/util/0.log" Dec 05 10:35:57 crc kubenswrapper[4997]: I1205 10:35:57.436701 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm_2414755f-2b83-4d3e-a9d2-16147925fc44/extract/0.log" Dec 05 10:35:57 crc kubenswrapper[4997]: I1205 10:35:57.450363 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210bkjvm_2414755f-2b83-4d3e-a9d2-16147925fc44/pull/0.log" Dec 05 10:35:57 crc kubenswrapper[4997]: I1205 10:35:57.600576 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z_e3e947fc-fe79-4011-b548-b9accaefd1b1/util/0.log" Dec 05 10:35:57 crc kubenswrapper[4997]: I1205 10:35:57.806036 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z_e3e947fc-fe79-4011-b548-b9accaefd1b1/pull/0.log" Dec 05 10:35:57 crc kubenswrapper[4997]: I1205 10:35:57.839456 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z_e3e947fc-fe79-4011-b548-b9accaefd1b1/util/0.log" Dec 05 10:35:57 crc kubenswrapper[4997]: I1205 10:35:57.852786 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z_e3e947fc-fe79-4011-b548-b9accaefd1b1/pull/0.log" Dec 05 10:35:58 crc kubenswrapper[4997]: I1205 10:35:58.032772 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z_e3e947fc-fe79-4011-b548-b9accaefd1b1/util/0.log" Dec 05 10:35:58 crc kubenswrapper[4997]: I1205 10:35:58.036088 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z_e3e947fc-fe79-4011-b548-b9accaefd1b1/pull/0.log" Dec 05 10:35:58 crc kubenswrapper[4997]: I1205 10:35:58.057225 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83r648z_e3e947fc-fe79-4011-b548-b9accaefd1b1/extract/0.log" Dec 05 10:35:58 crc kubenswrapper[4997]: I1205 10:35:58.269194 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cspmz_e23b6d26-af92-4d88-926c-a785fbfd7a03/extract-utilities/0.log" Dec 05 10:35:58 crc kubenswrapper[4997]: I1205 10:35:58.455251 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cspmz_e23b6d26-af92-4d88-926c-a785fbfd7a03/extract-content/0.log" Dec 05 10:35:58 crc kubenswrapper[4997]: I1205 10:35:58.457959 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cspmz_e23b6d26-af92-4d88-926c-a785fbfd7a03/extract-utilities/0.log" Dec 05 10:35:58 crc kubenswrapper[4997]: I1205 10:35:58.469294 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cspmz_e23b6d26-af92-4d88-926c-a785fbfd7a03/extract-content/0.log" Dec 05 10:35:58 crc kubenswrapper[4997]: I1205 10:35:58.625754 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cspmz_e23b6d26-af92-4d88-926c-a785fbfd7a03/extract-utilities/0.log" Dec 05 10:35:58 crc kubenswrapper[4997]: I1205 10:35:58.648282 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cspmz_e23b6d26-af92-4d88-926c-a785fbfd7a03/extract-content/0.log" Dec 05 10:35:58 crc kubenswrapper[4997]: I1205 10:35:58.892745 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nn5jp_d228c4a4-4656-4624-af35-94bf3db4c79e/extract-utilities/0.log" Dec 05 10:35:59 crc kubenswrapper[4997]: I1205 10:35:59.109784 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nn5jp_d228c4a4-4656-4624-af35-94bf3db4c79e/extract-content/0.log" Dec 05 10:35:59 crc kubenswrapper[4997]: I1205 10:35:59.155680 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nn5jp_d228c4a4-4656-4624-af35-94bf3db4c79e/extract-utilities/0.log" Dec 05 10:35:59 crc kubenswrapper[4997]: I1205 10:35:59.225439 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nn5jp_d228c4a4-4656-4624-af35-94bf3db4c79e/extract-content/0.log" Dec 05 10:35:59 crc kubenswrapper[4997]: I1205 10:35:59.565201 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cspmz_e23b6d26-af92-4d88-926c-a785fbfd7a03/registry-server/0.log" Dec 05 10:35:59 crc kubenswrapper[4997]: I1205 10:35:59.665056 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nn5jp_d228c4a4-4656-4624-af35-94bf3db4c79e/extract-content/0.log" Dec 05 10:35:59 crc kubenswrapper[4997]: I1205 10:35:59.680833 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nn5jp_d228c4a4-4656-4624-af35-94bf3db4c79e/extract-utilities/0.log" Dec 05 10:35:59 crc kubenswrapper[4997]: I1205 10:35:59.884487 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-k4n89_14ea843e-5c09-4461-88b4-4810e6d28cca/marketplace-operator/0.log" Dec 05 10:35:59 crc kubenswrapper[4997]: I1205 10:35:59.984591 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c46hv_0b51e69f-d870-494a-adc2-b912ec12e2e6/extract-utilities/0.log" Dec 05 10:36:00 crc kubenswrapper[4997]: I1205 10:36:00.144301 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c46hv_0b51e69f-d870-494a-adc2-b912ec12e2e6/extract-utilities/0.log" Dec 05 10:36:00 crc kubenswrapper[4997]: I1205 10:36:00.189497 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c46hv_0b51e69f-d870-494a-adc2-b912ec12e2e6/extract-content/0.log" Dec 05 10:36:00 crc kubenswrapper[4997]: I1205 10:36:00.248384 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c46hv_0b51e69f-d870-494a-adc2-b912ec12e2e6/extract-content/0.log" Dec 05 10:36:00 crc kubenswrapper[4997]: I1205 10:36:00.410910 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c46hv_0b51e69f-d870-494a-adc2-b912ec12e2e6/extract-utilities/0.log" Dec 05 10:36:00 crc kubenswrapper[4997]: I1205 10:36:00.419517 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c46hv_0b51e69f-d870-494a-adc2-b912ec12e2e6/extract-content/0.log" Dec 05 10:36:00 crc kubenswrapper[4997]: I1205 10:36:00.582937 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nn5jp_d228c4a4-4656-4624-af35-94bf3db4c79e/registry-server/0.log" Dec 05 10:36:00 crc kubenswrapper[4997]: I1205 10:36:00.656326 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ktvff_e5f08766-f19b-4b4f-b1f8-f25b4214457e/extract-utilities/0.log" Dec 05 10:36:00 crc kubenswrapper[4997]: I1205 10:36:00.869166 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ktvff_e5f08766-f19b-4b4f-b1f8-f25b4214457e/extract-utilities/0.log" Dec 05 10:36:00 crc kubenswrapper[4997]: I1205 10:36:00.932687 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ktvff_e5f08766-f19b-4b4f-b1f8-f25b4214457e/extract-content/0.log" Dec 05 10:36:00 crc kubenswrapper[4997]: I1205 10:36:00.954775 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ktvff_e5f08766-f19b-4b4f-b1f8-f25b4214457e/extract-content/0.log" Dec 05 10:36:01 crc kubenswrapper[4997]: I1205 10:36:01.080445 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c46hv_0b51e69f-d870-494a-adc2-b912ec12e2e6/registry-server/0.log" Dec 05 10:36:01 crc kubenswrapper[4997]: I1205 10:36:01.169571 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ktvff_e5f08766-f19b-4b4f-b1f8-f25b4214457e/extract-content/0.log" Dec 05 10:36:01 crc kubenswrapper[4997]: I1205 10:36:01.178227 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ktvff_e5f08766-f19b-4b4f-b1f8-f25b4214457e/extract-utilities/0.log" Dec 05 10:36:02 crc kubenswrapper[4997]: I1205 10:36:02.872623 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ktvff_e5f08766-f19b-4b4f-b1f8-f25b4214457e/registry-server/0.log" Dec 05 10:36:07 crc kubenswrapper[4997]: I1205 10:36:07.749274 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:36:07 crc kubenswrapper[4997]: E1205 10:36:07.750229 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:36:15 crc kubenswrapper[4997]: I1205 10:36:15.377439 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-4586l_68e3de45-5599-48c3-8b35-73cc4d02250a/prometheus-operator/0.log" Dec 05 10:36:15 crc kubenswrapper[4997]: I1205 10:36:15.602921 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-58cc577445-gtbr9_77d7110e-dc31-46fe-90d0-a1344dea977e/prometheus-operator-admission-webhook/0.log" Dec 05 10:36:15 crc kubenswrapper[4997]: I1205 10:36:15.695416 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-58cc577445-qz4cf_c4adb5f1-1596-4833-92e8-8fddef0bdd40/prometheus-operator-admission-webhook/0.log" Dec 05 10:36:15 crc kubenswrapper[4997]: I1205 10:36:15.819696 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-xzrfl_da0ed268-f4eb-4a9f-8ecf-bd7d318e0f09/operator/0.log" Dec 05 10:36:15 crc kubenswrapper[4997]: I1205 10:36:15.879373 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-rb78r_9f2abe7e-5864-476a-b5b8-f46b7d7d7350/perses-operator/0.log" Dec 05 10:36:19 crc kubenswrapper[4997]: I1205 10:36:19.759343 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:36:19 crc kubenswrapper[4997]: E1205 10:36:19.760173 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:36:33 crc kubenswrapper[4997]: I1205 10:36:33.748905 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:36:33 crc kubenswrapper[4997]: E1205 10:36:33.750667 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:36:38 crc kubenswrapper[4997]: E1205 10:36:38.316494 4997 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.155:56062->38.102.83.155:45077: write tcp 38.102.83.155:56062->38.102.83.155:45077: write: connection reset by peer Dec 05 10:36:45 crc kubenswrapper[4997]: I1205 10:36:45.749434 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:36:45 crc kubenswrapper[4997]: E1205 10:36:45.750420 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:36:57 crc kubenswrapper[4997]: I1205 10:36:57.749324 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:36:57 crc kubenswrapper[4997]: E1205 10:36:57.750128 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:37:12 crc kubenswrapper[4997]: I1205 10:37:12.749099 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:37:12 crc kubenswrapper[4997]: E1205 10:37:12.749851 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:37:26 crc kubenswrapper[4997]: I1205 10:37:26.749670 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:37:26 crc kubenswrapper[4997]: E1205 10:37:26.750564 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:37:40 crc kubenswrapper[4997]: I1205 10:37:40.748992 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:37:40 crc kubenswrapper[4997]: E1205 10:37:40.749777 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:37:53 crc kubenswrapper[4997]: I1205 10:37:53.749652 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:37:53 crc kubenswrapper[4997]: E1205 10:37:53.750942 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:38:04 crc kubenswrapper[4997]: I1205 10:38:04.749606 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:38:04 crc kubenswrapper[4997]: E1205 10:38:04.750500 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:38:16 crc kubenswrapper[4997]: I1205 10:38:16.751119 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:38:16 crc kubenswrapper[4997]: E1205 10:38:16.752224 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:38:27 crc kubenswrapper[4997]: I1205 10:38:27.749292 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:38:27 crc kubenswrapper[4997]: E1205 10:38:27.750100 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:38:42 crc kubenswrapper[4997]: I1205 10:38:42.749131 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:38:42 crc kubenswrapper[4997]: E1205 10:38:42.750030 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:38:55 crc kubenswrapper[4997]: I1205 10:38:55.750046 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:38:55 crc kubenswrapper[4997]: E1205 10:38:55.751010 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:39:08 crc kubenswrapper[4997]: I1205 10:39:08.750348 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:39:08 crc kubenswrapper[4997]: E1205 10:39:08.751239 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:39:21 crc kubenswrapper[4997]: I1205 10:39:21.750158 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:39:21 crc kubenswrapper[4997]: E1205 10:39:21.751151 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:39:36 crc kubenswrapper[4997]: I1205 10:39:36.749318 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:39:36 crc kubenswrapper[4997]: E1205 10:39:36.750137 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:39:48 crc kubenswrapper[4997]: I1205 10:39:48.748995 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:39:48 crc kubenswrapper[4997]: E1205 10:39:48.749912 4997 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l7hsk_openshift-machine-config-operator(28cde6da-1de5-455a-93c8-dd250f9ea4d0)\"" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" Dec 05 10:40:02 crc kubenswrapper[4997]: I1205 10:40:02.748365 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:40:03 crc kubenswrapper[4997]: I1205 10:40:03.701267 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"5fac2ba0fe3f622bea15d0ab9b645e028b7d2ad7b9379d559a0990ccb6508233"} Dec 05 10:40:18 crc kubenswrapper[4997]: I1205 10:40:18.865103 4997 generic.go:334] "Generic (PLEG): container finished" podID="b7478a1e-4ad6-421a-ab0f-9795d95be179" containerID="2050dfdcf13f3aec023eff86ac4567094633839815a0aae97102a7c2f8a21b52" exitCode=0 Dec 05 10:40:18 crc kubenswrapper[4997]: I1205 10:40:18.865668 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w7q74/must-gather-mnk5f" event={"ID":"b7478a1e-4ad6-421a-ab0f-9795d95be179","Type":"ContainerDied","Data":"2050dfdcf13f3aec023eff86ac4567094633839815a0aae97102a7c2f8a21b52"} Dec 05 10:40:18 crc kubenswrapper[4997]: I1205 10:40:18.866312 4997 scope.go:117] "RemoveContainer" containerID="2050dfdcf13f3aec023eff86ac4567094633839815a0aae97102a7c2f8a21b52" Dec 05 10:40:19 crc kubenswrapper[4997]: I1205 10:40:19.610767 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-w7q74_must-gather-mnk5f_b7478a1e-4ad6-421a-ab0f-9795d95be179/gather/0.log" Dec 05 10:40:31 crc kubenswrapper[4997]: I1205 10:40:31.831805 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-w7q74/must-gather-mnk5f"] Dec 05 10:40:31 crc kubenswrapper[4997]: I1205 10:40:31.833836 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-w7q74/must-gather-mnk5f" podUID="b7478a1e-4ad6-421a-ab0f-9795d95be179" containerName="copy" containerID="cri-o://6513b33bd9eca34b0e643050c3d9e2cea0cca4ce48f4f79879ec30f2ab93f68f" gracePeriod=2 Dec 05 10:40:31 crc kubenswrapper[4997]: I1205 10:40:31.845572 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-w7q74/must-gather-mnk5f"] Dec 05 10:40:32 crc kubenswrapper[4997]: I1205 10:40:32.022371 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-w7q74_must-gather-mnk5f_b7478a1e-4ad6-421a-ab0f-9795d95be179/copy/0.log" Dec 05 10:40:32 crc kubenswrapper[4997]: I1205 10:40:32.022801 4997 generic.go:334] "Generic (PLEG): container finished" podID="b7478a1e-4ad6-421a-ab0f-9795d95be179" containerID="6513b33bd9eca34b0e643050c3d9e2cea0cca4ce48f4f79879ec30f2ab93f68f" exitCode=143 Dec 05 10:40:32 crc kubenswrapper[4997]: I1205 10:40:32.354154 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-w7q74_must-gather-mnk5f_b7478a1e-4ad6-421a-ab0f-9795d95be179/copy/0.log" Dec 05 10:40:32 crc kubenswrapper[4997]: I1205 10:40:32.355026 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w7q74/must-gather-mnk5f" Dec 05 10:40:32 crc kubenswrapper[4997]: I1205 10:40:32.507514 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjc4d\" (UniqueName: \"kubernetes.io/projected/b7478a1e-4ad6-421a-ab0f-9795d95be179-kube-api-access-kjc4d\") pod \"b7478a1e-4ad6-421a-ab0f-9795d95be179\" (UID: \"b7478a1e-4ad6-421a-ab0f-9795d95be179\") " Dec 05 10:40:32 crc kubenswrapper[4997]: I1205 10:40:32.508493 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b7478a1e-4ad6-421a-ab0f-9795d95be179-must-gather-output\") pod \"b7478a1e-4ad6-421a-ab0f-9795d95be179\" (UID: \"b7478a1e-4ad6-421a-ab0f-9795d95be179\") " Dec 05 10:40:32 crc kubenswrapper[4997]: I1205 10:40:32.515638 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7478a1e-4ad6-421a-ab0f-9795d95be179-kube-api-access-kjc4d" (OuterVolumeSpecName: "kube-api-access-kjc4d") pod "b7478a1e-4ad6-421a-ab0f-9795d95be179" (UID: "b7478a1e-4ad6-421a-ab0f-9795d95be179"). InnerVolumeSpecName "kube-api-access-kjc4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:40:32 crc kubenswrapper[4997]: I1205 10:40:32.611515 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjc4d\" (UniqueName: \"kubernetes.io/projected/b7478a1e-4ad6-421a-ab0f-9795d95be179-kube-api-access-kjc4d\") on node \"crc\" DevicePath \"\"" Dec 05 10:40:32 crc kubenswrapper[4997]: I1205 10:40:32.787163 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7478a1e-4ad6-421a-ab0f-9795d95be179-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b7478a1e-4ad6-421a-ab0f-9795d95be179" (UID: "b7478a1e-4ad6-421a-ab0f-9795d95be179"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:40:32 crc kubenswrapper[4997]: I1205 10:40:32.816088 4997 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b7478a1e-4ad6-421a-ab0f-9795d95be179-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 05 10:40:33 crc kubenswrapper[4997]: I1205 10:40:33.033544 4997 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-w7q74_must-gather-mnk5f_b7478a1e-4ad6-421a-ab0f-9795d95be179/copy/0.log" Dec 05 10:40:33 crc kubenswrapper[4997]: I1205 10:40:33.034641 4997 scope.go:117] "RemoveContainer" containerID="6513b33bd9eca34b0e643050c3d9e2cea0cca4ce48f4f79879ec30f2ab93f68f" Dec 05 10:40:33 crc kubenswrapper[4997]: I1205 10:40:33.034672 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w7q74/must-gather-mnk5f" Dec 05 10:40:33 crc kubenswrapper[4997]: I1205 10:40:33.058201 4997 scope.go:117] "RemoveContainer" containerID="2050dfdcf13f3aec023eff86ac4567094633839815a0aae97102a7c2f8a21b52" Dec 05 10:40:33 crc kubenswrapper[4997]: I1205 10:40:33.764480 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7478a1e-4ad6-421a-ab0f-9795d95be179" path="/var/lib/kubelet/pods/b7478a1e-4ad6-421a-ab0f-9795d95be179/volumes" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.001293 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kgpn2"] Dec 05 10:41:34 crc kubenswrapper[4997]: E1205 10:41:34.003941 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0c8978e-08e9-4c27-8f13-d83b35a02ca7" containerName="registry-server" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.004168 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0c8978e-08e9-4c27-8f13-d83b35a02ca7" containerName="registry-server" Dec 05 10:41:34 crc kubenswrapper[4997]: E1205 10:41:34.004265 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7478a1e-4ad6-421a-ab0f-9795d95be179" containerName="copy" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.004341 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7478a1e-4ad6-421a-ab0f-9795d95be179" containerName="copy" Dec 05 10:41:34 crc kubenswrapper[4997]: E1205 10:41:34.004432 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7478a1e-4ad6-421a-ab0f-9795d95be179" containerName="gather" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.004504 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7478a1e-4ad6-421a-ab0f-9795d95be179" containerName="gather" Dec 05 10:41:34 crc kubenswrapper[4997]: E1205 10:41:34.004582 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0c8978e-08e9-4c27-8f13-d83b35a02ca7" containerName="extract-content" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.004715 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0c8978e-08e9-4c27-8f13-d83b35a02ca7" containerName="extract-content" Dec 05 10:41:34 crc kubenswrapper[4997]: E1205 10:41:34.004839 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0c8978e-08e9-4c27-8f13-d83b35a02ca7" containerName="extract-utilities" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.004929 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0c8978e-08e9-4c27-8f13-d83b35a02ca7" containerName="extract-utilities" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.005322 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7478a1e-4ad6-421a-ab0f-9795d95be179" containerName="gather" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.005420 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7478a1e-4ad6-421a-ab0f-9795d95be179" containerName="copy" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.005514 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0c8978e-08e9-4c27-8f13-d83b35a02ca7" containerName="registry-server" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.007859 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kgpn2" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.017100 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kgpn2"] Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.095294 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsrpq\" (UniqueName: \"kubernetes.io/projected/a1355762-5f72-42ea-b6bf-9f6a16a9529e-kube-api-access-qsrpq\") pod \"certified-operators-kgpn2\" (UID: \"a1355762-5f72-42ea-b6bf-9f6a16a9529e\") " pod="openshift-marketplace/certified-operators-kgpn2" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.095339 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1355762-5f72-42ea-b6bf-9f6a16a9529e-catalog-content\") pod \"certified-operators-kgpn2\" (UID: \"a1355762-5f72-42ea-b6bf-9f6a16a9529e\") " pod="openshift-marketplace/certified-operators-kgpn2" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.095437 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1355762-5f72-42ea-b6bf-9f6a16a9529e-utilities\") pod \"certified-operators-kgpn2\" (UID: \"a1355762-5f72-42ea-b6bf-9f6a16a9529e\") " pod="openshift-marketplace/certified-operators-kgpn2" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.196978 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsrpq\" (UniqueName: \"kubernetes.io/projected/a1355762-5f72-42ea-b6bf-9f6a16a9529e-kube-api-access-qsrpq\") pod \"certified-operators-kgpn2\" (UID: \"a1355762-5f72-42ea-b6bf-9f6a16a9529e\") " pod="openshift-marketplace/certified-operators-kgpn2" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.197025 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1355762-5f72-42ea-b6bf-9f6a16a9529e-catalog-content\") pod \"certified-operators-kgpn2\" (UID: \"a1355762-5f72-42ea-b6bf-9f6a16a9529e\") " pod="openshift-marketplace/certified-operators-kgpn2" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.197070 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1355762-5f72-42ea-b6bf-9f6a16a9529e-utilities\") pod \"certified-operators-kgpn2\" (UID: \"a1355762-5f72-42ea-b6bf-9f6a16a9529e\") " pod="openshift-marketplace/certified-operators-kgpn2" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.197765 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1355762-5f72-42ea-b6bf-9f6a16a9529e-catalog-content\") pod \"certified-operators-kgpn2\" (UID: \"a1355762-5f72-42ea-b6bf-9f6a16a9529e\") " pod="openshift-marketplace/certified-operators-kgpn2" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.197821 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1355762-5f72-42ea-b6bf-9f6a16a9529e-utilities\") pod \"certified-operators-kgpn2\" (UID: \"a1355762-5f72-42ea-b6bf-9f6a16a9529e\") " pod="openshift-marketplace/certified-operators-kgpn2" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.218801 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsrpq\" (UniqueName: \"kubernetes.io/projected/a1355762-5f72-42ea-b6bf-9f6a16a9529e-kube-api-access-qsrpq\") pod \"certified-operators-kgpn2\" (UID: \"a1355762-5f72-42ea-b6bf-9f6a16a9529e\") " pod="openshift-marketplace/certified-operators-kgpn2" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.339531 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kgpn2" Dec 05 10:41:34 crc kubenswrapper[4997]: I1205 10:41:34.918220 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kgpn2"] Dec 05 10:41:35 crc kubenswrapper[4997]: I1205 10:41:35.670349 4997 generic.go:334] "Generic (PLEG): container finished" podID="a1355762-5f72-42ea-b6bf-9f6a16a9529e" containerID="857d0ed9d6ec8ef8ec280f5a6a108e8a03e30e374c5fde00a0d958d810ad2a0d" exitCode=0 Dec 05 10:41:35 crc kubenswrapper[4997]: I1205 10:41:35.670418 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgpn2" event={"ID":"a1355762-5f72-42ea-b6bf-9f6a16a9529e","Type":"ContainerDied","Data":"857d0ed9d6ec8ef8ec280f5a6a108e8a03e30e374c5fde00a0d958d810ad2a0d"} Dec 05 10:41:35 crc kubenswrapper[4997]: I1205 10:41:35.670654 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgpn2" event={"ID":"a1355762-5f72-42ea-b6bf-9f6a16a9529e","Type":"ContainerStarted","Data":"668d41758ad62f551ee91d0bb0d1e072d376bd5b61bfd5c8512a3175f4da6eb2"} Dec 05 10:41:35 crc kubenswrapper[4997]: I1205 10:41:35.674123 4997 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:41:36 crc kubenswrapper[4997]: I1205 10:41:36.683900 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgpn2" event={"ID":"a1355762-5f72-42ea-b6bf-9f6a16a9529e","Type":"ContainerStarted","Data":"235c139540623ff1f18c38020f4b078d7edfed380b86fdd44efdea1e27752986"} Dec 05 10:41:37 crc kubenswrapper[4997]: I1205 10:41:37.694807 4997 generic.go:334] "Generic (PLEG): container finished" podID="a1355762-5f72-42ea-b6bf-9f6a16a9529e" containerID="235c139540623ff1f18c38020f4b078d7edfed380b86fdd44efdea1e27752986" exitCode=0 Dec 05 10:41:37 crc kubenswrapper[4997]: I1205 10:41:37.694871 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgpn2" event={"ID":"a1355762-5f72-42ea-b6bf-9f6a16a9529e","Type":"ContainerDied","Data":"235c139540623ff1f18c38020f4b078d7edfed380b86fdd44efdea1e27752986"} Dec 05 10:41:38 crc kubenswrapper[4997]: I1205 10:41:38.708438 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgpn2" event={"ID":"a1355762-5f72-42ea-b6bf-9f6a16a9529e","Type":"ContainerStarted","Data":"358d47935e59d625e5a1f56c441af4948b6a9a5d38bed426a2b6a5d9f2f870a0"} Dec 05 10:41:38 crc kubenswrapper[4997]: I1205 10:41:38.737543 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kgpn2" podStartSLOduration=3.2984889969999998 podStartE2EDuration="5.737517975s" podCreationTimestamp="2025-12-05 10:41:33 +0000 UTC" firstStartedPulling="2025-12-05 10:41:35.67384267 +0000 UTC m=+13596.202749941" lastFinishedPulling="2025-12-05 10:41:38.112871668 +0000 UTC m=+13598.641778919" observedRunningTime="2025-12-05 10:41:38.727294749 +0000 UTC m=+13599.256202020" watchObservedRunningTime="2025-12-05 10:41:38.737517975 +0000 UTC m=+13599.266425236" Dec 05 10:41:44 crc kubenswrapper[4997]: I1205 10:41:44.340516 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kgpn2" Dec 05 10:41:44 crc kubenswrapper[4997]: I1205 10:41:44.341098 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kgpn2" Dec 05 10:41:44 crc kubenswrapper[4997]: I1205 10:41:44.397608 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kgpn2" Dec 05 10:41:44 crc kubenswrapper[4997]: I1205 10:41:44.819048 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kgpn2" Dec 05 10:41:44 crc kubenswrapper[4997]: I1205 10:41:44.865241 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kgpn2"] Dec 05 10:41:46 crc kubenswrapper[4997]: I1205 10:41:46.792425 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kgpn2" podUID="a1355762-5f72-42ea-b6bf-9f6a16a9529e" containerName="registry-server" containerID="cri-o://358d47935e59d625e5a1f56c441af4948b6a9a5d38bed426a2b6a5d9f2f870a0" gracePeriod=2 Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.805651 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kgpn2" Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.805822 4997 generic.go:334] "Generic (PLEG): container finished" podID="a1355762-5f72-42ea-b6bf-9f6a16a9529e" containerID="358d47935e59d625e5a1f56c441af4948b6a9a5d38bed426a2b6a5d9f2f870a0" exitCode=0 Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.805850 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgpn2" event={"ID":"a1355762-5f72-42ea-b6bf-9f6a16a9529e","Type":"ContainerDied","Data":"358d47935e59d625e5a1f56c441af4948b6a9a5d38bed426a2b6a5d9f2f870a0"} Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.806573 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgpn2" event={"ID":"a1355762-5f72-42ea-b6bf-9f6a16a9529e","Type":"ContainerDied","Data":"668d41758ad62f551ee91d0bb0d1e072d376bd5b61bfd5c8512a3175f4da6eb2"} Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.806601 4997 scope.go:117] "RemoveContainer" containerID="358d47935e59d625e5a1f56c441af4948b6a9a5d38bed426a2b6a5d9f2f870a0" Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.827861 4997 scope.go:117] "RemoveContainer" containerID="235c139540623ff1f18c38020f4b078d7edfed380b86fdd44efdea1e27752986" Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.881973 4997 scope.go:117] "RemoveContainer" containerID="857d0ed9d6ec8ef8ec280f5a6a108e8a03e30e374c5fde00a0d958d810ad2a0d" Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.912819 4997 scope.go:117] "RemoveContainer" containerID="358d47935e59d625e5a1f56c441af4948b6a9a5d38bed426a2b6a5d9f2f870a0" Dec 05 10:41:47 crc kubenswrapper[4997]: E1205 10:41:47.913941 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"358d47935e59d625e5a1f56c441af4948b6a9a5d38bed426a2b6a5d9f2f870a0\": container with ID starting with 358d47935e59d625e5a1f56c441af4948b6a9a5d38bed426a2b6a5d9f2f870a0 not found: ID does not exist" containerID="358d47935e59d625e5a1f56c441af4948b6a9a5d38bed426a2b6a5d9f2f870a0" Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.914001 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"358d47935e59d625e5a1f56c441af4948b6a9a5d38bed426a2b6a5d9f2f870a0"} err="failed to get container status \"358d47935e59d625e5a1f56c441af4948b6a9a5d38bed426a2b6a5d9f2f870a0\": rpc error: code = NotFound desc = could not find container \"358d47935e59d625e5a1f56c441af4948b6a9a5d38bed426a2b6a5d9f2f870a0\": container with ID starting with 358d47935e59d625e5a1f56c441af4948b6a9a5d38bed426a2b6a5d9f2f870a0 not found: ID does not exist" Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.914037 4997 scope.go:117] "RemoveContainer" containerID="235c139540623ff1f18c38020f4b078d7edfed380b86fdd44efdea1e27752986" Dec 05 10:41:47 crc kubenswrapper[4997]: E1205 10:41:47.915662 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"235c139540623ff1f18c38020f4b078d7edfed380b86fdd44efdea1e27752986\": container with ID starting with 235c139540623ff1f18c38020f4b078d7edfed380b86fdd44efdea1e27752986 not found: ID does not exist" containerID="235c139540623ff1f18c38020f4b078d7edfed380b86fdd44efdea1e27752986" Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.915812 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"235c139540623ff1f18c38020f4b078d7edfed380b86fdd44efdea1e27752986"} err="failed to get container status \"235c139540623ff1f18c38020f4b078d7edfed380b86fdd44efdea1e27752986\": rpc error: code = NotFound desc = could not find container \"235c139540623ff1f18c38020f4b078d7edfed380b86fdd44efdea1e27752986\": container with ID starting with 235c139540623ff1f18c38020f4b078d7edfed380b86fdd44efdea1e27752986 not found: ID does not exist" Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.915919 4997 scope.go:117] "RemoveContainer" containerID="857d0ed9d6ec8ef8ec280f5a6a108e8a03e30e374c5fde00a0d958d810ad2a0d" Dec 05 10:41:47 crc kubenswrapper[4997]: E1205 10:41:47.916534 4997 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"857d0ed9d6ec8ef8ec280f5a6a108e8a03e30e374c5fde00a0d958d810ad2a0d\": container with ID starting with 857d0ed9d6ec8ef8ec280f5a6a108e8a03e30e374c5fde00a0d958d810ad2a0d not found: ID does not exist" containerID="857d0ed9d6ec8ef8ec280f5a6a108e8a03e30e374c5fde00a0d958d810ad2a0d" Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.916567 4997 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"857d0ed9d6ec8ef8ec280f5a6a108e8a03e30e374c5fde00a0d958d810ad2a0d"} err="failed to get container status \"857d0ed9d6ec8ef8ec280f5a6a108e8a03e30e374c5fde00a0d958d810ad2a0d\": rpc error: code = NotFound desc = could not find container \"857d0ed9d6ec8ef8ec280f5a6a108e8a03e30e374c5fde00a0d958d810ad2a0d\": container with ID starting with 857d0ed9d6ec8ef8ec280f5a6a108e8a03e30e374c5fde00a0d958d810ad2a0d not found: ID does not exist" Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.919356 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1355762-5f72-42ea-b6bf-9f6a16a9529e-utilities\") pod \"a1355762-5f72-42ea-b6bf-9f6a16a9529e\" (UID: \"a1355762-5f72-42ea-b6bf-9f6a16a9529e\") " Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.919596 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsrpq\" (UniqueName: \"kubernetes.io/projected/a1355762-5f72-42ea-b6bf-9f6a16a9529e-kube-api-access-qsrpq\") pod \"a1355762-5f72-42ea-b6bf-9f6a16a9529e\" (UID: \"a1355762-5f72-42ea-b6bf-9f6a16a9529e\") " Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.919754 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1355762-5f72-42ea-b6bf-9f6a16a9529e-catalog-content\") pod \"a1355762-5f72-42ea-b6bf-9f6a16a9529e\" (UID: \"a1355762-5f72-42ea-b6bf-9f6a16a9529e\") " Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.920790 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1355762-5f72-42ea-b6bf-9f6a16a9529e-utilities" (OuterVolumeSpecName: "utilities") pod "a1355762-5f72-42ea-b6bf-9f6a16a9529e" (UID: "a1355762-5f72-42ea-b6bf-9f6a16a9529e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.925972 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1355762-5f72-42ea-b6bf-9f6a16a9529e-kube-api-access-qsrpq" (OuterVolumeSpecName: "kube-api-access-qsrpq") pod "a1355762-5f72-42ea-b6bf-9f6a16a9529e" (UID: "a1355762-5f72-42ea-b6bf-9f6a16a9529e"). InnerVolumeSpecName "kube-api-access-qsrpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:41:47 crc kubenswrapper[4997]: I1205 10:41:47.979942 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1355762-5f72-42ea-b6bf-9f6a16a9529e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a1355762-5f72-42ea-b6bf-9f6a16a9529e" (UID: "a1355762-5f72-42ea-b6bf-9f6a16a9529e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:41:48 crc kubenswrapper[4997]: I1205 10:41:48.022163 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsrpq\" (UniqueName: \"kubernetes.io/projected/a1355762-5f72-42ea-b6bf-9f6a16a9529e-kube-api-access-qsrpq\") on node \"crc\" DevicePath \"\"" Dec 05 10:41:48 crc kubenswrapper[4997]: I1205 10:41:48.022203 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1355762-5f72-42ea-b6bf-9f6a16a9529e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:41:48 crc kubenswrapper[4997]: I1205 10:41:48.022212 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1355762-5f72-42ea-b6bf-9f6a16a9529e-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:41:48 crc kubenswrapper[4997]: I1205 10:41:48.815963 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kgpn2" Dec 05 10:41:48 crc kubenswrapper[4997]: I1205 10:41:48.854773 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kgpn2"] Dec 05 10:41:48 crc kubenswrapper[4997]: I1205 10:41:48.866229 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kgpn2"] Dec 05 10:41:49 crc kubenswrapper[4997]: I1205 10:41:49.767029 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1355762-5f72-42ea-b6bf-9f6a16a9529e" path="/var/lib/kubelet/pods/a1355762-5f72-42ea-b6bf-9f6a16a9529e/volumes" Dec 05 10:42:19 crc kubenswrapper[4997]: I1205 10:42:19.770058 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:42:19 crc kubenswrapper[4997]: I1205 10:42:19.770743 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:42:49 crc kubenswrapper[4997]: I1205 10:42:49.773718 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:42:49 crc kubenswrapper[4997]: I1205 10:42:49.774417 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:43:19 crc kubenswrapper[4997]: I1205 10:43:19.769940 4997 patch_prober.go:28] interesting pod/machine-config-daemon-l7hsk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:43:19 crc kubenswrapper[4997]: I1205 10:43:19.770451 4997 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:43:19 crc kubenswrapper[4997]: I1205 10:43:19.770508 4997 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" Dec 05 10:43:19 crc kubenswrapper[4997]: I1205 10:43:19.770990 4997 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5fac2ba0fe3f622bea15d0ab9b645e028b7d2ad7b9379d559a0990ccb6508233"} pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 10:43:19 crc kubenswrapper[4997]: I1205 10:43:19.771045 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" podUID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerName="machine-config-daemon" containerID="cri-o://5fac2ba0fe3f622bea15d0ab9b645e028b7d2ad7b9379d559a0990ccb6508233" gracePeriod=600 Dec 05 10:43:20 crc kubenswrapper[4997]: I1205 10:43:20.837722 4997 generic.go:334] "Generic (PLEG): container finished" podID="28cde6da-1de5-455a-93c8-dd250f9ea4d0" containerID="5fac2ba0fe3f622bea15d0ab9b645e028b7d2ad7b9379d559a0990ccb6508233" exitCode=0 Dec 05 10:43:20 crc kubenswrapper[4997]: I1205 10:43:20.837822 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerDied","Data":"5fac2ba0fe3f622bea15d0ab9b645e028b7d2ad7b9379d559a0990ccb6508233"} Dec 05 10:43:20 crc kubenswrapper[4997]: I1205 10:43:20.838194 4997 scope.go:117] "RemoveContainer" containerID="804c682b5f0e1a94fa58421930c8a864fa8b8cbd490292031a2b9a787c9ef648" Dec 05 10:43:21 crc kubenswrapper[4997]: I1205 10:43:21.851585 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l7hsk" event={"ID":"28cde6da-1de5-455a-93c8-dd250f9ea4d0","Type":"ContainerStarted","Data":"261bfb394c6b27633f87b47a76e64b21396c457110f929539b23f5cea636bde8"} Dec 05 10:43:42 crc kubenswrapper[4997]: I1205 10:43:42.820116 4997 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zkwc2"] Dec 05 10:43:42 crc kubenswrapper[4997]: E1205 10:43:42.821275 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1355762-5f72-42ea-b6bf-9f6a16a9529e" containerName="extract-utilities" Dec 05 10:43:42 crc kubenswrapper[4997]: I1205 10:43:42.821301 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1355762-5f72-42ea-b6bf-9f6a16a9529e" containerName="extract-utilities" Dec 05 10:43:42 crc kubenswrapper[4997]: E1205 10:43:42.821316 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1355762-5f72-42ea-b6bf-9f6a16a9529e" containerName="extract-content" Dec 05 10:43:42 crc kubenswrapper[4997]: I1205 10:43:42.821323 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1355762-5f72-42ea-b6bf-9f6a16a9529e" containerName="extract-content" Dec 05 10:43:42 crc kubenswrapper[4997]: E1205 10:43:42.821364 4997 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1355762-5f72-42ea-b6bf-9f6a16a9529e" containerName="registry-server" Dec 05 10:43:42 crc kubenswrapper[4997]: I1205 10:43:42.821372 4997 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1355762-5f72-42ea-b6bf-9f6a16a9529e" containerName="registry-server" Dec 05 10:43:42 crc kubenswrapper[4997]: I1205 10:43:42.821642 4997 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1355762-5f72-42ea-b6bf-9f6a16a9529e" containerName="registry-server" Dec 05 10:43:42 crc kubenswrapper[4997]: I1205 10:43:42.823724 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zkwc2" Dec 05 10:43:42 crc kubenswrapper[4997]: I1205 10:43:42.835783 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zkwc2"] Dec 05 10:43:42 crc kubenswrapper[4997]: I1205 10:43:42.940710 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ab2d982-1c12-4585-bea3-83c246b00ed9-catalog-content\") pod \"redhat-marketplace-zkwc2\" (UID: \"4ab2d982-1c12-4585-bea3-83c246b00ed9\") " pod="openshift-marketplace/redhat-marketplace-zkwc2" Dec 05 10:43:42 crc kubenswrapper[4997]: I1205 10:43:42.940900 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzxg7\" (UniqueName: \"kubernetes.io/projected/4ab2d982-1c12-4585-bea3-83c246b00ed9-kube-api-access-dzxg7\") pod \"redhat-marketplace-zkwc2\" (UID: \"4ab2d982-1c12-4585-bea3-83c246b00ed9\") " pod="openshift-marketplace/redhat-marketplace-zkwc2" Dec 05 10:43:42 crc kubenswrapper[4997]: I1205 10:43:42.941038 4997 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ab2d982-1c12-4585-bea3-83c246b00ed9-utilities\") pod \"redhat-marketplace-zkwc2\" (UID: \"4ab2d982-1c12-4585-bea3-83c246b00ed9\") " pod="openshift-marketplace/redhat-marketplace-zkwc2" Dec 05 10:43:43 crc kubenswrapper[4997]: I1205 10:43:43.043307 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ab2d982-1c12-4585-bea3-83c246b00ed9-utilities\") pod \"redhat-marketplace-zkwc2\" (UID: \"4ab2d982-1c12-4585-bea3-83c246b00ed9\") " pod="openshift-marketplace/redhat-marketplace-zkwc2" Dec 05 10:43:43 crc kubenswrapper[4997]: I1205 10:43:43.043449 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ab2d982-1c12-4585-bea3-83c246b00ed9-catalog-content\") pod \"redhat-marketplace-zkwc2\" (UID: \"4ab2d982-1c12-4585-bea3-83c246b00ed9\") " pod="openshift-marketplace/redhat-marketplace-zkwc2" Dec 05 10:43:43 crc kubenswrapper[4997]: I1205 10:43:43.043489 4997 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzxg7\" (UniqueName: \"kubernetes.io/projected/4ab2d982-1c12-4585-bea3-83c246b00ed9-kube-api-access-dzxg7\") pod \"redhat-marketplace-zkwc2\" (UID: \"4ab2d982-1c12-4585-bea3-83c246b00ed9\") " pod="openshift-marketplace/redhat-marketplace-zkwc2" Dec 05 10:43:43 crc kubenswrapper[4997]: I1205 10:43:43.044312 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ab2d982-1c12-4585-bea3-83c246b00ed9-utilities\") pod \"redhat-marketplace-zkwc2\" (UID: \"4ab2d982-1c12-4585-bea3-83c246b00ed9\") " pod="openshift-marketplace/redhat-marketplace-zkwc2" Dec 05 10:43:43 crc kubenswrapper[4997]: I1205 10:43:43.044558 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ab2d982-1c12-4585-bea3-83c246b00ed9-catalog-content\") pod \"redhat-marketplace-zkwc2\" (UID: \"4ab2d982-1c12-4585-bea3-83c246b00ed9\") " pod="openshift-marketplace/redhat-marketplace-zkwc2" Dec 05 10:43:43 crc kubenswrapper[4997]: I1205 10:43:43.071433 4997 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzxg7\" (UniqueName: \"kubernetes.io/projected/4ab2d982-1c12-4585-bea3-83c246b00ed9-kube-api-access-dzxg7\") pod \"redhat-marketplace-zkwc2\" (UID: \"4ab2d982-1c12-4585-bea3-83c246b00ed9\") " pod="openshift-marketplace/redhat-marketplace-zkwc2" Dec 05 10:43:43 crc kubenswrapper[4997]: I1205 10:43:43.148984 4997 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zkwc2" Dec 05 10:43:43 crc kubenswrapper[4997]: I1205 10:43:43.719406 4997 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zkwc2"] Dec 05 10:43:44 crc kubenswrapper[4997]: I1205 10:43:44.125404 4997 generic.go:334] "Generic (PLEG): container finished" podID="4ab2d982-1c12-4585-bea3-83c246b00ed9" containerID="b9e636e8feda21dec91d3f0551000f5e1d7e908474baebab32cd453d3722f453" exitCode=0 Dec 05 10:43:44 crc kubenswrapper[4997]: I1205 10:43:44.125448 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkwc2" event={"ID":"4ab2d982-1c12-4585-bea3-83c246b00ed9","Type":"ContainerDied","Data":"b9e636e8feda21dec91d3f0551000f5e1d7e908474baebab32cd453d3722f453"} Dec 05 10:43:44 crc kubenswrapper[4997]: I1205 10:43:44.125472 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkwc2" event={"ID":"4ab2d982-1c12-4585-bea3-83c246b00ed9","Type":"ContainerStarted","Data":"0bfe1a8789652b50830319a927c4fdfd64fe9ef2645914637db93001f5a54ccd"} Dec 05 10:43:45 crc kubenswrapper[4997]: I1205 10:43:45.137179 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkwc2" event={"ID":"4ab2d982-1c12-4585-bea3-83c246b00ed9","Type":"ContainerStarted","Data":"cd51cce7bc15aa7b699ccdfacdf004b5de33abd9b5aa2b17bf63a685c4416c38"} Dec 05 10:43:46 crc kubenswrapper[4997]: I1205 10:43:46.153914 4997 generic.go:334] "Generic (PLEG): container finished" podID="4ab2d982-1c12-4585-bea3-83c246b00ed9" containerID="cd51cce7bc15aa7b699ccdfacdf004b5de33abd9b5aa2b17bf63a685c4416c38" exitCode=0 Dec 05 10:43:46 crc kubenswrapper[4997]: I1205 10:43:46.155648 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkwc2" event={"ID":"4ab2d982-1c12-4585-bea3-83c246b00ed9","Type":"ContainerDied","Data":"cd51cce7bc15aa7b699ccdfacdf004b5de33abd9b5aa2b17bf63a685c4416c38"} Dec 05 10:43:47 crc kubenswrapper[4997]: I1205 10:43:47.168232 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkwc2" event={"ID":"4ab2d982-1c12-4585-bea3-83c246b00ed9","Type":"ContainerStarted","Data":"c17028571a59f2a336092a387268f433889eddeaca2ff616171319a41f9dab36"} Dec 05 10:43:47 crc kubenswrapper[4997]: I1205 10:43:47.195062 4997 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zkwc2" podStartSLOduration=2.765379878 podStartE2EDuration="5.195044089s" podCreationTimestamp="2025-12-05 10:43:42 +0000 UTC" firstStartedPulling="2025-12-05 10:43:44.128288995 +0000 UTC m=+13724.657196286" lastFinishedPulling="2025-12-05 10:43:46.557953236 +0000 UTC m=+13727.086860497" observedRunningTime="2025-12-05 10:43:47.183402124 +0000 UTC m=+13727.712309415" watchObservedRunningTime="2025-12-05 10:43:47.195044089 +0000 UTC m=+13727.723951350" Dec 05 10:43:53 crc kubenswrapper[4997]: I1205 10:43:53.150582 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zkwc2" Dec 05 10:43:53 crc kubenswrapper[4997]: I1205 10:43:53.151097 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zkwc2" Dec 05 10:43:53 crc kubenswrapper[4997]: I1205 10:43:53.213834 4997 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zkwc2" Dec 05 10:43:53 crc kubenswrapper[4997]: I1205 10:43:53.301569 4997 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zkwc2" Dec 05 10:43:53 crc kubenswrapper[4997]: I1205 10:43:53.463166 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zkwc2"] Dec 05 10:43:55 crc kubenswrapper[4997]: I1205 10:43:55.267761 4997 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zkwc2" podUID="4ab2d982-1c12-4585-bea3-83c246b00ed9" containerName="registry-server" containerID="cri-o://c17028571a59f2a336092a387268f433889eddeaca2ff616171319a41f9dab36" gracePeriod=2 Dec 05 10:43:56 crc kubenswrapper[4997]: I1205 10:43:56.284321 4997 generic.go:334] "Generic (PLEG): container finished" podID="4ab2d982-1c12-4585-bea3-83c246b00ed9" containerID="c17028571a59f2a336092a387268f433889eddeaca2ff616171319a41f9dab36" exitCode=0 Dec 05 10:43:56 crc kubenswrapper[4997]: I1205 10:43:56.284399 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkwc2" event={"ID":"4ab2d982-1c12-4585-bea3-83c246b00ed9","Type":"ContainerDied","Data":"c17028571a59f2a336092a387268f433889eddeaca2ff616171319a41f9dab36"} Dec 05 10:43:56 crc kubenswrapper[4997]: I1205 10:43:56.429248 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zkwc2" Dec 05 10:43:56 crc kubenswrapper[4997]: I1205 10:43:56.475230 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzxg7\" (UniqueName: \"kubernetes.io/projected/4ab2d982-1c12-4585-bea3-83c246b00ed9-kube-api-access-dzxg7\") pod \"4ab2d982-1c12-4585-bea3-83c246b00ed9\" (UID: \"4ab2d982-1c12-4585-bea3-83c246b00ed9\") " Dec 05 10:43:56 crc kubenswrapper[4997]: I1205 10:43:56.475308 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ab2d982-1c12-4585-bea3-83c246b00ed9-utilities\") pod \"4ab2d982-1c12-4585-bea3-83c246b00ed9\" (UID: \"4ab2d982-1c12-4585-bea3-83c246b00ed9\") " Dec 05 10:43:56 crc kubenswrapper[4997]: I1205 10:43:56.475392 4997 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ab2d982-1c12-4585-bea3-83c246b00ed9-catalog-content\") pod \"4ab2d982-1c12-4585-bea3-83c246b00ed9\" (UID: \"4ab2d982-1c12-4585-bea3-83c246b00ed9\") " Dec 05 10:43:56 crc kubenswrapper[4997]: I1205 10:43:56.476325 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ab2d982-1c12-4585-bea3-83c246b00ed9-utilities" (OuterVolumeSpecName: "utilities") pod "4ab2d982-1c12-4585-bea3-83c246b00ed9" (UID: "4ab2d982-1c12-4585-bea3-83c246b00ed9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:43:56 crc kubenswrapper[4997]: I1205 10:43:56.484886 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ab2d982-1c12-4585-bea3-83c246b00ed9-kube-api-access-dzxg7" (OuterVolumeSpecName: "kube-api-access-dzxg7") pod "4ab2d982-1c12-4585-bea3-83c246b00ed9" (UID: "4ab2d982-1c12-4585-bea3-83c246b00ed9"). InnerVolumeSpecName "kube-api-access-dzxg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:43:56 crc kubenswrapper[4997]: I1205 10:43:56.495921 4997 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ab2d982-1c12-4585-bea3-83c246b00ed9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ab2d982-1c12-4585-bea3-83c246b00ed9" (UID: "4ab2d982-1c12-4585-bea3-83c246b00ed9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:43:56 crc kubenswrapper[4997]: I1205 10:43:56.578474 4997 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzxg7\" (UniqueName: \"kubernetes.io/projected/4ab2d982-1c12-4585-bea3-83c246b00ed9-kube-api-access-dzxg7\") on node \"crc\" DevicePath \"\"" Dec 05 10:43:56 crc kubenswrapper[4997]: I1205 10:43:56.578595 4997 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ab2d982-1c12-4585-bea3-83c246b00ed9-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:43:56 crc kubenswrapper[4997]: I1205 10:43:56.578692 4997 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ab2d982-1c12-4585-bea3-83c246b00ed9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:43:57 crc kubenswrapper[4997]: I1205 10:43:57.300254 4997 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkwc2" event={"ID":"4ab2d982-1c12-4585-bea3-83c246b00ed9","Type":"ContainerDied","Data":"0bfe1a8789652b50830319a927c4fdfd64fe9ef2645914637db93001f5a54ccd"} Dec 05 10:43:57 crc kubenswrapper[4997]: I1205 10:43:57.300656 4997 scope.go:117] "RemoveContainer" containerID="c17028571a59f2a336092a387268f433889eddeaca2ff616171319a41f9dab36" Dec 05 10:43:57 crc kubenswrapper[4997]: I1205 10:43:57.300342 4997 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zkwc2" Dec 05 10:43:57 crc kubenswrapper[4997]: I1205 10:43:57.340701 4997 scope.go:117] "RemoveContainer" containerID="cd51cce7bc15aa7b699ccdfacdf004b5de33abd9b5aa2b17bf63a685c4416c38" Dec 05 10:43:57 crc kubenswrapper[4997]: I1205 10:43:57.368577 4997 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zkwc2"] Dec 05 10:43:57 crc kubenswrapper[4997]: I1205 10:43:57.381526 4997 scope.go:117] "RemoveContainer" containerID="b9e636e8feda21dec91d3f0551000f5e1d7e908474baebab32cd453d3722f453" Dec 05 10:43:57 crc kubenswrapper[4997]: I1205 10:43:57.383316 4997 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zkwc2"] Dec 05 10:43:57 crc kubenswrapper[4997]: I1205 10:43:57.760807 4997 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ab2d982-1c12-4585-bea3-83c246b00ed9" path="/var/lib/kubelet/pods/4ab2d982-1c12-4585-bea3-83c246b00ed9/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114533654024454 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114533654017371 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114477741016520 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114477741015470 5ustar corecore